Documentation ¶
Overview ¶
Package pubsub provides an easy way to publish and receive Google Cloud Pub/Sub messages, hiding the details of the underlying server RPCs. Pub/Sub is a many-to-many, asynchronous messaging system that decouples senders and receivers.
More information about Pub/Sub is available at https://cloud.google.com/pubsub/docs
See https://godoc.org/cloud.google.com/go for authentication, timeouts, connection pooling and similar aspects of this package.
Publishing ¶
Pub/Sub messages are published to topics. A Topic may be created using Client.CreateTopic like so:
topic, err := pubsubClient.CreateTopic(context.Background(), "topic-name")
Messages may then be published to a Topic:
res := topic.Publish(ctx, &pubsub.Message{Data: []byte("payload")})
Topic.Publish queues the message for publishing and returns immediately. When enough messages have accumulated, or enough time has elapsed, the batch of messages is sent to the Pub/Sub service.
Topic.Publish returns a PublishResult, which behaves like a future: its Get method blocks until the message has been sent to the service.
The first time you call Topic.Publish on a Topic, goroutines are started in the background. To clean up these goroutines, call Topic.Stop:
topic.Stop()
Receiving ¶
To receive messages published to a Topic, clients create a Subscription for the topic. There may be more than one subscription per topic ; each message that is published to the topic will be delivered to all associated subscriptions.
A Subscription may be created like so:
sub, err := pubsubClient.CreateSubscription(context.Background(), "sub-name", pubsub.SubscriptionConfig{Topic: topic})
Messages are then consumed from a Subscription via callback.
err := sub.Receive(context.Background(), func(ctx context.Context, m *Message) { log.Printf("Got message: %s", m.Data) m.Ack() }) if err != nil && !errors.Is(err, context.Canceled) { // Handle error. }
The callback is invoked concurrently by multiple goroutines, maximizing throughput. To terminate a call to Subscription.Receive, cancel its context.
Once client code has processed the Message, it must call Message.Ack or Message.Nack; otherwise the Message will eventually be redelivered. Ack/Nack MUST be called within the Subscription.Receive handler function, and not from a goroutine. Otherwise, flow control (e.g. ReceiveSettings.MaxOutstandingMessages) will not be respected, and messages can get orphaned when cancelling Receive.
If the client cannot or doesn't want to process the message, it can call Message.Nack to speed redelivery. For more information and configuration options, see Ack Deadlines below.
Note: It is possible for a Message to be redelivered even if Message.Ack has been called. Client code must be robust to multiple deliveries of messages.
Note: This uses pubsub's streaming pull feature. This feature has properties that may be surprising. Please take a look at https://cloud.google.com/pubsub/docs/pull#streamingpull for more details on how streaming pull behaves compared to the synchronous pull method.
Streams Management ¶
The number of StreamingPull connections can be configured by setting NumGoroutines in ReceiveSettings. The default value of 10 means the client library will maintain 10 StreamingPull connections. This is more than sufficient for most use cases, as StreamingPull connections can handle up to 10 MB/s https://cloud.google.com/pubsub/quotas#resource_limits. In some cases, using too many streams can lead to client library behaving poorly as the application becomes I/O bound.
By default, the number of connections in the gRPC conn pool is min(4,GOMAXPROCS). Each connection supports up to 100 streams. Thus, if you have 4 or more CPU cores, the default setting allows a maximum of 400 streams which is already excessive for most use cases. If you want to change the limits on the number of streams, you can change the number of connections in the gRPC connection pool as shown below:
opts := []option.ClientOption{ option.WithGRPCConnectionPool(2), } client, err := pubsub.NewClient(ctx, projID, opts...)
Ack Deadlines ¶
The default pubsub deadlines are suitable for most use cases, but may be overridden. This section describes the tradeoffs that should be considered when overriding the defaults.
Behind the scenes, each message returned by the Pub/Sub server has an associated lease, known as an "ack deadline". Unless a message is acknowledged within the ack deadline, or the client requests that the ack deadline be extended, the message will become eligible for redelivery.
As a convenience, the pubsub client will automatically extend deadlines until either:
- Message.Ack or Message.Nack is called, or
- The "MaxExtension" duration elapses from the time the message is fetched from the server. This defaults to 60m.
Ack deadlines are extended periodically by the client. The period between extensions, as well as the length of the extension, automatically adjusts based on the time it takes the subscriber application to ack messages (based on the 99th percentile of ack latency). By default, this extension period is capped at 10m, but this limit can be configured by the "MaxExtensionPeriod" setting. This has the effect that subscribers that process messages quickly have their message ack deadlines extended for a short amount, whereas subscribers that process message slowly have their message ack deadlines extended for a large amount. The net effect is fewer RPCs sent from the client library.
For example, consider a subscriber that takes 3 minutes to process each message. Since the library has already recorded several 3-minute "ack latencies"s in a percentile distribution, future message extensions are sent with a value of 3 minutes, every 3 minutes. Suppose the application crashes 5 seconds after the library sends such an extension: the Pub/Sub server would wait the remaining 2m55s before re-sending the messages out to other subscribers.
Please note that by default, the client library does not use the subscription's AckDeadline for the MaxExtension value.
Slow Message Processing ¶
For use cases where message processing exceeds 30 minutes, we recommend using the base client in a pull model, since long-lived streams are periodically killed by firewalls. See the example at https://godoc.org/cloud.google.com/go/pubsub/apiv1#example-SubscriberClient-Pull-LengthyClientProcessing
Emulator ¶
To use an emulator with this library, you can set the PUBSUB_EMULATOR_HOST environment variable to the address at which your emulator is running. This will send requests to that address instead of to Pub/Sub. You can then create and use a client as usual:
// Set PUBSUB_EMULATOR_HOST environment variable. err := os.Setenv("PUBSUB_EMULATOR_HOST", "localhost:9000") if err != nil { // TODO: Handle error. } // Create client as usual. client, err := pubsub.NewClient(ctx, "my-project-id") if err != nil { // TODO: Handle error. } defer client.Close()
Index ¶
- Constants
- Variables
- func NewMessageCarrierFromPB(msg *pb.PubsubMessage) propagation.TextMapCarrier
- type AWSKinesisState
- type AckResult
- type AcknowledgeStatus
- type AuthenticationMethod
- type BigQueryConfig
- type BigQueryConfigState
- type Client
- func (c *Client) Close() error
- func (c *Client) CreateSubscription(ctx context.Context, id string, cfg SubscriptionConfig) (*Subscription, error)
- func (c *Client) CreateTopic(ctx context.Context, topicID string) (*Topic, error)
- func (c *Client) CreateTopicWithConfig(ctx context.Context, topicID string, tc *TopicConfig) (*Topic, error)
- func (c *Client) DetachSubscription(ctx context.Context, sub string) (*DetachSubscriptionResult, error)
- func (c *Client) Project() string
- func (c *Client) Snapshot(id string) *Snapshot
- func (c *Client) Snapshots(ctx context.Context) *SnapshotConfigIterator
- func (c *Client) Subscription(id string) *Subscription
- func (c *Client) SubscriptionInProject(id, projectID string) *Subscription
- func (c *Client) Subscriptions(ctx context.Context) *SubscriptionIterator
- func (c *Client) Topic(id string) *Topic
- func (c *Client) TopicInProject(id, projectID string) *Topic
- func (c *Client) Topics(ctx context.Context) *TopicIterator
- type ClientConfig
- type CloudStorageConfig
- type CloudStorageConfigState
- type CloudStorageIngestionState
- type CloudStorageOutputFormatAvroConfig
- type CloudStorageOutputFormatTextConfig
- type DeadLetterPolicy
- type DetachSubscriptionResult
- type ErrPublishingPaused
- type FlowControlSettings
- type IngestionDataSource
- type IngestionDataSourceAWSKinesis
- type IngestionDataSourceCloudStorage
- type IngestionDataSourceCloudStorageAvroFormat
- type IngestionDataSourceCloudStoragePubSubAvroFormat
- type IngestionDataSourceCloudStorageTextFormat
- type IngestionDataSourceSettings
- type LimitExceededBehavior
- type Message
- type MessageStoragePolicy
- type NoWrapper
- type OIDCToken
- type PlatformLogsSettings
- type PlatformLogsSeverity
- type PublishResult
- type PublishSettings
- type PubsubWrapper
- type PushConfig
- type ReceiveSettings
- type RetryPolicy
- type SchemaClient
- func (s *SchemaClient) Close() error
- func (c *SchemaClient) CommitSchema(ctx context.Context, schemaID string, s SchemaConfig) (*SchemaConfig, error)
- func (c *SchemaClient) CreateSchema(ctx context.Context, schemaID string, s SchemaConfig) (*SchemaConfig, error)
- func (c *SchemaClient) DeleteSchema(ctx context.Context, schemaID string) error
- func (c *SchemaClient) DeleteSchemaRevision(ctx context.Context, schemaID, revisionID string) (*SchemaConfig, error)
- func (c *SchemaClient) ListSchemaRevisions(ctx context.Context, schemaID string, view SchemaView) *SchemaIterator
- func (c *SchemaClient) RollbackSchema(ctx context.Context, schemaID, revisionID string) (*SchemaConfig, error)
- func (c *SchemaClient) Schema(ctx context.Context, schemaID string, view SchemaView) (*SchemaConfig, error)
- func (c *SchemaClient) Schemas(ctx context.Context, view SchemaView) *SchemaIterator
- func (c *SchemaClient) ValidateMessageWithConfig(ctx context.Context, msg []byte, encoding SchemaEncoding, config SchemaConfig) (*ValidateMessageResult, error)
- func (c *SchemaClient) ValidateMessageWithID(ctx context.Context, msg []byte, encoding SchemaEncoding, schemaID string) (*ValidateMessageResult, error)
- func (c *SchemaClient) ValidateSchema(ctx context.Context, schema SchemaConfig) (*ValidateSchemaResult, error)
- type SchemaConfig
- type SchemaEncoding
- type SchemaIterator
- type SchemaSettings
- type SchemaType
- type SchemaView
- type Snapshot
- type SnapshotConfig
- type SnapshotConfigIterator
- type Subscription
- func (s *Subscription) Config(ctx context.Context) (SubscriptionConfig, error)
- func (s *Subscription) CreateSnapshot(ctx context.Context, name string) (*SnapshotConfig, error)
- func (s *Subscription) Delete(ctx context.Context) error
- func (s *Subscription) Exists(ctx context.Context) (bool, error)
- func (s *Subscription) IAM() *iam.Handle
- func (s *Subscription) ID() string
- func (s *Subscription) Receive(ctx context.Context, f func(context.Context, *Message)) error
- func (s *Subscription) SeekToSnapshot(ctx context.Context, snap *Snapshot) error
- func (s *Subscription) SeekToTime(ctx context.Context, t time.Time) error
- func (s *Subscription) String() string
- func (s *Subscription) Update(ctx context.Context, cfg SubscriptionConfigToUpdate) (SubscriptionConfig, error)
- type SubscriptionConfig
- type SubscriptionConfigToUpdate
- type SubscriptionIterator
- type SubscriptionState
- type Topic
- func (t *Topic) Config(ctx context.Context) (TopicConfig, error)
- func (t *Topic) Delete(ctx context.Context) error
- func (t *Topic) Exists(ctx context.Context) (bool, error)
- func (t *Topic) Flush()
- func (t *Topic) IAM() *iam.Handle
- func (t *Topic) ID() string
- func (t *Topic) Publish(ctx context.Context, msg *Message) *PublishResult
- func (t *Topic) ResumePublish(orderingKey string)
- func (t *Topic) Stop()
- func (t *Topic) String() string
- func (t *Topic) Subscriptions(ctx context.Context) *SubscriptionIterator
- func (t *Topic) Update(ctx context.Context, cfg TopicConfigToUpdate) (TopicConfig, error)
- type TopicConfig
- type TopicConfigToUpdate
- type TopicIterator
- type TopicState
- type ValidateMessageResult
- type ValidateSchemaResult
- type Wrapper
Examples ¶
- Client.CreateSubscription
- Client.CreateSubscription (NeverExpire)
- Client.CreateTopic
- Client.CreateTopicWithConfig
- Client.Snapshots
- Client.Subscriptions
- Client.TopicInProject
- Client.Topics
- NewClient
- Snapshot.Delete
- SnapshotConfigIterator.Next
- Subscription.Config
- Subscription.CreateSnapshot
- Subscription.Delete
- Subscription.Exists
- Subscription.Receive
- Subscription.Receive (MaxExtension)
- Subscription.Receive (MaxOutstanding)
- Subscription.SeekToSnapshot
- Subscription.SeekToTime
- Subscription.Update
- Subscription.Update (PushConfigAuthenticationMethod)
- SubscriptionIterator.Next
- Topic.Delete
- Topic.Exists
- Topic.Publish
- Topic.Subscriptions
- Topic.Update
- Topic.Update (ResetMessageStoragePolicy)
- TopicIterator.Next
Constants ¶
const ( // ScopePubSub grants permissions to view and manage Pub/Sub // topics and subscriptions. ScopePubSub = "https://www.googleapis.com/auth/pubsub" // ScopeCloudPlatform grants permissions to view and manage your data // across Google Cloud Platform services. ScopeCloudPlatform = "https://www.googleapis.com/auth/cloud-platform" )
const ( // BigQueryConfigStateUnspecified is the default value. This value is unused. BigQueryConfigStateUnspecified = iota // BigQueryConfigActive means the subscription can actively send messages to BigQuery. BigQueryConfigActive // BigQueryConfigPermissionDenied means the subscription cannot write to the BigQuery table because of permission denied errors. BigQueryConfigPermissionDenied // BigQueryConfigNotFound means the subscription cannot write to the BigQuery table because it does not exist. BigQueryConfigNotFound // BigQueryConfigSchemaMismatch means the subscription cannot write to the BigQuery table due to a schema mismatch. BigQueryConfigSchemaMismatch )
const ( // CloudStorageConfigStateUnspecified is the default value. This value is unused. CloudStorageConfigStateUnspecified = iota // CloudStorageConfigActive means the subscription can actively send messages to Cloud Storage. CloudStorageConfigActive // CloudStorageConfigPermissionDenied means the subscription cannot write to the Cloud storage bucket because of permission denied errors. CloudStorageConfigPermissionDenied // CloudStorageConfigNotFound means the subscription cannot write to the Cloud Storage bucket because it does not exist. CloudStorageConfigNotFound )
const ( // SubscriptionStateUnspecified is the default value. This value is unused. SubscriptionStateUnspecified = iota // SubscriptionStateActive means the subscription can actively send messages to BigQuery. SubscriptionStateActive // SubscriptionStateResourceError means the subscription receive messages because of an // error with the resource to which it pushes messages. // See the more detailed error state in the corresponding configuration. SubscriptionStateResourceError )
const ( // MaxPublishRequestCount is the maximum number of messages that can be in // a single publish request, as defined by the PubSub service. MaxPublishRequestCount = 1000 // MaxPublishRequestBytes is the maximum size of a single publish request // in bytes, as defined by the PubSub service. MaxPublishRequestBytes = 1e7 )
const ( // TopicStateUnspecified is the default value. This value is unused. TopicStateUnspecified = iota // TopicStateActive means the topic does not have any persistent errors. TopicStateActive // TopicStateIngestionResourceError means ingestion from the data source // has encountered a permanent error. // See the more detailed error state in the corresponding ingestion // source configuration. TopicStateIngestionResourceError )
const ( // AWSKinesisStateUnspecified is the default value. This value is unused. AWSKinesisStateUnspecified = iota // AWSKinesisStateActive means ingestion is active. AWSKinesisStateActive // AWSKinesisStatePermissionDenied means encountering an error while consumign data from Kinesis. // This can happen if: // - The provided `aws_role_arn` does not exist or does not have the // appropriate permissions attached. // - The provided `aws_role_arn` is not set up properly for Identity // Federation using `gcp_service_account`. // - The Pub/Sub SA is not granted the // `iam.serviceAccounts.getOpenIdToken` permission on // `gcp_service_account`. AWSKinesisStatePermissionDenied // AWSKinesisStatePublishPermissionDenied means permission denied encountered while publishing to the topic. // This can happen due to Pub/Sub SA has not been granted the appropriate publish // permissions https://cloud.google.com/pubsub/docs/access-control#pubsub.publisher AWSKinesisStatePublishPermissionDenied // AWSKinesisStateStreamNotFound means the Kinesis stream does not exist. AWSKinesisStateStreamNotFound // AWSKinesisStateConsumerNotFound means the Kinesis consumer does not exist. AWSKinesisStateConsumerNotFound )
const ( // CloudStorageIngestionStateUnspecified is the default value. This value is unused. CloudStorageIngestionStateUnspecified = iota // CloudStorageIngestionStateActive means ingestion is active. CloudStorageIngestionStateActive // CloudStorageIngestionPermissionDenied means encountering an error while calling the Cloud Storage API. // This can happen if the Pub/Sub SA has not been granted the // [appropriate permissions](https://cloud.google.com/storage/docs/access-control/iam-permissions): // - storage.objects.list: to list the objects in a bucket. // - storage.objects.get: to read the objects in a bucket. // - storage.buckets.get: to verify the bucket exists. CloudStorageIngestionPermissionDenied // CloudStorageIngestionPublishPermissionDenied means encountering an error when publishing to the topic. // This can happen if the Pub/Sub SA has not been granted the [appropriate publish // permissions](https://cloud.google.com/pubsub/docs/access-control#pubsub.publisher) CloudStorageIngestionPublishPermissionDenied // CloudStorageIngestionBucketNotFound means the provided bucket doesn't exist. CloudStorageIngestionBucketNotFound // CloudStorageIngestionTooManyObjects means the bucket has too many objects, ingestion will be paused. CloudStorageIngestionTooManyObjects )
const DetectProjectID = "*detect-project-id*"
DetectProjectID is a sentinel value that instructs NewClient to detect the project ID. It is given in place of the projectID argument. NewClient will use the project ID from the given credentials or the default credentials (https://developers.google.com/accounts/docs/application-default-credentials) if no credentials were provided. When providing credentials, not all options will allow NewClient to extract the project ID. Specifically a JWT does not have the project ID encoded.
Variables ¶
var ( // ErrFlowControllerMaxOutstandingMessages indicates that outstanding messages exceeds MaxOutstandingMessages. ErrFlowControllerMaxOutstandingMessages = errors.New("pubsub: MaxOutstandingMessages flow controller limit exceeded") // ErrFlowControllerMaxOutstandingBytes indicates that outstanding bytes of messages exceeds MaxOutstandingBytes. ErrFlowControllerMaxOutstandingBytes = errors.New("pubsub: MaxOutstandingBytes flow control limit exceeded") )
var ( // PublishedMessages is a measure of the number of messages published, which may include errors. // It is EXPERIMENTAL and subject to change or removal without notice. PublishedMessages = stats.Int64(statsPrefix+"published_messages", "Number of PubSub message published", stats.UnitDimensionless) // PublishLatency is a measure of the number of milliseconds it took to publish a bundle, // which may consist of one or more messages. // It is EXPERIMENTAL and subject to change or removal without notice. PublishLatency = stats.Float64(statsPrefix+"publish_roundtrip_latency", "The latency in milliseconds per publish batch", stats.UnitMilliseconds) // PullCount is a measure of the number of messages pulled. // It is EXPERIMENTAL and subject to change or removal without notice. PullCount = stats.Int64(statsPrefix+"pull_count", "Number of PubSub messages pulled", stats.UnitDimensionless) // AckCount is a measure of the number of messages acked. // It is EXPERIMENTAL and subject to change or removal without notice. AckCount = stats.Int64(statsPrefix+"ack_count", "Number of PubSub messages acked", stats.UnitDimensionless) // NackCount is a measure of the number of messages nacked. // It is EXPERIMENTAL and subject to change or removal without notice. NackCount = stats.Int64(statsPrefix+"nack_count", "Number of PubSub messages nacked", stats.UnitDimensionless) // ModAckCount is a measure of the number of messages whose ack-deadline was modified. // It is EXPERIMENTAL and subject to change or removal without notice. ModAckCount = stats.Int64(statsPrefix+"mod_ack_count", "Number of ack-deadlines modified", stats.UnitDimensionless) // ModAckTimeoutCount is a measure of the number ModifyAckDeadline RPCs that timed out. // It is EXPERIMENTAL and subject to change or removal without notice. ModAckTimeoutCount = stats.Int64(statsPrefix+"mod_ack_timeout_count", "Number of ModifyAckDeadline RPCs that timed out", stats.UnitDimensionless) // StreamOpenCount is a measure of the number of times a streaming-pull stream was opened. // It is EXPERIMENTAL and subject to change or removal without notice. StreamOpenCount = stats.Int64(statsPrefix+"stream_open_count", "Number of calls opening a new streaming pull", stats.UnitDimensionless) // StreamRetryCount is a measure of the number of times a streaming-pull operation was retried. // It is EXPERIMENTAL and subject to change or removal without notice. StreamRetryCount = stats.Int64(statsPrefix+"stream_retry_count", "Number of retries of a stream send or receive", stats.UnitDimensionless) // StreamRequestCount is a measure of the number of requests sent on a streaming-pull stream. // It is EXPERIMENTAL and subject to change or removal without notice. StreamRequestCount = stats.Int64(statsPrefix+"stream_request_count", "Number gRPC StreamingPull request messages sent", stats.UnitDimensionless) // StreamResponseCount is a measure of the number of responses received on a streaming-pull stream. // It is EXPERIMENTAL and subject to change or removal without notice. StreamResponseCount = stats.Int64(statsPrefix+"stream_response_count", "Number of gRPC StreamingPull response messages received", stats.UnitDimensionless) // OutstandingMessages is a measure of the number of outstanding messages held by the client before they are processed. // It is EXPERIMENTAL and subject to change or removal without notice. OutstandingMessages = stats.Int64(statsPrefix+"outstanding_messages", "Number of outstanding Pub/Sub messages", stats.UnitDimensionless) // OutstandingBytes is a measure of the number of bytes all outstanding messages held by the client take up. // It is EXPERIMENTAL and subject to change or removal without notice. OutstandingBytes = stats.Int64(statsPrefix+"outstanding_bytes", "Number of outstanding bytes", stats.UnitDimensionless) // PublisherOutstandingMessages is a measure of the number of published outstanding messages held by the client before they are processed. // It is EXPERIMENTAL and subject to change or removal without notice. PublisherOutstandingMessages = stats.Int64(statsPrefix+"publisher_outstanding_messages", "Number of outstanding publish messages", stats.UnitDimensionless) // PublisherOutstandingBytes is a measure of the number of bytes all outstanding publish messages held by the client take up. // It is EXPERIMENTAL and subject to change or removal without notice. PublisherOutstandingBytes = stats.Int64(statsPrefix+"publisher_outstanding_bytes", "Number of outstanding publish bytes", stats.UnitDimensionless) )
The following are measures recorded in publish/subscribe flows.
var ( // PublishedMessagesView is a cumulative sum of PublishedMessages. // It is EXPERIMENTAL and subject to change or removal without notice. PublishedMessagesView *view.View // PublishLatencyView is a distribution of PublishLatency. // It is EXPERIMENTAL and subject to change or removal without notice. PublishLatencyView *view.View // PullCountView is a cumulative sum of PullCount. // It is EXPERIMENTAL and subject to change or removal without notice. PullCountView *view.View // AckCountView is a cumulative sum of AckCount. // It is EXPERIMENTAL and subject to change or removal without notice. AckCountView *view.View // NackCountView is a cumulative sum of NackCount. // It is EXPERIMENTAL and subject to change or removal without notice. NackCountView *view.View // ModAckCountView is a cumulative sum of ModAckCount. // It is EXPERIMENTAL and subject to change or removal without notice. ModAckCountView *view.View // ModAckTimeoutCountView is a cumulative sum of ModAckTimeoutCount. // It is EXPERIMENTAL and subject to change or removal without notice. ModAckTimeoutCountView *view.View // StreamOpenCountView is a cumulative sum of StreamOpenCount. // It is EXPERIMENTAL and subject to change or removal without notice. StreamOpenCountView *view.View // StreamRetryCountView is a cumulative sum of StreamRetryCount. // It is EXPERIMENTAL and subject to change or removal without notice. StreamRetryCountView *view.View // StreamRequestCountView is a cumulative sum of StreamRequestCount. // It is EXPERIMENTAL and subject to change or removal without notice. StreamRequestCountView *view.View // StreamResponseCountView is a cumulative sum of StreamResponseCount. // It is EXPERIMENTAL and subject to change or removal without notice. StreamResponseCountView *view.View // OutstandingMessagesView is the last value of OutstandingMessages // It is EXPERIMENTAL and subject to change or removal without notice. OutstandingMessagesView *view.View // OutstandingBytesView is the last value of OutstandingBytes // It is EXPERIMENTAL and subject to change or removal without notice. OutstandingBytesView *view.View // PublisherOutstandingMessagesView is the last value of OutstandingMessages // It is EXPERIMENTAL and subject to change or removal without notice. PublisherOutstandingMessagesView *view.View // PublisherOutstandingBytesView is the last value of OutstandingBytes // It is EXPERIMENTAL and subject to change or removal without notice. PublisherOutstandingBytesView *view.View )
var ( DefaultPublishViews []*view.View DefaultSubscribeViews []*view.View )
These arrays hold the default OpenCensus views that keep track of publish/subscribe operations. It is EXPERIMENTAL and subject to change or removal without notice.
var DefaultPublishSettings = PublishSettings{ DelayThreshold: 10 * time.Millisecond, CountThreshold: 100, ByteThreshold: 1e6, Timeout: 60 * time.Second, BufferedByteLimit: 10 * MaxPublishRequestBytes, FlowControlSettings: FlowControlSettings{ MaxOutstandingMessages: 1000, MaxOutstandingBytes: -1, LimitExceededBehavior: FlowControlIgnore, }, EnableCompression: false, CompressionBytesThreshold: 240, }
DefaultPublishSettings holds the default values for topics' PublishSettings.
var DefaultReceiveSettings = ReceiveSettings{ MaxExtension: 60 * time.Minute, MaxExtensionPeriod: 0, MinExtensionPeriod: 0, MaxOutstandingMessages: 1000, MaxOutstandingBytes: 1e9, NumGoroutines: 10, }
DefaultReceiveSettings holds the default values for ReceiveSettings.
var ErrEmptyProjectID = errors.New("pubsub: projectID string is empty")
ErrEmptyProjectID denotes that the project string passed into NewClient was empty. Please provide a valid project ID or use the DetectProjectID sentinel value to detect project ID from well defined sources.
var ErrOversizedMessage = bundler.ErrOversizedItem
ErrOversizedMessage indicates that a message's size exceeds MaxPublishRequestBytes.
var ErrTopicStopped = errors.New("pubsub: Stop has been called for this topic")
ErrTopicStopped indicates that topic has been stopped and further publishing will fail.
Functions ¶
func NewMessageCarrierFromPB ¶ added in v1.43.0
func NewMessageCarrierFromPB(msg *pb.PubsubMessage) propagation.TextMapCarrier
NewMessageCarrierFromPB creates a propagation.TextMapCarrier that can be used to extract the trace context from a protobuf PubsubMessage.
Example: ctx = propagation.TraceContext{}.Extract(ctx, pubsub.NewMessageCarrierFromPB(msg))
Types ¶
type AWSKinesisState ¶ added in v1.37.0
type AWSKinesisState int
AWSKinesisState denotes the possible states for ingestion from Amazon Kinesis Data Streams.
type AckResult ¶ added in v1.25.0
AckResult holds the result from a call to Ack or Nack.
Call Get to obtain the result of the Ack/NackWithResult call. Example:
// Get blocks until Ack/NackWithResult completes or ctx is done. ackStatus, err := r.Get(ctx) if err != nil { // TODO: Handle error. }
type AcknowledgeStatus ¶ added in v1.25.0
type AcknowledgeStatus = ipubsub.AcknowledgeStatus
AcknowledgeStatus represents the status of an Ack or Nack request.
const ( // AcknowledgeStatusSuccess indicates the request was a success. AcknowledgeStatusSuccess AcknowledgeStatus = iota // AcknowledgeStatusPermissionDenied indicates the caller does not have sufficient permissions. AcknowledgeStatusPermissionDenied // AcknowledgeStatusFailedPrecondition indicates the request encountered a FailedPrecondition error. AcknowledgeStatusFailedPrecondition // AcknowledgeStatusInvalidAckID indicates one or more of the ack IDs sent were invalid. AcknowledgeStatusInvalidAckID // AcknowledgeStatusOther indicates another unknown error was returned. AcknowledgeStatusOther )
type AuthenticationMethod ¶
type AuthenticationMethod interface {
// contains filtered or unexported methods
}
AuthenticationMethod is used by push subscriptions to verify the source of push requests.
type BigQueryConfig ¶ added in v1.23.0
type BigQueryConfig struct { // The name of the table to which to write data, of the form // {projectId}:{datasetId}.{tableId} Table string // When true, use the topic's schema as the columns to write to in BigQuery, // if it exists. UseTopicSchema bool // When true, write the subscription name, message_id, publish_time, // attributes, and ordering_key to additional columns in the table. The // subscription name, message_id, and publish_time fields are put in their own // columns while all other message properties (other than data) are written to // a JSON object in the attributes column. WriteMetadata bool // When true and use_topic_schema is true, any fields that are a part of the // topic schema that are not part of the BigQuery table schema are dropped // when writing to BigQuery. Otherwise, the schemas must be kept in sync and // any messages with extra fields are not written and remain in the // subscription's backlog. DropUnknownFields bool // This is an output-only field that indicates whether or not the subscription can // receive messages. This field is set only in responses from the server; // it is ignored if it is set in any requests. State BigQueryConfigState }
BigQueryConfig configures the subscription to deliver to a BigQuery table.
type BigQueryConfigState ¶ added in v1.23.0
type BigQueryConfigState int
BigQueryConfigState denotes the possible states for a BigQuery Subscription.
type Client ¶
type Client struct {
// contains filtered or unexported fields
}
Client is a Google Pub/Sub client scoped to a single project.
Clients should be reused rather than being created as needed. A Client may be shared by multiple goroutines.
func NewClient ¶
func NewClient(ctx context.Context, projectID string, opts ...option.ClientOption) (c *Client, err error)
NewClient creates a new PubSub client. It uses a default configuration.
Example ¶
package main import ( "context" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() _, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } // See the other examples to learn how to use the Client. }
Output:
func NewClientWithConfig ¶ added in v1.15.0
func NewClientWithConfig(ctx context.Context, projectID string, config *ClientConfig, opts ...option.ClientOption) (*Client, error)
NewClientWithConfig creates a new PubSub client.
func (*Client) Close ¶
Close releases any resources held by the client, such as memory and goroutines.
If the client is available for the lifetime of the program, then Close need not be called at exit.
func (*Client) CreateSubscription ¶
func (c *Client) CreateSubscription(ctx context.Context, id string, cfg SubscriptionConfig) (*Subscription, error)
CreateSubscription creates a new subscription on a topic.
id is the name of the subscription to create. It must start with a letter, and contain only letters ([A-Za-z]), numbers ([0-9]), dashes (-), underscores (_), periods (.), tildes (~), plus (+) or percent signs (%). It must be between 3 and 255 characters in length, and must not start with "goog".
cfg.Topic is the topic from which the subscription should receive messages. It need not belong to the same project as the subscription. This field is required.
cfg.AckDeadline is the maximum time after a subscriber receives a message before the subscriber should acknowledge the message. It must be between 10 and 600 seconds (inclusive), and is rounded down to the nearest second. If the provided ackDeadline is 0, then the default value of 10 seconds is used. Note: messages which are obtained via Subscription.Receive need not be acknowledged within this deadline, as the deadline will be automatically extended.
cfg.PushConfig may be set to configure this subscription for push delivery.
If the subscription already exists an error will be returned.
Example ¶
package main import ( "context" "time" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } // Create a new topic with the given name. topic, err := client.CreateTopic(ctx, "topicName") if err != nil { // TODO: Handle error. } // Create a new subscription to the previously created topic // with the given name. sub, err := client.CreateSubscription(ctx, "subName", pubsub.SubscriptionConfig{ Topic: topic, AckDeadline: 10 * time.Second, ExpirationPolicy: 25 * time.Hour, }) if err != nil { // TODO: Handle error. } _ = sub // TODO: use the subscription. }
Output:
Example (NeverExpire) ¶
package main import ( "context" "time" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } // Create a new topic with the given name. topic, err := client.CreateTopic(ctx, "topicName") if err != nil { // TODO: Handle error. } // Create a new subscription to the previously // created topic and ensure it never expires. sub, err := client.CreateSubscription(ctx, "subName", pubsub.SubscriptionConfig{ Topic: topic, AckDeadline: 10 * time.Second, ExpirationPolicy: time.Duration(0), }) if err != nil { // TODO: Handle error. } _ = sub // TODO: Use the subscription }
Output:
func (*Client) CreateTopic ¶
CreateTopic creates a new topic.
The specified topic ID must start with a letter, and contain only letters ([A-Za-z]), numbers ([0-9]), dashes (-), underscores (_), periods (.), tildes (~), plus (+) or percent signs (%). It must be between 3 and 255 characters in length, and must not start with "goog". For more information, see: https://cloud.google.com/pubsub/docs/admin#resource_names
If the topic already exists an error will be returned.
Example ¶
package main import ( "context" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } // Create a new topic with the given name. topic, err := client.CreateTopic(ctx, "topicName") if err != nil { // TODO: Handle error. } _ = topic // TODO: use the topic. }
Output:
func (*Client) CreateTopicWithConfig ¶
func (c *Client) CreateTopicWithConfig(ctx context.Context, topicID string, tc *TopicConfig) (*Topic, error)
CreateTopicWithConfig creates a topic from TopicConfig.
The specified topic ID must start with a letter, and contain only letters ([A-Za-z]), numbers ([0-9]), dashes (-), underscores (_), periods (.), tildes (~), plus (+) or percent signs (%). It must be between 3 and 255 characters in length, and must not start with "goog". For more information, see: https://cloud.google.com/pubsub/docs/admin#resource_names.
If the topic already exists, an error will be returned.
Example ¶
package main import ( "context" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } // Create a new topic with the given name and config. topicConfig := &pubsub.TopicConfig{ KMSKeyName: "projects/project-id/locations/global/keyRings/my-key-ring/cryptoKeys/my-key", MessageStoragePolicy: pubsub.MessageStoragePolicy{ AllowedPersistenceRegions: []string{"us-east1"}, }, } topic, err := client.CreateTopicWithConfig(ctx, "topicName", topicConfig) if err != nil { // TODO: Handle error. } _ = topic // TODO: use the topic. }
Output:
func (*Client) DetachSubscription ¶ added in v1.5.0
func (c *Client) DetachSubscription(ctx context.Context, sub string) (*DetachSubscriptionResult, error)
DetachSubscription detaches a subscription from its topic. All messages retained in the subscription are dropped. Subsequent `Pull` and `StreamingPull` requests will return FAILED_PRECONDITION. If the subscription is a push subscription, pushes to the endpoint will stop.
func (*Client) Project ¶ added in v1.32.0
Project returns the project ID or number for this instance of the client, which may have either been explicitly specified or autodetected.
func (*Client) Snapshots ¶
func (c *Client) Snapshots(ctx context.Context) *SnapshotConfigIterator
Snapshots returns an iterator which returns snapshots for this project.
Example ¶
package main import ( "context" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } // List all snapshots for the project. iter := client.Snapshots(ctx) _ = iter // TODO: iterate using Next. }
Output:
func (*Client) Subscription ¶
func (c *Client) Subscription(id string) *Subscription
Subscription creates a reference to a subscription.
func (*Client) SubscriptionInProject ¶
func (c *Client) SubscriptionInProject(id, projectID string) *Subscription
SubscriptionInProject creates a reference to a subscription in a given project.
func (*Client) Subscriptions ¶
func (c *Client) Subscriptions(ctx context.Context) *SubscriptionIterator
Subscriptions returns an iterator which returns all of the subscriptions for the client's project.
Example ¶
package main import ( "context" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } // List all subscriptions of the project. it := client.Subscriptions(ctx) _ = it // TODO: iterate using Next. }
Output:
func (*Client) Topic ¶
Topic creates a reference to a topic in the client's project.
If a Topic's Publish method is called, it has background goroutines associated with it. Clean them up by calling Topic.Stop.
Avoid creating many Topic instances if you use them to publish.
func (*Client) TopicInProject ¶
TopicInProject creates a reference to a topic in the given project.
If a Topic's Publish method is called, it has background goroutines associated with it. Clean them up by calling Topic.Stop.
Avoid creating many Topic instances if you use them to publish.
Example ¶
Use TopicInProject to refer to a topic that is not in the client's project, such as a public topic.
package main import ( "context" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } topic := client.TopicInProject("topicName", "another-project-id") _ = topic // TODO: use the topic. }
Output:
func (*Client) Topics ¶
func (c *Client) Topics(ctx context.Context) *TopicIterator
Topics returns an iterator which returns all of the topics for the client's project.
Example ¶
package main import ( "context" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } it := client.Topics(ctx) _ = it // TODO: iterate using Next. }
Output:
type ClientConfig ¶ added in v1.15.0
type ClientConfig struct { PublisherCallOptions *vkit.PublisherCallOptions SubscriberCallOptions *vkit.SubscriberCallOptions // EnableOpenTelemetryTracing enables tracing for this client. // This option allows selectively disabling Pub/Sub traces. // This defaults to false. // OpenTelemetry tracing standards are in active development, and thus // attributes, links, and span names are EXPERIMENTAL and subject to // change or removal without notice. EnableOpenTelemetryTracing bool }
ClientConfig has configurations for the client.
type CloudStorageConfig ¶ added in v1.32.0
type CloudStorageConfig struct { // User-provided name for the Cloud Storage bucket. // The bucket must be created by the user. The bucket name must be without // any prefix like "gs://". See the [bucket naming // requirements] (https://cloud.google.com/storage/docs/buckets#naming). Bucket string // User-provided prefix for Cloud Storage filename. See the [object naming // requirements](https://cloud.google.com/storage/docs/objects#naming). FilenamePrefix string // User-provided suffix for Cloud Storage filename. See the [object naming // requirements](https://cloud.google.com/storage/docs/objects#naming). FilenameSuffix string // Configuration for how to write message data. Options are // CloudStorageOutputFormat_TextConfig and CloudStorageOutputFormat_AvroConfig. // Defaults to text format. OutputFormat isCloudStorageOutputFormat // The maximum duration that can elapse before a new Cloud Storage file is // created. Min 1 minute, max 10 minutes, default 5 minutes. May not exceed // the subscription's acknowledgement deadline. MaxDuration optional.Duration // The maximum bytes that can be written to a Cloud Storage file before a new // file is created. Min 1 KB, max 10 GiB. The max_bytes limit may be exceeded // in cases where messages are larger than the limit. MaxBytes int64 // Output only. An output-only field that indicates whether or not the // subscription can receive messages. State CloudStorageConfigState }
CloudStorageConfig configures the subscription to deliver to Cloud Storage.
type CloudStorageConfigState ¶ added in v1.32.0
type CloudStorageConfigState int
CloudStorageConfigState denotes the possible states for a Cloud Storage Subscription.
type CloudStorageIngestionState ¶ added in v1.44.0
type CloudStorageIngestionState int
CloudStorageIngestionState denotes the possible states for ingestion from Cloud Storage.
type CloudStorageOutputFormatAvroConfig ¶ added in v1.32.0
type CloudStorageOutputFormatAvroConfig struct { // When true, write the subscription name, message_id, publish_time, // attributes, and ordering_key as additional fields in the output. WriteMetadata bool }
CloudStorageOutputFormatAvroConfig is the configuration for writing message data in Avro format. Message payloads and metadata will be written to the files as an Avro binary.
type CloudStorageOutputFormatTextConfig ¶ added in v1.32.0
type CloudStorageOutputFormatTextConfig struct{}
CloudStorageOutputFormatTextConfig is the configuration for writing message data in text format. Message payloads will be written to files as raw text, separated by a newline.
type DeadLetterPolicy ¶ added in v1.2.0
DeadLetterPolicy specifies the conditions for dead lettering messages in a subscription.
type DetachSubscriptionResult ¶ added in v1.5.0
type DetachSubscriptionResult struct{}
DetachSubscriptionResult is the response for the DetachSubscription method. Reserved for future use.
type ErrPublishingPaused ¶ added in v1.31.0
type ErrPublishingPaused struct {
OrderingKey string
}
ErrPublishingPaused is a custom error indicating that the publish paused for the specified ordering key.
func (ErrPublishingPaused) Error ¶ added in v1.31.0
func (e ErrPublishingPaused) Error() string
type FlowControlSettings ¶ added in v1.17.0
type FlowControlSettings struct { // MaxOutstandingMessages is the maximum number of buffered messages to be published. // If less than or equal to zero, this is disabled. MaxOutstandingMessages int // MaxOutstandingBytes is the maximum size of buffered messages to be published. // If less than or equal to zero, this is disabled. MaxOutstandingBytes int // LimitExceededBehavior configures the behavior when trying to publish // additional messages while the flow controller is full. The available options // are Ignore (disable, default), Block, and SignalError (publish // results will return an error). LimitExceededBehavior LimitExceededBehavior }
FlowControlSettings controls flow control for messages while publishing or subscribing.
type IngestionDataSource ¶ added in v1.37.0
type IngestionDataSource interface {
// contains filtered or unexported methods
}
IngestionDataSource is the kind of ingestion source to be used.
type IngestionDataSourceAWSKinesis ¶ added in v1.37.0
type IngestionDataSourceAWSKinesis struct { // State is an output-only field indicating the state of the kinesis connection. State AWSKinesisState // StreamARN is the Kinesis stream ARN to ingest data from. StreamARN string // ConsumerARn is the Kinesis consumer ARN to used for ingestion in Enhanced // Fan-Out mode. The consumer must be already created and ready to be used. ConsumerARN string // AWSRoleARn is the AWS role ARN to be used for Federated Identity authentication // with Kinesis. Check the Pub/Sub docs for how to set up this role and the // required permissions that need to be attached to it. AWSRoleARN string // GCPServiceAccount is the GCP service account to be used for Federated Identity // authentication with Kinesis (via a `AssumeRoleWithWebIdentity` call for // the provided role). The `aws_role_arn` must be set up with // `accounts.google.com:sub` equals to this service account number. GCPServiceAccount string }
IngestionDataSourceAWSKinesis are ingestion settings for Amazon Kinesis Data Streams.
type IngestionDataSourceCloudStorage ¶ added in v1.44.0
type IngestionDataSourceCloudStorage struct { // State is an output-only field indicating the state of the Cloud storage ingestion source. State CloudStorageIngestionState // Bucket is the Cloud Storage bucket. The bucket name must be without any // prefix like "gs://". See the bucket naming requirements (https://cloud.google.com/storage/docs/buckets#naming). Bucket string // InputFormat is the format of objects in Cloud Storage. // Defaults to TextFormat. InputFormat ingestionDataSourceCloudStorageInputFormat // MinimumObjectCreateTime means objects with a larger or equal creation timestamp will be // ingested. MinimumObjectCreateTime time.Time // MatchGlob is the pattern used to match objects that will be ingested. If // empty, all objects will be ingested. See the [supported // patterns](https://cloud.google.com/storage/docs/json_api/v1/objects/list#list-objects-and-prefixes-using-glob). MatchGlob string }
IngestionDataSourceCloudStorage are ingestion settings for Cloud Storage.
type IngestionDataSourceCloudStorageAvroFormat ¶ added in v1.44.0
type IngestionDataSourceCloudStorageAvroFormat struct{}
IngestionDataSourceCloudStorageAvroFormat means Cloud Storage data will be interpreted in Avro format.
type IngestionDataSourceCloudStoragePubSubAvroFormat ¶ added in v1.44.0
type IngestionDataSourceCloudStoragePubSubAvroFormat struct{}
IngestionDataSourceCloudStoragePubSubAvroFormat is used assuming the data was written using Cloud Storage subscriptions https://cloud.google.com/pubsub/docs/cloudstorage.
type IngestionDataSourceCloudStorageTextFormat ¶ added in v1.44.0
type IngestionDataSourceCloudStorageTextFormat struct {
Delimiter string
}
IngestionDataSourceCloudStorageTextFormat means Cloud Storage data will be interpreted as text.
type IngestionDataSourceSettings ¶ added in v1.37.0
type IngestionDataSourceSettings struct { Source IngestionDataSource PlatformLogsSettings *PlatformLogsSettings }
IngestionDataSourceSettings enables ingestion from a data source into this topic.
type LimitExceededBehavior ¶ added in v1.17.0
type LimitExceededBehavior int
LimitExceededBehavior configures the behavior that flowController can use in case the flow control limits are exceeded.
const ( // FlowControlIgnore disables flow control. FlowControlIgnore LimitExceededBehavior = iota // FlowControlBlock signals to wait until the request can be made without exceeding the limit. FlowControlBlock // FlowControlSignalError signals an error to the caller of acquire. FlowControlSignalError )
type Message ¶
Message represents a Pub/Sub message.
Message can be passed to Topic.Publish for publishing.
If received in the callback passed to Subscription.Receive, client code must call Message.Ack or Message.Nack when finished processing the Message. Calls to Ack or Nack have no effect after the first call.
Ack indicates successful processing of a Message. If message acknowledgement fails, the Message will be redelivered. Nack indicates that the client will not or cannot process a Message. Nack will result in the Message being redelivered more quickly than if it were allowed to expire.
If using exactly once delivery, you should call Message.AckWithResult and Message.NackWithResult instead. These methods will return an AckResult, which tracks the state of acknowledgement operation. If the AckResult returns successful, the message is guaranteed NOT to be re-delivered. Otherwise, the AckResult will return an error with more details about the failure and the message may be re-delivered.
type MessageStoragePolicy ¶
type MessageStoragePolicy struct { // AllowedPersistenceRegions is the list of GCP regions where messages that are published // to the topic may be persisted in storage. Messages published by publishers running in // non-allowed GCP regions (or running outside of GCP altogether) will be // routed for storage in one of the allowed regions. // // If empty, it indicates a misconfiguration at the project or organization level, which // will result in all Publish operations failing. This field cannot be empty in updates. // // If nil, then the policy is not defined on a topic level. When used in updates, it resets // the regions back to the organization level Resource Location Restriction policy. // // For more information, see // https://cloud.google.com/pubsub/docs/resource-location-restriction#pubsub-storage-locations. AllowedPersistenceRegions []string }
MessageStoragePolicy constrains how messages published to the topic may be stored. It is determined when the topic is created based on the policy configured at the project level.
type NoWrapper ¶ added in v1.33.0
type NoWrapper struct {
WriteMetadata bool
}
NoWrapper denotes not wrapping the payload sent to the push endpoint.
type OIDCToken ¶
type OIDCToken struct { // Audience to be used when generating OIDC token. The audience claim // identifies the recipients that the JWT is intended for. The audience // value is a single case-sensitive string. Having multiple values (array) // for the audience field is not supported. More info about the OIDC JWT // token audience here: https://tools.ietf.org/html/rfc7519#section-4.1.3 // Note: if not specified, the Push endpoint URL will be used. Audience string // The service account email to be used for generating the OpenID Connect token. // The caller of: // * CreateSubscription // * UpdateSubscription // * ModifyPushConfig // calls must have the iam.serviceAccounts.actAs permission for the service account. // See https://cloud.google.com/iam/docs/understanding-roles#service-accounts-roles. ServiceAccountEmail string }
OIDCToken allows PushConfigs to be authenticated using the OpenID Connect protocol https://openid.net/connect/
type PlatformLogsSettings ¶ added in v1.45.0
type PlatformLogsSettings struct {
Severity PlatformLogsSeverity
}
PlatformLogsSettings configures logging produced by Pub/Sub. Currently only valid on Cloud Storage ingestion topics.
type PlatformLogsSeverity ¶ added in v1.45.0
type PlatformLogsSeverity int32
PlatformLogsSeverity are the severity levels of Platform Logs.
const ( // PlatformLogsSeverityUnspecified is the default value. Logs level is unspecified. Logs will be disabled. PlatformLogsSeverityUnspecified PlatformLogsSeverity = iota // PlatformLogsSeverityDisabled means logs will be disabled. PlatformLogsSeverityDisabled // PlatformLogsSeverityDebug means debug logs and higher-severity logs will be written. PlatformLogsSeverityDebug // PlatformLogsSeverityInfo means info logs and higher-severity logs will be written. PlatformLogsSeverityInfo // PlatformLogsSeverityWarning means warning logs and higher-severity logs will be written. PlatformLogsSeverityWarning // PlatformLogsSeverityError means only error logs will be written. PlatformLogsSeverityError )
type PublishResult ¶
type PublishResult = ipubsub.PublishResult
A PublishResult holds the result from a call to Publish.
Call Get to obtain the result of the Publish call. Example:
// Get blocks until Publish completes or ctx is done. id, err := r.Get(ctx) if err != nil { // TODO: Handle error. }
type PublishSettings ¶
type PublishSettings struct { // Publish a non-empty batch after this delay has passed. DelayThreshold time.Duration // Publish a batch when it has this many messages. The maximum is // MaxPublishRequestCount. CountThreshold int // Publish a batch when its size in bytes reaches this value. ByteThreshold int // The number of goroutines used in each of the data structures that are // involved along the the Publish path. Adjusting this value adjusts // concurrency along the publish path. // // Defaults to a multiple of GOMAXPROCS. NumGoroutines int // The maximum time that the client will attempt to publish a bundle of messages. Timeout time.Duration // The maximum number of bytes that the Bundler will keep in memory before // returning ErrOverflow. This is now superseded by FlowControlSettings.MaxOutstandingBytes. // If MaxOutstandingBytes is set, that value will override BufferedByteLimit. // // Defaults to DefaultPublishSettings.BufferedByteLimit. // Deprecated: Set `Topic.PublishSettings.FlowControlSettings.MaxOutstandingBytes` instead. BufferedByteLimit int // FlowControlSettings defines publisher flow control settings. FlowControlSettings FlowControlSettings // EnableCompression enables transport compression for Publish operations EnableCompression bool // CompressionBytesThreshold defines the threshold (in bytes) above which messages // are compressed for transport. Only takes effect if EnableCompression is true. CompressionBytesThreshold int }
PublishSettings control the bundling of published messages.
type PubsubWrapper ¶ added in v1.33.0
type PubsubWrapper struct{}
PubsubWrapper denotes sending the payload to the push endpoint in the form of the JSON representation of a PubsubMessage (https://cloud.google.com/pubsub/docs/reference/rpc/google.pubsub.v1#pubsubmessage).
type PushConfig ¶
type PushConfig struct { // A URL locating the endpoint to which messages should be pushed. Endpoint string // Endpoint configuration attributes. See https://cloud.google.com/pubsub/docs/reference/rest/v1/projects.subscriptions#pushconfig for more details. Attributes map[string]string // AuthenticationMethod is used by push endpoints to verify the source // of push requests. // It can be used with push endpoints that are private by default to // allow requests only from the Cloud Pub/Sub system, for example. // This field is optional and should be set only by users interested in // authenticated push. AuthenticationMethod AuthenticationMethod // The format of the delivered message to the push endpoint is defined by // the chosen wrapper. When unset, `PubsubWrapper` is used. Wrapper Wrapper }
PushConfig contains configuration for subscriptions that operate in push mode.
type ReceiveSettings ¶
type ReceiveSettings struct { // MaxExtension is the maximum period for which the Subscription should // automatically extend the ack deadline for each message. // // The Subscription will automatically extend the ack deadline of all // fetched Messages up to the duration specified. Automatic deadline // extension beyond the initial receipt may be disabled by specifying a // duration less than 0. MaxExtension time.Duration // MaxExtensionPeriod is the maximum duration by which to extend the ack // deadline at a time. The ack deadline will continue to be extended by up // to this duration until MaxExtension is reached. Setting MaxExtensionPeriod // bounds the maximum amount of time before a message redelivery in the // event the subscriber fails to extend the deadline. // // MaxExtensionPeriod must be between 10s and 600s (inclusive). This configuration // can be disabled by specifying a duration less than (or equal to) 0. MaxExtensionPeriod time.Duration // MinExtensionPeriod is the the min duration for a single lease extension attempt. // By default the 99th percentile of ack latency is used to determine lease extension // periods but this value can be set to minimize the number of extraneous RPCs sent. // // MinExtensionPeriod must be between 10s and 600s (inclusive). This configuration // can be disabled by specifying a duration less than (or equal to) 0. // Defaults to off but set to 60 seconds if the subscription has exactly-once delivery enabled, // which will be added in a future release. MinExtensionPeriod time.Duration // MaxOutstandingMessages is the maximum number of unprocessed messages // (unacknowledged but not yet expired). If MaxOutstandingMessages is 0, it // will be treated as if it were DefaultReceiveSettings.MaxOutstandingMessages. // If the value is negative, then there will be no limit on the number of // unprocessed messages. MaxOutstandingMessages int // MaxOutstandingBytes is the maximum size of unprocessed messages // (unacknowledged but not yet expired). If MaxOutstandingBytes is 0, it will // be treated as if it were DefaultReceiveSettings.MaxOutstandingBytes. If // the value is negative, then there will be no limit on the number of bytes // for unprocessed messages. MaxOutstandingBytes int // UseLegacyFlowControl disables enforcing flow control settings at the Cloud // PubSub server and the less accurate method of only enforcing flow control // at the client side is used. // The default is false. UseLegacyFlowControl bool // NumGoroutines sets the number of StreamingPull streams to pull messages // from the subscription. // // NumGoroutines defaults to DefaultReceiveSettings.NumGoroutines. // // NumGoroutines does not limit the number of messages that can be processed // concurrently. Even with one goroutine, many messages might be processed at // once, because that goroutine may continually receive messages and invoke the // function passed to Receive on them. To limit the number of messages being // processed concurrently, set MaxOutstandingMessages. NumGoroutines int // Synchronous switches the underlying receiving mechanism to unary Pull. // When Synchronous is false, the more performant StreamingPull is used. // StreamingPull also has the benefit of subscriber affinity when using // ordered delivery. // When Synchronous is true, NumGoroutines is set to 1 and only one Pull // RPC will be made to poll messages at a time. // The default is false. // // Deprecated. // Previously, users might use Synchronous mode since StreamingPull had a limitation // where MaxOutstandingMessages was not always respected with large batches of // small messages. With server side flow control, this is no longer an issue // and we recommend switching to the default StreamingPull mode by setting // Synchronous to false. // Synchronous mode does not work with exactly once delivery. Synchronous bool }
ReceiveSettings configure the Receive method. A zero ReceiveSettings will result in values equivalent to DefaultReceiveSettings.
type RetryPolicy ¶ added in v1.5.0
type RetryPolicy struct { // MinimumBackoff is the minimum delay between consecutive deliveries of a // given message. Value should be between 0 and 600 seconds. Defaults to 10 seconds. MinimumBackoff optional.Duration // MaximumBackoff is the maximum delay between consecutive deliveries of a // given message. Value should be between 0 and 600 seconds. Defaults to 600 seconds. MaximumBackoff optional.Duration }
RetryPolicy specifies how Cloud Pub/Sub retries message delivery.
Retry delay will be exponential based on provided minimum and maximum backoffs. https://en.wikipedia.org/wiki/Exponential_backoff.
RetryPolicy will be triggered on NACKs or acknowledgement deadline exceeded events for a given message.
Retry Policy is implemented on a best effort basis. At times, the delay between consecutive deliveries may not match the configuration. That is, delay can be more or less than configured backoff.
type SchemaClient ¶ added in v1.15.0
type SchemaClient struct {
// contains filtered or unexported fields
}
SchemaClient is a Pub/Sub schema client scoped to a single project.
func NewSchemaClient ¶ added in v1.15.0
func NewSchemaClient(ctx context.Context, projectID string, opts ...option.ClientOption) (*SchemaClient, error)
NewSchemaClient creates a new Pub/Sub Schema client.
func (*SchemaClient) Close ¶ added in v1.15.0
func (s *SchemaClient) Close() error
Close closes the schema client and frees up resources.
func (*SchemaClient) CommitSchema ¶ added in v1.29.0
func (c *SchemaClient) CommitSchema(ctx context.Context, schemaID string, s SchemaConfig) (*SchemaConfig, error)
CommitSchema commits a new schema revision to an existing schema.
func (*SchemaClient) CreateSchema ¶ added in v1.15.0
func (c *SchemaClient) CreateSchema(ctx context.Context, schemaID string, s SchemaConfig) (*SchemaConfig, error)
CreateSchema creates a new schema with the given schemaID and config. Schemas cannot be updated after creation.
func (*SchemaClient) DeleteSchema ¶ added in v1.15.0
func (c *SchemaClient) DeleteSchema(ctx context.Context, schemaID string) error
DeleteSchema deletes an existing schema given a schema ID.
func (*SchemaClient) DeleteSchemaRevision ¶ added in v1.29.0
func (c *SchemaClient) DeleteSchemaRevision(ctx context.Context, schemaID, revisionID string) (*SchemaConfig, error)
DeleteSchemaRevision deletes a specific schema revision.
func (*SchemaClient) ListSchemaRevisions ¶ added in v1.29.0
func (c *SchemaClient) ListSchemaRevisions(ctx context.Context, schemaID string, view SchemaView) *SchemaIterator
ListSchemaRevisions lists all schema revisions for the named schema.
func (*SchemaClient) RollbackSchema ¶ added in v1.29.0
func (c *SchemaClient) RollbackSchema(ctx context.Context, schemaID, revisionID string) (*SchemaConfig, error)
RollbackSchema creates a new schema revision that is a copy of the provided revision.
func (*SchemaClient) Schema ¶ added in v1.15.0
func (c *SchemaClient) Schema(ctx context.Context, schemaID string, view SchemaView) (*SchemaConfig, error)
Schema retrieves the configuration of a schema given a schemaID and a view.
func (*SchemaClient) Schemas ¶ added in v1.15.0
func (c *SchemaClient) Schemas(ctx context.Context, view SchemaView) *SchemaIterator
Schemas returns an iterator which returns all of the schemas for the client's project.
func (*SchemaClient) ValidateMessageWithConfig ¶ added in v1.15.0
func (c *SchemaClient) ValidateMessageWithConfig(ctx context.Context, msg []byte, encoding SchemaEncoding, config SchemaConfig) (*ValidateMessageResult, error)
ValidateMessageWithConfig validates a message against an schema specified by a schema config.
func (*SchemaClient) ValidateMessageWithID ¶ added in v1.15.0
func (c *SchemaClient) ValidateMessageWithID(ctx context.Context, msg []byte, encoding SchemaEncoding, schemaID string) (*ValidateMessageResult, error)
ValidateMessageWithID validates a message against an schema specified by the schema ID of an existing schema.
func (*SchemaClient) ValidateSchema ¶ added in v1.15.0
func (c *SchemaClient) ValidateSchema(ctx context.Context, schema SchemaConfig) (*ValidateSchemaResult, error)
ValidateSchema validates a schema config and returns an error if invalid.
type SchemaConfig ¶ added in v1.15.0
type SchemaConfig struct { // Name of the schema. // Format is `projects/{project}/schemas/{schema}` Name string // The type of the schema definition. Type SchemaType // The definition of the schema. This should contain a string representing // the full definition of the schema that is a valid schema definition of // the type specified in `type`. Definition string // RevisionID is the revision ID of the schema. // This field is output only. RevisionID string // RevisionCreateTime is the timestamp that the revision was created. // This field is output only. RevisionCreateTime time.Time }
SchemaConfig is a reference to a PubSub schema.
type SchemaEncoding ¶ added in v1.15.0
SchemaEncoding is the encoding expected for messages.
const ( // EncodingUnspecified is the default unused value. EncodingUnspecified SchemaEncoding = 0 // EncodingJSON is the JSON encoding type for a message. EncodingJSON SchemaEncoding = 1 // EncodingBinary is the binary encoding type for a message. // For some schema types, binary encoding may not be available. EncodingBinary SchemaEncoding = 2 )
type SchemaIterator ¶ added in v1.15.0
type SchemaIterator struct {
// contains filtered or unexported fields
}
SchemaIterator is a struct used to iterate over schemas.
func (*SchemaIterator) Next ¶ added in v1.15.0
func (s *SchemaIterator) Next() (*SchemaConfig, error)
Next returns the next schema. If there are no more schemas, iterator.Done will be returned.
type SchemaSettings ¶ added in v1.15.0
type SchemaSettings struct { // The name of the schema that messages published should be // validated against. Format is `projects/{project}/schemas/{schema}` Schema string // The encoding of messages validated against the schema. Encoding SchemaEncoding // The minimum (inclusive) revision allowed for validating messages. If empty // or not present, allow any revision to be validated against LastRevisionID or // any revision created before. FirstRevisionID string // The maximum (inclusive) revision allowed for validating messages. If empty // or not present, allow any revision to be validated against FirstRevisionID // or any revision created after. LastRevisionID string }
SchemaSettings are settings for validating messages published against a schema.
type SchemaType ¶ added in v1.15.0
type SchemaType pb.Schema_Type
SchemaType is the possible schema definition types.
const ( // SchemaTypeUnspecified is the unused default value. SchemaTypeUnspecified SchemaType = 0 // SchemaProtocolBuffer is a protobuf schema definition. SchemaProtocolBuffer SchemaType = 1 // SchemaAvro is an Avro schema definition. SchemaAvro SchemaType = 2 )
type SchemaView ¶ added in v1.15.0
type SchemaView pb.SchemaView
SchemaView is a view of Schema object fields to be returned by GetSchema and ListSchemas.
const ( // SchemaViewUnspecified is the default/unset value. SchemaViewUnspecified SchemaView = 0 // SchemaViewBasic includes the name and type of the schema, but not the definition. SchemaViewBasic SchemaView = 1 // SchemaViewFull includes all Schema object fields. SchemaViewFull SchemaView = 2 )
type Snapshot ¶
type Snapshot struct {
// contains filtered or unexported fields
}
Snapshot is a reference to a PubSub snapshot.
func (*Snapshot) Delete ¶
Delete deletes a snapshot.
Example ¶
package main import ( "context" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } snap := client.Snapshot("snapshotName") if err := snap.Delete(ctx); err != nil { // TODO: Handle error. } }
Output:
type SnapshotConfig ¶
type SnapshotConfig struct { *Snapshot Topic *Topic Expiration time.Time // The set of labels for the snapshot. Labels map[string]string }
SnapshotConfig contains the details of a Snapshot.
type SnapshotConfigIterator ¶
type SnapshotConfigIterator struct {
// contains filtered or unexported fields
}
SnapshotConfigIterator is an iterator that returns a series of snapshots.
func (*SnapshotConfigIterator) Next ¶
func (snaps *SnapshotConfigIterator) Next() (*SnapshotConfig, error)
Next returns the next SnapshotConfig. Its second return value is iterator.Done if there are no more results. Once Next returns iterator.Done, all subsequent calls will return iterator.Done.
Example ¶
package main import ( "context" "cloud.google.com/go/pubsub" "google.golang.org/api/iterator" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } // List all snapshots for the project. iter := client.Snapshots(ctx) for { snapConfig, err := iter.Next() if err == iterator.Done { break } if err != nil { // TODO: Handle error. } _ = snapConfig // TODO: use the SnapshotConfig. } }
Output:
type Subscription ¶
type Subscription struct { // Settings for pulling messages. Configure these before calling Receive. ReceiveSettings ReceiveSettings // contains filtered or unexported fields }
Subscription is a reference to a PubSub subscription.
func (*Subscription) Config ¶
func (s *Subscription) Config(ctx context.Context) (SubscriptionConfig, error)
Config fetches the current configuration for the subscription.
Example ¶
package main import ( "context" "fmt" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } sub := client.Subscription("subName") config, err := sub.Config(ctx) if err != nil { // TODO: Handle error. } fmt.Println(config) }
Output:
func (*Subscription) CreateSnapshot ¶
func (s *Subscription) CreateSnapshot(ctx context.Context, name string) (*SnapshotConfig, error)
CreateSnapshot creates a new snapshot from this subscription. The snapshot will be for the topic this subscription is subscribed to. If the name is empty string, a unique name is assigned.
The created snapshot is guaranteed to retain:
(a) The existing backlog on the subscription. More precisely, this is defined as the messages in the subscription's backlog that are unacknowledged when Snapshot returns without error. (b) Any messages published to the subscription's topic following Snapshot returning without error.
Example ¶
package main import ( "context" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } sub := client.Subscription("subName") snapConfig, err := sub.CreateSnapshot(ctx, "snapshotName") if err != nil { // TODO: Handle error. } _ = snapConfig // TODO: Use SnapshotConfig. }
Output:
func (*Subscription) Delete ¶
func (s *Subscription) Delete(ctx context.Context) error
Delete deletes the subscription.
Example ¶
package main import ( "context" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } sub := client.Subscription("subName") if err := sub.Delete(ctx); err != nil { // TODO: Handle error. } }
Output:
func (*Subscription) Exists ¶
func (s *Subscription) Exists(ctx context.Context) (bool, error)
Exists reports whether the subscription exists on the server.
Example ¶
package main import ( "context" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } sub := client.Subscription("subName") ok, err := sub.Exists(ctx) if err != nil { // TODO: Handle error. } if !ok { // Subscription doesn't exist. } }
Output:
func (*Subscription) IAM ¶
func (s *Subscription) IAM() *iam.Handle
IAM returns the subscription's IAM handle.
func (*Subscription) ID ¶
func (s *Subscription) ID() string
ID returns the unique identifier of the subscription within its project.
func (*Subscription) Receive ¶
Receive calls f with the outstanding messages from the subscription. It blocks until ctx is done, or the service returns a non-retryable error.
The standard way to terminate a Receive is to cancel its context:
cctx, cancel := context.WithCancel(ctx) err := sub.Receive(cctx, callback) // Call cancel from callback, or another goroutine.
If the service returns a non-retryable error, Receive returns that error after all of the outstanding calls to f have returned. If ctx is done, Receive returns nil after all of the outstanding calls to f have returned and all messages have been acknowledged or have expired.
Receive calls f concurrently from multiple goroutines. It is encouraged to process messages synchronously in f, even if that processing is relatively time-consuming; Receive will spawn new goroutines for incoming messages, limited by MaxOutstandingMessages and MaxOutstandingBytes in ReceiveSettings.
The context passed to f will be canceled when ctx is Done or there is a fatal service error.
Receive will send an ack deadline extension on message receipt, then automatically extend the ack deadline of all fetched Messages up to the period specified by s.ReceiveSettings.MaxExtension.
Each Subscription may have only one invocation of Receive active at a time.
Example ¶
package main import ( "context" "errors" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } sub := client.Subscription("subName") err = sub.Receive(ctx, func(ctx context.Context, m *pubsub.Message) { // TODO: Handle message. // NOTE: May be called concurrently; synchronize access to shared memory. m.Ack() }) if err != nil && !errors.Is(err, context.Canceled) { // TODO: Handle error. } }
Output:
Example (MaxExtension) ¶
This example shows how to configure keepalive so that unacknoweldged messages expire quickly, allowing other subscribers to take them.
package main import ( "context" "errors" "time" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } sub := client.Subscription("subName") // This program is expected to process and acknowledge messages in 30 seconds. If // not, the Pub/Sub API will assume the message is not acknowledged. sub.ReceiveSettings.MaxExtension = 30 * time.Second err = sub.Receive(ctx, func(ctx context.Context, m *pubsub.Message) { // TODO: Handle message. m.Ack() }) if err != nil && !errors.Is(err, context.Canceled) { // TODO: Handle error. } }
Output:
Example (MaxOutstanding) ¶
This example shows how to throttle Subscription.Receive, which aims for high throughput by default. By limiting the number of messages and/or bytes being processed at once, you can bound your program's resource consumption.
package main import ( "context" "errors" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } sub := client.Subscription("subName") sub.ReceiveSettings.MaxOutstandingMessages = 5 sub.ReceiveSettings.MaxOutstandingBytes = 10e6 err = sub.Receive(ctx, func(ctx context.Context, m *pubsub.Message) { // TODO: Handle message. m.Ack() }) if err != nil && !errors.Is(err, context.Canceled) { // TODO: Handle error. } }
Output:
func (*Subscription) SeekToSnapshot ¶
func (s *Subscription) SeekToSnapshot(ctx context.Context, snap *Snapshot) error
SeekToSnapshot seeks the subscription to a snapshot.
The snapshot need not be created from this subscription, but it must be for the topic this subscription is subscribed to.
Example ¶
package main import ( "context" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } sub := client.Subscription("subName") snap := client.Snapshot("snapshotName") if err := sub.SeekToSnapshot(ctx, snap); err != nil { // TODO: Handle error. } }
Output:
func (*Subscription) SeekToTime ¶
SeekToTime seeks the subscription to a point in time.
Messages retained in the subscription that were published before this time are marked as acknowledged, and messages retained in the subscription that were published after this time are marked as unacknowledged. Note that this operation affects only those messages retained in the subscription (configured by SnapshotConfig). For example, if `time` corresponds to a point before the message retention window (or to a point before the system's notion of the subscription creation time), only retained messages will be marked as unacknowledged, and already-expunged messages will not be restored.
Example ¶
package main import ( "context" "time" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } sub := client.Subscription("subName") if err := sub.SeekToTime(ctx, time.Now().Add(-time.Hour)); err != nil { // TODO: Handle error. } }
Output:
func (*Subscription) String ¶
func (s *Subscription) String() string
String returns the globally unique printable name of the subscription.
func (*Subscription) Update ¶
func (s *Subscription) Update(ctx context.Context, cfg SubscriptionConfigToUpdate) (SubscriptionConfig, error)
Update changes an existing subscription according to the fields set in cfg. It returns the new SubscriptionConfig.
Update returns an error if no fields were modified.
Example ��
package main import ( "context" "time" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } sub := client.Subscription("subName") subConfig, err := sub.Update(ctx, pubsub.SubscriptionConfigToUpdate{ PushConfig: &pubsub.PushConfig{Endpoint: "https://example.com/push"}, // Make the subscription never expire. ExpirationPolicy: time.Duration(0), }) if err != nil { // TODO: Handle error. } _ = subConfig // TODO: Use SubscriptionConfig. }
Output:
Example (PushConfigAuthenticationMethod) ¶
package main import ( "context" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } sub := client.Subscription("subName") subConfig, err := sub.Update(ctx, pubsub.SubscriptionConfigToUpdate{ PushConfig: &pubsub.PushConfig{ Endpoint: "https://example.com/push", AuthenticationMethod: &pubsub.OIDCToken{ ServiceAccountEmail: "service-account-email", Audience: "client-12345", }, }, }) if err != nil { // TODO: Handle error. } _ = subConfig // TODO: Use SubscriptionConfig. }
Output:
type SubscriptionConfig ¶
type SubscriptionConfig struct { // The topic from which this subscription is receiving messages. Topic *Topic // If push delivery is used with this subscription, this field is // used to configure it. At most one of `PushConfig`, `BigQueryConfig`, // or `CloudStorageConfig` can be set. If all are empty, then the // subscriber will pull and ack messages using API methods. PushConfig PushConfig // If delivery to BigQuery is used with this subscription, this field is // used to configure it. At most one of `PushConfig`, `BigQueryConfig`, // or `CloudStorageConfig` can be set. If all are empty, then the // subscriber will pull and ack messages using API methods. BigQueryConfig BigQueryConfig // If delivery to Cloud Storage is used with this subscription, this field is // used to configure it. At most one of `PushConfig`, `BigQueryConfig`, // or `CloudStorageConfig` can be set. If all are empty, then the // subscriber will pull and ack messages using API methods. CloudStorageConfig CloudStorageConfig // The default maximum time after a subscriber receives a message before // the subscriber should acknowledge the message. Note: messages which are // obtained via Subscription.Receive need not be acknowledged within this // deadline, as the deadline will be automatically extended. AckDeadline time.Duration // Whether to retain acknowledged messages. If true, acknowledged messages // will not be expunged until they fall out of the RetentionDuration window. RetainAckedMessages bool // How long to retain messages in backlog, from the time of publish. If // RetainAckedMessages is true, this duration affects the retention of // acknowledged messages, otherwise only unacknowledged messages are retained. // Defaults to 7 days. Cannot be longer than 7 days or shorter than 10 minutes. RetentionDuration time.Duration // Expiration policy specifies the conditions for a subscription's expiration. // A subscription is considered active as long as any connected subscriber is // successfully consuming messages from the subscription or is issuing // operations on the subscription. If `expiration_policy` is not set, a // *default policy* with `ttl` of 31 days will be used. The minimum allowed // value for `expiration_policy.ttl` is 1 day. // // Use time.Duration(0) to indicate that the subscription should never expire. ExpirationPolicy optional.Duration // The set of labels for the subscription. Labels map[string]string // EnableMessageOrdering enables message ordering on this subscription. // This value is only used for subscription creation and update, and // is not read locally in calls like Subscription.Receive(). // // If set to false, even if messages are published with ordering keys, // messages will not be delivered in order. // // When calling Subscription.Receive(), the client will check this // value with a call to Subscription.Config(), which requires the // roles/viewer or roles/pubsub.viewer role on your service account. // If that call fails, mesages with ordering keys will be delivered in order. EnableMessageOrdering bool // DeadLetterPolicy specifies the conditions for dead lettering messages in // a subscription. If not set, dead lettering is disabled. DeadLetterPolicy *DeadLetterPolicy // Filter is an expression written in the Cloud Pub/Sub filter language. If // non-empty, then only `PubsubMessage`s whose `attributes` field matches the // filter are delivered on this subscription. If empty, then no messages are // filtered out. Cannot be changed after the subscription is created. Filter string // RetryPolicy specifies how Cloud Pub/Sub retries message delivery. RetryPolicy *RetryPolicy // Detached indicates whether the subscription is detached from its topic. // Detached subscriptions don't receive messages from their topic and don't // retain any backlog. `Pull` and `StreamingPull` requests will return // FAILED_PRECONDITION. If the subscription is a push subscription, pushes to // the endpoint will not be made. Detached bool // TopicMessageRetentionDuration indicates the minimum duration for which a message is // retained after it is published to the subscription's topic. If this field is // set, messages published to the subscription's topic in the last // `TopicMessageRetentionDuration` are always available to subscribers. // You can enable both topic and subscription retention for the same topic. // In this situation, the maximum of the retention durations takes effect. // // This is an output only field, meaning it will only appear in responses from the backend // and will be ignored if sent in a request. TopicMessageRetentionDuration time.Duration // EnableExactlyOnceDelivery configures Pub/Sub to provide the following guarantees // for the delivery of a message with a given MessageID on this subscription: // // The message sent to a subscriber is guaranteed not to be resent // before the message's acknowledgement deadline expires. // An acknowledged message will not be resent to a subscriber. // // Note that subscribers may still receive multiple copies of a message // when `enable_exactly_once_delivery` is true if the message was published // multiple times by a publisher client. These copies are considered distinct // by Pub/Sub and have distinct MessageID values. // // Lastly, to guarantee messages have been acked or nacked properly, you must // call Message.AckWithResult() or Message.NackWithResult(). These return an // AckResult which will be ready if the message has been acked (or failed to be acked). EnableExactlyOnceDelivery bool // State indicates whether or not the subscription can receive messages. // This is an output-only field that indicates whether or not the subscription can // receive messages. This field is set only in responses from the server; // it is ignored if it is set in any requests. State SubscriptionState // contains filtered or unexported fields }
SubscriptionConfig describes the configuration of a subscription. If none of PushConfig, BigQueryConfig, or CloudStorageConfig is set, then the subscriber will pull and ack messages using API methods. At most one of these fields may be set.
func (*SubscriptionConfig) ID ¶ added in v1.17.1
func (s *SubscriptionConfig) ID() string
ID returns the unique identifier of the subscription within its project. This method only works when the subscription config is returned from the server, such as when calling `client.Subscription` or `client.Subscriptions`. Otherwise, this will return an empty string.
func (*SubscriptionConfig) String ¶ added in v1.17.1
func (s *SubscriptionConfig) String() string
String returns the globally unique printable name of the subscription config. This method only works when the subscription config is returned from the server, such as when calling `client.Subscription` or `client.Subscriptions`. Otherwise, this will return an empty string.
type SubscriptionConfigToUpdate ¶
type SubscriptionConfigToUpdate struct { // If non-nil, the push config is changed. At most one of PushConfig, BigQueryConfig, or CloudStorageConfig // can be set. // If currently in push mode, set this value to the zero value to revert to a Pull based subscription. PushConfig *PushConfig // If non-nil, the bigquery config is changed. At most one of PushConfig, BigQueryConfig, or CloudStorageConfig // can be set. // If currently in bigquery mode, set this value to the zero value to revert to a Pull based subscription, BigQueryConfig *BigQueryConfig // If non-nil, the Cloud Storage config is changed. At most one of PushConfig, BigQueryConfig, or CloudStorageConfig // can be set. // If currently in CloudStorage mode, set this value to the zero value to revert to a Pull based subscription, CloudStorageConfig *CloudStorageConfig // If non-zero, the ack deadline is changed. AckDeadline time.Duration // If set, RetainAckedMessages is changed. RetainAckedMessages optional.Bool // If non-zero, RetentionDuration is changed. RetentionDuration time.Duration // If non-zero, Expiration is changed. ExpirationPolicy optional.Duration // If non-nil, DeadLetterPolicy is changed. To remove dead lettering from // a subscription, use the zero value for this struct. DeadLetterPolicy *DeadLetterPolicy // If non-nil, the current set of labels is completely // replaced by the new set. // This field has beta status. It is not subject to the stability guarantee // and may change. Labels map[string]string // If non-nil, RetryPolicy is changed. To remove an existing retry policy // (to redeliver messages as soon as possible) use a pointer to the zero value // for this struct. RetryPolicy *RetryPolicy // If set, EnableExactlyOnce is changed. EnableExactlyOnceDelivery optional.Bool }
SubscriptionConfigToUpdate describes how to update a subscription.
type SubscriptionIterator ¶
type SubscriptionIterator struct {
// contains filtered or unexported fields
}
SubscriptionIterator is an iterator that returns a series of subscriptions.
func (*SubscriptionIterator) Next ¶
func (subs *SubscriptionIterator) Next() (*Subscription, error)
Next returns the next subscription. If there are no more subscriptions, iterator.Done will be returned.
Example ¶
package main import ( "context" "fmt" "cloud.google.com/go/pubsub" "google.golang.org/api/iterator" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } // List all subscriptions of the project. it := client.Subscriptions(ctx) for { sub, err := it.Next() if err == iterator.Done { break } if err != nil { // TODO: Handle error. } fmt.Println(sub) } }
Output:
func (*SubscriptionIterator) NextConfig ¶ added in v1.17.0
func (subs *SubscriptionIterator) NextConfig() (*SubscriptionConfig, error)
NextConfig returns the next subscription config. If there are no more subscriptions, iterator.Done will be returned. This call shares the underlying iterator with calls to `SubscriptionIterator.Next`. If you wish to use mix calls, create separate iterator instances for both.
type SubscriptionState ¶ added in v1.23.0
type SubscriptionState int
SubscriptionState denotes the possible states for a Subscription.
type Topic ¶
type Topic struct { // Settings for publishing messages. All changes must be made before the // first call to Publish. The default is DefaultPublishSettings. PublishSettings PublishSettings // EnableMessageOrdering enables delivery of ordered keys. EnableMessageOrdering bool // contains filtered or unexported fields }
Topic is a reference to a PubSub topic.
The methods of Topic are safe for use by multiple goroutines.
func (*Topic) Config ¶
func (t *Topic) Config(ctx context.Context) (TopicConfig, error)
Config returns the TopicConfig for the topic.
func (*Topic) Delete ¶
Delete deletes the topic.
Example ¶
package main import ( "context" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } topic := client.Topic("topicName") if err := topic.Delete(ctx); err != nil { // TODO: Handle error. } }
Output:
func (*Topic) Exists ¶
Exists reports whether the topic exists on the server.
Example ¶
package main import ( "context" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } topic := client.Topic("topicName") ok, err := topic.Exists(ctx) if err != nil { // TODO: Handle error. } if !ok { // Topic doesn't exist. } }
Output:
func (*Topic) Flush ¶ added in v1.11.0
func (t *Topic) Flush()
Flush blocks until all remaining messages are sent.
func (*Topic) Publish ¶
func (t *Topic) Publish(ctx context.Context, msg *Message) *PublishResult
Publish publishes msg to the topic asynchronously. Messages are batched and sent according to the topic's PublishSettings. Publish never blocks.
Publish returns a non-nil PublishResult which will be ready when the message has been sent (or has failed to be sent) to the server.
Publish creates goroutines for batching and sending messages. These goroutines need to be stopped by calling t.Stop(). Once stopped, future calls to Publish will immediately return a PublishResult with an error.
Example ¶
package main import ( "context" "fmt" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } topic := client.Topic("topicName") defer topic.Stop() var results []*pubsub.PublishResult r := topic.Publish(ctx, &pubsub.Message{ Data: []byte("hello world"), }) results = append(results, r) // Do other work ... for _, r := range results { id, err := r.Get(ctx) if err != nil { // TODO: Handle error. } fmt.Printf("Published a message with a message ID: %s\n", id) } }
Output:
func (*Topic) ResumePublish ¶ added in v1.4.0
ResumePublish resumes accepting messages for the provided ordering key. Publishing using an ordering key might be paused if an error is encountered while publishing, to prevent messages from being published out of order.
func (*Topic) Stop ¶
func (t *Topic) Stop()
Stop sends all remaining published messages and stop goroutines created for handling publishing. Returns once all outstanding messages have been sent or have failed to be sent.
func (*Topic) Subscriptions ¶
func (t *Topic) Subscriptions(ctx context.Context) *SubscriptionIterator
Subscriptions returns an iterator which returns the subscriptions for this topic.
Some of the returned subscriptions may belong to a project other than t.
Example ¶
package main import ( "context" "cloud.google.com/go/pubsub" "google.golang.org/api/iterator" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } topic := client.Topic("topic-name") // List all subscriptions of the topic (maybe of multiple projects). for subs := topic.Subscriptions(ctx); ; { sub, err := subs.Next() if err == iterator.Done { break } if err != nil { // TODO: Handle error. } _ = sub // TODO: use the subscription. } }
Output:
func (*Topic) Update ¶
func (t *Topic) Update(ctx context.Context, cfg TopicConfigToUpdate) (TopicConfig, error)
Update changes an existing topic according to the fields set in cfg. It returns the new TopicConfig.
Example ¶
package main import ( "context" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error.V } topic := client.Topic("topic-name") topicConfig, err := topic.Update(ctx, pubsub.TopicConfigToUpdate{ MessageStoragePolicy: &pubsub.MessageStoragePolicy{ AllowedPersistenceRegions: []string{ "asia-east1", "asia-northeast1", "asia-southeast1", "australia-southeast1", "europe-north1", "europe-west1", "europe-west2", "europe-west3", "europe-west4", "us-central1", "us-central2", "us-east1", "us-east4", "us-west1", "us-west2"}, }, }) if err != nil { // TODO: Handle error. } _ = topicConfig // TODO: Use TopicConfig }
Output:
Example (ResetMessageStoragePolicy) ¶
package main import ( "context" "cloud.google.com/go/pubsub" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error.V } topic := client.Topic("topic-name") topicConfig, err := topic.Update(ctx, pubsub.TopicConfigToUpdate{ // Just use a non-nil MessageStoragePolicy without any fields. MessageStoragePolicy: &pubsub.MessageStoragePolicy{}, }) if err != nil { // TODO: Handle error. } _ = topicConfig // TODO: Use TopicConfig }
Output:
type TopicConfig ¶
type TopicConfig struct { // The set of labels for the topic. Labels map[string]string // The topic's message storage policy. MessageStoragePolicy MessageStoragePolicy // The name of the Cloud KMS key to be used to protect access to messages // published to this topic, in the format // "projects/P/locations/L/keyRings/R/cryptoKeys/K". KMSKeyName string // Schema defines the schema settings upon topic creation. SchemaSettings *SchemaSettings // RetentionDuration configures the minimum duration to retain a message // after it is published to the topic. If this field is set, messages published // to the topic in the last `RetentionDuration` are always available to subscribers. // For instance, it allows any attached subscription to [seek to a // timestamp](https://cloud.google.com/pubsub/docs/replay-overview#seek_to_a_time) // that is up to `RetentionDuration` in the past. If this field is // not set, message retention is controlled by settings on individual // subscriptions. Cannot be more than 31 days or less than 10 minutes. // // For more information, see https://cloud.google.com/pubsub/docs/replay-overview#topic_message_retention. RetentionDuration optional.Duration // State is an output-only field indicating the state of the topic. State TopicState // IngestionDataSourceSettings are settings for ingestion from a // data source into this topic. IngestionDataSourceSettings *IngestionDataSourceSettings // contains filtered or unexported fields }
TopicConfig describes the configuration of a topic.
func (*TopicConfig) ID ¶ added in v1.17.1
func (t *TopicConfig) ID() string
ID returns the unique identifier of the topic within its project. This method only works when the topic config is returned from the server, such as when calling `client.Topic` or `client.Topics`. Otherwise, this will return an empty string.
func (*TopicConfig) String ¶ added in v1.17.1
func (t *TopicConfig) String() string
String returns the printable globally unique name for the topic config. This method only works when the topic config is returned from the server, such as when calling `client.Topic` or `client.Topics`. Otherwise, this will return an empty string.
type TopicConfigToUpdate ¶
type TopicConfigToUpdate struct { // If non-nil, the current set of labels is completely // replaced by the new set. Labels map[string]string // If non-nil, the existing policy (containing the list of regions) // is completely replaced by the new policy. // // Use the zero value &MessageStoragePolicy{} to reset the topic back to // using the organization's Resource Location Restriction policy. // // If nil, the policy remains unchanged. // // This field has beta status. It is not subject to the stability guarantee // and may change. MessageStoragePolicy *MessageStoragePolicy // If set to a positive duration between 10 minutes and 31 days, RetentionDuration is changed. // If set to a negative value, this clears RetentionDuration from the topic. // If nil, the retention duration remains unchanged. RetentionDuration optional.Duration // Schema defines the schema settings upon topic creation. // // Use the zero value &SchemaSettings{} to remove the schema from the topic. SchemaSettings *SchemaSettings // IngestionDataSourceSettings are settings for ingestion from a // data source into this topic. // // When changing this value, the entire data source settings object must be applied, // rather than just the differences. This includes the source and logging settings. // // Use the zero value &IngestionDataSourceSettings{} to remove the ingestion settings from the topic. IngestionDataSourceSettings *IngestionDataSourceSettings }
TopicConfigToUpdate describes how to update a topic.
type TopicIterator ¶
type TopicIterator struct {
// contains filtered or unexported fields
}
TopicIterator is an iterator that returns a series of topics.
func (*TopicIterator) Next ¶
func (tps *TopicIterator) Next() (*Topic, error)
Next returns the next topic. If there are no more topics, iterator.Done will be returned.
Example ¶
package main import ( "context" "fmt" "cloud.google.com/go/pubsub" "google.golang.org/api/iterator" ) func main() { ctx := context.Background() client, err := pubsub.NewClient(ctx, "project-id") if err != nil { // TODO: Handle error. } // List all topics. it := client.Topics(ctx) for { t, err := it.Next() if err == iterator.Done { break } if err != nil { // TODO: Handle error. } fmt.Println(t) } }
Output:
func (*TopicIterator) NextConfig ¶ added in v1.17.0
func (t *TopicIterator) NextConfig() (*TopicConfig, error)
NextConfig returns the next topic config. If there are no more topics, iterator.Done will be returned. This call shares the underlying iterator with calls to `TopicIterator.Next`. If you wish to use mix calls, create separate iterator instances for both.
type TopicState ¶ added in v1.37.0
type TopicState int
TopicState denotes the possible states for a topic.
type ValidateMessageResult ¶ added in v1.15.0
type ValidateMessageResult struct{}
ValidateMessageResult is the response for the ValidateMessage method. Reserved for future use.
type ValidateSchemaResult ¶ added in v1.15.0
type ValidateSchemaResult struct{}
ValidateSchemaResult is the response for the ValidateSchema method. Reserved for future use.
Source Files ¶
Directories ¶
Path | Synopsis |
---|---|
Package pubsub is an auto-generated package for the Cloud Pub/Sub API.
|
Package pubsub is an auto-generated package for the Cloud Pub/Sub API. |
benchwrapper
Package main wraps the client library in a gRPC interface that a benchmarker can communicate through.
|
Package main wraps the client library in a gRPC interface that a benchmarker can communicate through. |
Package loadtest implements load testing for pubsub, following the interface defined in https://github.com/GoogleCloudPlatform/pubsub/tree/master/load-test-framework/ .
|
Package loadtest implements load testing for pubsub, following the interface defined in https://github.com/GoogleCloudPlatform/pubsub/tree/master/load-test-framework/ . |
Package pstest provides a fake Cloud PubSub service for testing.
|
Package pstest provides a fake Cloud PubSub service for testing. |