f3c80c4eef
Protobuf Refactoring for Multi-Module Cleanliness This commit includes the following: Moves all packages that were within proto/ to proto/private Rewrites imports to account for the packages being moved Adds in buf.work.yaml to enable buf workspaces Names the proto-public buf module so that we can override the Go package imports within proto/buf.yaml Bumps the buf version dependency to 1.14.0 (I was trying out the version to see if it would get around an issue - it didn't but it also doesn't break things and it seemed best to keep up with the toolchain changes) Why: In the future we will need to consume other protobuf dependencies such as the Google HTTP annotations for openapi generation or grpc-gateway usage. There were some recent changes to have our own ratelimiting annotations. The two combined were not working when I was trying to use them together (attempting to rebase another branch) Buf workspaces should be the solution to the problem Buf workspaces means that each module will have generated Go code that embeds proto file names relative to the proto dir and not the top level repo root. This resulted in proto file name conflicts in the Go global protobuf type registry. The solution to that was to add in a private/ directory into the path within the proto/ directory. That then required rewriting all the imports. Is this safe? AFAICT yes The gRPC wire protocol doesn't seem to care about the proto file names (although the Go grpc code does tack on the proto file name as Metadata in the ServiceDesc) Other than imports, there were no changes to any generated code as a result of this.
108 lines
3.1 KiB
Go
108 lines
3.1 KiB
Go
package cachetype
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"strconv"
|
|
"time"
|
|
|
|
external "github.com/hashicorp/consul/agent/grpc-external"
|
|
"github.com/hashicorp/consul/proto/private/pbpeering"
|
|
"github.com/mitchellh/hashstructure"
|
|
"google.golang.org/grpc"
|
|
"google.golang.org/grpc/codes"
|
|
"google.golang.org/grpc/status"
|
|
|
|
"github.com/hashicorp/consul/agent/cache"
|
|
"github.com/hashicorp/consul/agent/structs"
|
|
)
|
|
|
|
// PeeringListName is the recommended name for registration.
|
|
const PeeringListName = "peers"
|
|
|
|
type PeeringListRequest struct {
|
|
Request *pbpeering.PeeringListRequest
|
|
structs.QueryOptions
|
|
}
|
|
|
|
func (r *PeeringListRequest) CacheInfo() cache.RequestInfo {
|
|
info := cache.RequestInfo{
|
|
Token: r.Token,
|
|
Datacenter: "",
|
|
MinIndex: 0,
|
|
Timeout: 0,
|
|
MustRevalidate: false,
|
|
|
|
// OPTIMIZE(peering): Cache.notifyPollingQuery polls at this interval. We need to revisit how that polling works.
|
|
// Using an exponential backoff when the result hasn't changed may be preferable.
|
|
MaxAge: 1 * time.Second,
|
|
}
|
|
|
|
v, err := hashstructure.Hash([]interface{}{
|
|
r.Request.Partition,
|
|
}, nil)
|
|
if err == nil {
|
|
// If there is an error, we don't set the key. A blank key forces
|
|
// no cache for this request so the request is forwarded directly
|
|
// to the server.
|
|
info.Key = strconv.FormatUint(v, 10)
|
|
}
|
|
|
|
return info
|
|
}
|
|
|
|
// Peerings supports fetching the list of peers for a given partition or wildcard-specifier.
|
|
type Peerings struct {
|
|
RegisterOptionsNoRefresh
|
|
Client PeeringLister
|
|
}
|
|
|
|
//go:generate mockery --name PeeringLister --inpackage --filename mock_PeeringLister_test.go
|
|
type PeeringLister interface {
|
|
PeeringList(
|
|
ctx context.Context, in *pbpeering.PeeringListRequest, opts ...grpc.CallOption,
|
|
) (*pbpeering.PeeringListResponse, error)
|
|
}
|
|
|
|
func (t *Peerings) Fetch(_ cache.FetchOptions, req cache.Request) (cache.FetchResult, error) {
|
|
var result cache.FetchResult
|
|
|
|
// The request should be a PeeringListRequest.
|
|
// We do not need to make a copy of this request type like in other cache types
|
|
// because the RequestInfo is synthetic.
|
|
reqReal, ok := req.(*PeeringListRequest)
|
|
if !ok {
|
|
return result, fmt.Errorf(
|
|
"Internal cache failure: request wrong type: %T", req)
|
|
}
|
|
|
|
// Always allow stale - there's no point in hitting leader if the request is
|
|
// going to be served from cache and end up arbitrarily stale anyway. This
|
|
// allows cached service-discover to automatically read scale across all
|
|
// servers too.
|
|
reqReal.QueryOptions.SetAllowStale(true)
|
|
|
|
ctx, err := external.ContextWithQueryOptions(context.Background(), reqReal.QueryOptions)
|
|
if err != nil {
|
|
return result, err
|
|
}
|
|
|
|
// Fetch
|
|
reply, err := t.Client.PeeringList(ctx, reqReal.Request)
|
|
if err != nil {
|
|
// Return an empty result if the error is due to peering being disabled.
|
|
// This allows mesh gateways to receive an update and confirm that the watch is set.
|
|
if e, ok := status.FromError(err); ok && e.Code() == codes.FailedPrecondition {
|
|
result.Index = 1
|
|
result.Value = &pbpeering.PeeringListResponse{}
|
|
return result, nil
|
|
}
|
|
return result, err
|
|
}
|
|
|
|
result.Value = reply
|
|
result.Index = reply.Index
|
|
|
|
return result, nil
|
|
}
|