6d368b5eed
When deleting a peering we do not want to delete the peering and all imported data in a single operation, since deleting a large amount of data at once could overload Consul. Instead we defer deletion of peerings so that: 1. When a peering deletion request is received via gRPC the peering is marked for deletion by setting the DeletedAt field. 2. A leader routine will monitor for peerings that are marked for deletion and kick off a throttled deletion of all imported resources before deleting the peering itself. This commit mostly addresses point #1 by modifying the peering service to mark peerings for deletion. Another key change is to add a PeeringListDeleted state store function which can return all peerings marked for deletion. This function is what will be watched by the deferred deletion leader routine. |
||
---|---|---|
.. | ||
watch | ||
.golangci.yml | ||
acl.go | ||
acl_test.go | ||
agent.go | ||
agent_test.go | ||
api.go | ||
api_test.go | ||
catalog.go | ||
catalog_test.go | ||
config_entry.go | ||
config_entry_discoverychain.go | ||
config_entry_discoverychain_test.go | ||
config_entry_exports.go | ||
config_entry_exports_test.go | ||
config_entry_gateways.go | ||
config_entry_gateways_test.go | ||
config_entry_intentions.go | ||
config_entry_intentions_test.go | ||
config_entry_mesh.go | ||
config_entry_test.go | ||
connect.go | ||
connect_ca.go | ||
connect_ca_test.go | ||
connect_intention.go | ||
connect_intention_test.go | ||
coordinate.go | ||
coordinate_test.go | ||
debug.go | ||
debug_test.go | ||
discovery_chain.go | ||
discovery_chain_test.go | ||
event.go | ||
event_test.go | ||
go.mod | ||
go.sum | ||
health.go | ||
health_test.go | ||
kv.go | ||
kv_test.go | ||
lock.go | ||
lock_test.go | ||
mock_api_test.go | ||
namespace.go | ||
namespace_test.go | ||
operator.go | ||
operator_area.go | ||
operator_autopilot.go | ||
operator_autopilot_test.go | ||
operator_keyring.go | ||
operator_keyring_test.go | ||
operator_license.go | ||
operator_raft.go | ||
operator_raft_test.go | ||
operator_segment.go | ||
oss.go | ||
partition.go | ||
peering.go | ||
peering_test.go | ||
prepared_query.go | ||
prepared_query_test.go | ||
raw.go | ||
README.md | ||
semaphore.go | ||
semaphore_test.go | ||
session.go | ||
session_test.go | ||
snapshot.go | ||
snapshot_test.go | ||
status.go | ||
status_test.go | ||
txn.go | ||
txn_test.go |
Consul API client
This package provides the api
package which attempts to
provide programmatic access to the full Consul API.
Currently, all of the Consul APIs included in version 0.6.0 are supported.
Documentation
The full documentation is available on Godoc
Usage
Below is an example of using the Consul client:
package main
import "github.com/hashicorp/consul/api"
import "fmt"
func main() {
// Get a new client
client, err := api.NewClient(api.DefaultConfig())
if err != nil {
panic(err)
}
// Get a handle to the KV API
kv := client.KV()
// PUT a new KV pair
p := &api.KVPair{Key: "REDIS_MAXCLIENTS", Value: []byte("1000")}
_, err = kv.Put(p, nil)
if err != nil {
panic(err)
}
// Lookup the pair
pair, _, err := kv.Get("REDIS_MAXCLIENTS", nil)
if err != nil {
panic(err)
}
fmt.Printf("KV: %v %s\n", pair.Key, pair.Value)
}
To run this example, start a Consul server:
consul agent -dev
Copy the code above into a file such as main.go
.
Install and run. You'll see a key (REDIS_MAXCLIENTS
) and value (1000
) printed.
$ go get
$ go run main.go
KV: REDIS_MAXCLIENTS 1000
After running the code, you can also view the values in the Consul UI on your local machine at http://localhost:8500/ui/dc1/kv