open-vault/physical/s3/s3.go

323 lines
7.4 KiB
Go
Raw Normal View History

package s3
2015-05-20 14:54:26 +00:00
import (
"bytes"
"context"
2015-05-20 14:54:26 +00:00
"fmt"
"io"
"net/http"
2015-05-20 14:54:26 +00:00
"os"
"path"
2015-05-20 14:54:26 +00:00
"sort"
2017-03-26 18:32:26 +00:00
"strconv"
2015-05-20 14:54:26 +00:00
"strings"
"time"
"github.com/armon/go-metrics"
2015-06-03 19:02:49 +00:00
"github.com/aws/aws-sdk-go/aws"
"github.com/aws/aws-sdk-go/aws/awserr"
2015-10-30 22:22:48 +00:00
"github.com/aws/aws-sdk-go/aws/session"
2015-06-03 19:02:49 +00:00
"github.com/aws/aws-sdk-go/service/s3"
"github.com/hashicorp/go-cleanhttp"
log "github.com/hashicorp/go-hclog"
"github.com/hashicorp/go-secure-stdlib/awsutil"
"github.com/hashicorp/go-secure-stdlib/parseutil"
"github.com/hashicorp/vault/sdk/helper/consts"
"github.com/hashicorp/vault/sdk/physical"
2015-05-20 14:54:26 +00:00
)
// Verify S3Backend satisfies the correct interfaces
var _ physical.Backend = (*S3Backend)(nil)
2015-05-20 14:54:26 +00:00
// S3Backend is a physical backend that stores data
// within an S3 bucket.
type S3Backend struct {
bucket string
path string
kmsKeyId string
client *s3.S3
logger log.Logger
permitPool *physical.PermitPool
2015-05-20 14:54:26 +00:00
}
// NewS3Backend constructs a S3 backend using a pre-existing
2015-05-20 14:54:26 +00:00
// bucket. Credentials can be provided to the backend, sourced
// from the environment, AWS credential files or by IAM role.
func NewS3Backend(conf map[string]string, logger log.Logger) (physical.Backend, error) {
bucket := os.Getenv("AWS_S3_BUCKET")
if bucket == "" {
bucket = conf["bucket"]
if bucket == "" {
return nil, fmt.Errorf("'bucket' must be set")
}
2015-05-20 14:54:26 +00:00
}
path := conf["path"]
accessKey, ok := conf["access_key"]
2015-05-20 14:54:26 +00:00
if !ok {
accessKey = ""
2015-05-20 14:54:26 +00:00
}
secretKey, ok := conf["secret_key"]
2015-05-20 14:54:26 +00:00
if !ok {
secretKey = ""
2015-05-20 14:54:26 +00:00
}
sessionToken, ok := conf["session_token"]
if !ok {
sessionToken = ""
}
endpoint := os.Getenv("AWS_S3_ENDPOINT")
if endpoint == "" {
endpoint = conf["endpoint"]
}
region := os.Getenv("AWS_REGION")
if region == "" {
region = os.Getenv("AWS_DEFAULT_REGION")
2015-05-20 14:54:26 +00:00
if region == "" {
region = conf["region"]
if region == "" {
region = "us-east-1"
}
2015-05-20 14:54:26 +00:00
}
}
s3ForcePathStyleStr, ok := conf["s3_force_path_style"]
if !ok {
s3ForcePathStyleStr = "false"
}
s3ForcePathStyleBool, err := parseutil.ParseBool(s3ForcePathStyleStr)
if err != nil {
return nil, fmt.Errorf("invalid boolean set for s3_force_path_style: %q", s3ForcePathStyleStr)
}
disableSSLStr, ok := conf["disable_ssl"]
if !ok {
disableSSLStr = "false"
}
disableSSLBool, err := parseutil.ParseBool(disableSSLStr)
if err != nil {
return nil, fmt.Errorf("invalid boolean set for disable_ssl: %q", disableSSLStr)
}
2015-05-20 14:54:26 +00:00
credsConfig := &awsutil.CredentialsConfig{
AccessKey: accessKey,
SecretKey: secretKey,
SessionToken: sessionToken,
Logger: logger,
}
creds, err := credsConfig.GenerateCredentialChain()
if err != nil {
return nil, err
}
2015-05-20 14:54:26 +00:00
pooledTransport := cleanhttp.DefaultPooledTransport()
pooledTransport.MaxIdleConnsPerHost = consts.ExpirationRestoreWorkerCount
sess, err := session.NewSession(&aws.Config{
2015-05-20 14:54:26 +00:00
Credentials: creds,
HTTPClient: &http.Client{
Transport: pooledTransport,
},
Endpoint: aws.String(endpoint),
Region: aws.String(region),
S3ForcePathStyle: aws.Bool(s3ForcePathStyleBool),
DisableSSL: aws.Bool(disableSSLBool),
})
if err != nil {
return nil, err
}
s3conn := s3.New(sess)
2015-05-20 14:54:26 +00:00
_, err = s3conn.ListObjects(&s3.ListObjectsInput{Bucket: &bucket})
2015-05-20 14:54:26 +00:00
if err != nil {
return nil, fmt.Errorf("unable to access bucket %q in region %q: %w", bucket, region, err)
2015-05-20 14:54:26 +00:00
}
2017-03-26 18:32:26 +00:00
maxParStr, ok := conf["max_parallel"]
var maxParInt int
if ok {
maxParInt, err = strconv.Atoi(maxParStr)
if err != nil {
return nil, fmt.Errorf("failed parsing max_parallel parameter: %w", err)
2017-03-26 18:32:26 +00:00
}
if logger.IsDebug() {
logger.Debug("max_parallel set", "max_parallel", maxParInt)
2017-03-26 18:32:26 +00:00
}
}
kmsKeyId, ok := conf["kms_key_id"]
if !ok {
kmsKeyId = ""
}
2019-02-08 14:12:55 +00:00
2015-05-20 14:54:26 +00:00
s := &S3Backend{
client: s3conn,
bucket: bucket,
path: path,
kmsKeyId: kmsKeyId,
logger: logger,
permitPool: physical.NewPermitPool(maxParInt),
2015-05-20 14:54:26 +00:00
}
return s, nil
}
// Put is used to insert or update an entry
func (s *S3Backend) Put(ctx context.Context, entry *physical.Entry) error {
2015-05-20 14:54:26 +00:00
defer metrics.MeasureSince([]string{"s3", "put"}, time.Now())
2017-03-26 18:32:26 +00:00
s.permitPool.Acquire()
defer s.permitPool.Release()
// Setup key
key := path.Join(s.path, entry.Key)
putObjectInput := &s3.PutObjectInput{
2015-05-20 14:54:26 +00:00
Bucket: aws.String(s.bucket),
Key: aws.String(key),
2015-05-20 14:54:26 +00:00
Body: bytes.NewReader(entry.Value),
}
2015-05-20 14:54:26 +00:00
if s.kmsKeyId != "" {
putObjectInput.ServerSideEncryption = aws.String("aws:kms")
putObjectInput.SSEKMSKeyId = aws.String(s.kmsKeyId)
}
2019-02-08 14:12:55 +00:00
_, err := s.client.PutObject(putObjectInput)
2015-05-20 14:54:26 +00:00
if err != nil {
return err
}
return nil
}
// Get is used to fetch an entry
func (s *S3Backend) Get(ctx context.Context, key string) (*physical.Entry, error) {
2015-05-20 14:54:26 +00:00
defer metrics.MeasureSince([]string{"s3", "get"}, time.Now())
2017-03-26 18:32:26 +00:00
s.permitPool.Acquire()
defer s.permitPool.Release()
// Setup key
key = path.Join(s.path, key)
2015-05-20 14:54:26 +00:00
resp, err := s.client.GetObject(&s3.GetObjectInput{
Bucket: aws.String(s.bucket),
Key: aws.String(key),
})
if resp != nil && resp.Body != nil {
defer resp.Body.Close()
}
if awsErr, ok := err.(awserr.RequestFailure); ok {
2015-05-20 14:54:26 +00:00
// Return nil on 404s, error on anything else
if awsErr.StatusCode() == 404 {
2015-05-20 14:54:26 +00:00
return nil, nil
}
2017-03-26 18:32:26 +00:00
return nil, err
2015-05-20 14:54:26 +00:00
}
if err != nil {
return nil, err
}
if resp == nil {
return nil, fmt.Errorf("got nil response from S3 but no error")
}
2015-05-20 14:54:26 +00:00
data := bytes.NewBuffer(nil)
if resp.ContentLength != nil {
data = bytes.NewBuffer(make([]byte, 0, *resp.ContentLength))
}
_, err = io.Copy(data, resp.Body)
2015-05-20 14:54:26 +00:00
if err != nil {
return nil, err
}
// Strip path prefix
if s.path != "" {
key = strings.TrimPrefix(key, s.path+"/")
}
ent := &physical.Entry{
2015-05-20 14:54:26 +00:00
Key: key,
Value: data.Bytes(),
2015-05-20 14:54:26 +00:00
}
return ent, nil
}
// Delete is used to permanently delete an entry
func (s *S3Backend) Delete(ctx context.Context, key string) error {
2015-05-20 14:54:26 +00:00
defer metrics.MeasureSince([]string{"s3", "delete"}, time.Now())
2017-03-26 18:32:26 +00:00
s.permitPool.Acquire()
defer s.permitPool.Release()
// Setup key
key = path.Join(s.path, key)
_, err := s.client.DeleteObject(&s3.DeleteObjectInput{
2015-05-20 14:54:26 +00:00
Bucket: aws.String(s.bucket),
Key: aws.String(key),
2015-05-20 14:54:26 +00:00
})
if err != nil {
return err
}
return nil
}
// List is used to list all the keys under a given
// prefix, up to the next prefix.
func (s *S3Backend) List(ctx context.Context, prefix string) ([]string, error) {
2015-05-20 14:54:26 +00:00
defer metrics.MeasureSince([]string{"s3", "list"}, time.Now())
2017-03-26 18:32:26 +00:00
s.permitPool.Acquire()
defer s.permitPool.Release()
// Setup prefix
prefix = path.Join(s.path, prefix)
// Validate prefix (if present) is ending with a "/"
if prefix != "" && !strings.HasSuffix(prefix, "/") {
prefix += "/"
}
params := &s3.ListObjectsV2Input{
2017-06-16 15:09:15 +00:00
Bucket: aws.String(s.bucket),
Prefix: aws.String(prefix),
Delimiter: aws.String("/"),
}
2015-05-20 14:54:26 +00:00
keys := []string{}
err := s.client.ListObjectsV2Pages(params,
func(page *s3.ListObjectsV2Output, lastPage bool) bool {
if page != nil {
2017-06-16 15:09:15 +00:00
// Add truncated 'folder' paths
for _, commonPrefix := range page.CommonPrefixes {
// Avoid panic
if commonPrefix == nil {
continue
}
commonPrefix := strings.TrimPrefix(*commonPrefix.Prefix, prefix)
keys = append(keys, commonPrefix)
}
// Add objects only from the current 'folder'
for _, key := range page.Contents {
// Avoid panic
if key == nil {
continue
}
key := strings.TrimPrefix(*key.Key, prefix)
2017-06-16 15:09:15 +00:00
keys = append(keys, key)
}
}
return true
})
if err != nil {
return nil, err
2015-05-20 14:54:26 +00:00
}
sort.Strings(keys)
return keys, nil
}