Implemented GCS bucket objects
This commit is contained in:
parent
455ba7bdb1
commit
602acadde8
|
@ -6,6 +6,7 @@ FEATURES:
|
|||
* **New resource: `cloudstack_loadbalancer_rule`** [GH-2934]
|
||||
* **New resource: `google_compute_project_metadata`** [GH-3065]
|
||||
* **New resources: `aws_ami`, `aws_ami_copy`, `aws_ami_from_instance`** [GH-2874]
|
||||
* **New resource: `google_storage_bucket_object`** [GH-3192]
|
||||
|
||||
IMPROVEMENTS:
|
||||
|
||||
|
|
|
@ -52,6 +52,7 @@ func Provider() terraform.ResourceProvider {
|
|||
"google_dns_record_set": resourceDnsRecordSet(),
|
||||
"google_compute_instance_group_manager": resourceComputeInstanceGroupManager(),
|
||||
"google_storage_bucket": resourceStorageBucket(),
|
||||
"google_storage_bucket_object": resourceStorageBucketObject(),
|
||||
},
|
||||
|
||||
ConfigureFunc: providerConfigure,
|
||||
|
|
|
@ -0,0 +1,132 @@
|
|||
package google
|
||||
|
||||
import (
|
||||
"os"
|
||||
"fmt"
|
||||
|
||||
"github.com/hashicorp/terraform/helper/schema"
|
||||
|
||||
"google.golang.org/api/storage/v1"
|
||||
)
|
||||
|
||||
func resourceStorageBucketObject() *schema.Resource {
|
||||
return &schema.Resource{
|
||||
Create: resourceStorageBucketObjectCreate,
|
||||
Read: resourceStorageBucketObjectRead,
|
||||
Update: resourceStorageBucketObjectUpdate,
|
||||
Delete: resourceStorageBucketObjectDelete,
|
||||
|
||||
Schema: map[string]*schema.Schema{
|
||||
"bucket": &schema.Schema{
|
||||
Type: schema.TypeString,
|
||||
Required: true,
|
||||
ForceNew: true,
|
||||
},
|
||||
"name": &schema.Schema{
|
||||
Type: schema.TypeString,
|
||||
Required: true,
|
||||
ForceNew: true,
|
||||
},
|
||||
"source": &schema.Schema{
|
||||
Type: schema.TypeString,
|
||||
Required: true,
|
||||
ForceNew: true,
|
||||
},
|
||||
"predefined_acl": &schema.Schema{
|
||||
Type: schema.TypeString,
|
||||
Default: "projectPrivate",
|
||||
Optional: true,
|
||||
ForceNew: true,
|
||||
},
|
||||
"md5hash": &schema.Schema{
|
||||
Type: schema.TypeString,
|
||||
Computed: true,
|
||||
},
|
||||
"crc32c": &schema.Schema{
|
||||
Type: schema.TypeString,
|
||||
Computed: true,
|
||||
},
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
func objectGetId(object *storage.Object) string {
|
||||
return object.Bucket + "-" + object.Name
|
||||
}
|
||||
|
||||
func resourceStorageBucketObjectCreate(d *schema.ResourceData, meta interface{}) error {
|
||||
config := meta.(*Config)
|
||||
|
||||
bucket := d.Get("bucket").(string)
|
||||
name := d.Get("name").(string)
|
||||
source := d.Get("source").(string)
|
||||
acl := d.Get("predefined_acl").(string)
|
||||
|
||||
file, err := os.Open(source)
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error opening %s: %s", source, err)
|
||||
}
|
||||
|
||||
objectsService := storage.NewObjectsService(config.clientStorage)
|
||||
object := &storage.Object{Bucket: bucket}
|
||||
|
||||
insertCall := objectsService.Insert(bucket, object)
|
||||
insertCall.Name(name)
|
||||
insertCall.Media(file)
|
||||
insertCall.PredefinedAcl(acl)
|
||||
|
||||
_, err = insertCall.Do()
|
||||
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error uploading contents of object %s from %s: %s", name, source, err)
|
||||
}
|
||||
|
||||
return resourceStorageBucketObjectRead(d, meta)
|
||||
}
|
||||
|
||||
func resourceStorageBucketObjectRead(d *schema.ResourceData, meta interface{}) error {
|
||||
config := meta.(*Config)
|
||||
|
||||
bucket := d.Get("bucket").(string)
|
||||
name := d.Get("name").(string)
|
||||
|
||||
objectsService := storage.NewObjectsService(config.clientStorage)
|
||||
getCall := objectsService.Get(bucket, name)
|
||||
|
||||
res, err := getCall.Do()
|
||||
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error retrieving contents of object %s: %s", name, err)
|
||||
}
|
||||
|
||||
d.Set("md5hash", res.Md5Hash)
|
||||
d.Set("crc32c", res.Crc32c)
|
||||
|
||||
d.SetId(objectGetId(res))
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
func resourceStorageBucketObjectUpdate(d *schema.ResourceData, meta interface{}) error {
|
||||
// The Cloud storage API doesn't support updating object data contents,
|
||||
// only metadata. So once we implement metadata we'll have work to do here
|
||||
return nil
|
||||
}
|
||||
|
||||
func resourceStorageBucketObjectDelete(d *schema.ResourceData, meta interface{}) error {
|
||||
config := meta.(*Config)
|
||||
|
||||
bucket := d.Get("bucket").(string)
|
||||
name := d.Get("name").(string)
|
||||
|
||||
objectsService := storage.NewObjectsService(config.clientStorage)
|
||||
|
||||
DeleteCall := objectsService.Delete(bucket, name)
|
||||
err := DeleteCall.Do()
|
||||
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error deleting contents of object %s: %s", name, err)
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
|
@ -0,0 +1,102 @@
|
|||
package google
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"testing"
|
||||
"io/ioutil"
|
||||
"crypto/md5"
|
||||
"encoding/base64"
|
||||
|
||||
"github.com/hashicorp/terraform/helper/resource"
|
||||
"github.com/hashicorp/terraform/terraform"
|
||||
|
||||
"google.golang.org/api/storage/v1"
|
||||
)
|
||||
|
||||
var tf, err = ioutil.TempFile("", "tf-gce-test")
|
||||
var bucketName = "tf-gce-bucket-test"
|
||||
var objectName = "tf-gce-test"
|
||||
|
||||
func TestAccGoogleStorageObject_basic(t *testing.T) {
|
||||
data := []byte("data data data")
|
||||
h := md5.New()
|
||||
h.Write(data)
|
||||
data_md5 := base64.StdEncoding.EncodeToString(h.Sum(nil))
|
||||
|
||||
ioutil.WriteFile(tf.Name(), data, 0644)
|
||||
resource.Test(t, resource.TestCase{
|
||||
PreCheck: func() {
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
testAccPreCheck(t)
|
||||
},
|
||||
Providers: testAccProviders,
|
||||
CheckDestroy: testAccGoogleStorageObjectDestroy,
|
||||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
Config: testGoogleStorageBucketsObjectBasic,
|
||||
Check: testAccCheckGoogleStorageObject(bucketName, objectName, data_md5),
|
||||
},
|
||||
},
|
||||
})
|
||||
}
|
||||
|
||||
func testAccCheckGoogleStorageObject(bucket, object, md5 string) resource.TestCheckFunc {
|
||||
return func(s *terraform.State) error {
|
||||
config := testAccProvider.Meta().(*Config)
|
||||
|
||||
objectsService := storage.NewObjectsService(config.clientStorage)
|
||||
|
||||
|
||||
getCall := objectsService.Get(bucket, object)
|
||||
res, err := getCall.Do()
|
||||
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error retrieving contents of object %s: %s", object, err)
|
||||
}
|
||||
|
||||
if (md5 != res.Md5Hash) {
|
||||
return fmt.Errorf("Error contents of %s garbled, md5 hashes don't match (%s, %s)", object, md5, res.Md5Hash)
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
func testAccGoogleStorageObjectDestroy(s *terraform.State) error {
|
||||
config := testAccProvider.Meta().(*Config)
|
||||
|
||||
for _, rs := range s.RootModule().Resources {
|
||||
if rs.Type != "google_storage_bucket_object" {
|
||||
continue
|
||||
}
|
||||
|
||||
bucket := rs.Primary.Attributes["bucket"]
|
||||
name := rs.Primary.Attributes["name"]
|
||||
|
||||
objectsService := storage.NewObjectsService(config.clientStorage)
|
||||
|
||||
getCall := objectsService.Get(bucket, name)
|
||||
_, err := getCall.Do()
|
||||
|
||||
if err == nil {
|
||||
return fmt.Errorf("Object %s still exists", name)
|
||||
}
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
var testGoogleStorageBucketsObjectBasic = fmt.Sprintf(`
|
||||
resource "google_storage_bucket" "bucket" {
|
||||
name = "%s"
|
||||
}
|
||||
|
||||
resource "google_storage_bucket_object" "object" {
|
||||
name = "%s"
|
||||
bucket = "${google_storage_bucket.bucket.name}"
|
||||
source = "%s"
|
||||
predefined_acl = "projectPrivate"
|
||||
}
|
||||
`, bucketName, objectName, tf.Name())
|
|
@ -0,0 +1,42 @@
|
|||
---
|
||||
layout: "google"
|
||||
page_title: "Google: google_storage_bucket_object"
|
||||
sidebar_current: "docs-google-resource-storage-object"
|
||||
description: |-
|
||||
Creates a new object inside a specified bucket
|
||||
---
|
||||
|
||||
# google\_storage\_bucket\_object
|
||||
|
||||
Creates a new object inside an exisiting bucket in Google cloud storage service (GCS). Currently, it does not support creating custom ACLs. For more information see [the official documentation](https://cloud.google.com/storage/docs/overview) and [API](https://cloud.google.com/storage/docs/json_api).
|
||||
|
||||
|
||||
## Example Usage
|
||||
|
||||
Example creating a public object in an existing `image-store` bucket.
|
||||
|
||||
```
|
||||
resource "google_storage_bucket_object" "picture" {
|
||||
name = "butterfly01"
|
||||
source = "/images/nature/garden-tiger-moth.jpg"
|
||||
bucket = "image-store"
|
||||
predefined_acl = "publicRead"
|
||||
}
|
||||
|
||||
```
|
||||
|
||||
## Argument Reference
|
||||
|
||||
The following arguments are supported:
|
||||
|
||||
* `name` - (Required) The name of the object.
|
||||
* `bucket` - (Required) The name of the containing bucket.
|
||||
* `source` - (Required) A path to the data you want to upload.
|
||||
* `predefined_acl` - (Optional, Default: 'projectPrivate') The [canned GCS ACL](https://cloud.google.com/storage/docs/access-control#predefined-acl) apply.
|
||||
|
||||
## Attributes Reference
|
||||
|
||||
The following attributes are exported:
|
||||
|
||||
* `md5hash` - (Computed) Base 64 MD5 hash of the uploaded data.
|
||||
* `crc32c` - (Computed) Base 64 CRC32 hash of the uploaded data.
|
|
@ -84,6 +84,10 @@
|
|||
<li<%= sidebar_current("docs-google-resource-storage-bucket") %>>
|
||||
<a href="/docs/providers/google/r/storage_bucket.html">google_storage_bucket</a>
|
||||
</li>
|
||||
|
||||
<li<%= sidebar_current("docs-google-resource-storage-bucket-object") %>>
|
||||
<a href="/docs/providers/google/r/storage_bucket_object.html">google_storage_bucket_object</a>
|
||||
</li>
|
||||
</ul>
|
||||
</li>
|
||||
</ul>
|
||||
|
|
Loading…
Reference in New Issue