Awesome
Carrierwave AWS Storage
Use the officially supported AWS-SDK library for S3 storage rather than relying on fog. There are several things going for it:
- Full featured, it supports more of the API than Fog
- Significantly smaller footprint
- Fewer dependencies
- Clear documentation
Here is a simple comparison table [07/17/2013]
Library | Disk Space | Lines of Code | Boot Time | Runtime Deps | Develop Deps |
---|---|---|---|---|---|
fog | 28.0M | 133469 | 0.693 | 9 | 11 |
aws-sdk | 5.4M | 90290 | 0.098 | 3 | 8 |
Installation
Add this line to your application's Gemfile:
gem 'carrierwave-aws'
Run the bundle command from your shell to install it:
bundle install
Usage
Configure and use it just like you would Fog. The only notable difference is
the use of aws_bucket
instead of fog_directory
, and aws_acl
instead of
fog_public
.
CarrierWave.configure do |config|
config.storage = :aws
config.aws_bucket = ENV.fetch('S3_BUCKET_NAME') # for AWS-side bucket access permissions config, see section below
config.aws_acl = 'private'
# Optionally define an asset host for configurations that are fronted by a
# content host, such as CloudFront.
config.asset_host = 'http://example.com'
# config.asset_host = proc { |file| ... } # or can be a proc
# The maximum period for authenticated_urls is only 7 days.
config.aws_authenticated_url_expiration = 60 * 60 * 24 * 7
# Set custom options such as cache control to leverage browser caching.
# You can use either a static Hash or a Proc.
config.aws_attributes = -> { {
expires: 1.week.from_now.httpdate,
cache_control: 'max-age=604800'
} }
config.aws_credentials = {
access_key_id: ENV.fetch('AWS_ACCESS_KEY_ID'),
secret_access_key: ENV.fetch('AWS_SECRET_ACCESS_KEY'),
region: ENV.fetch('AWS_REGION'), # Required
stub_responses: Rails.env.test? # Optional, avoid hitting S3 actual during tests
}
# Optional: Signing of download urls, e.g. for serving private content through
# CloudFront. Be sure you have the `cloudfront-signer` gem installed and
# configured:
# config.aws_signer = -> (unsigned_url, options) do
# Aws::CF::Signer.sign_url(unsigned_url, options)
# end
end
Custom options for S3 endpoint
If you are using a non-standard endpoint for S3 service (eg: Swiss-based Exoscale S3) you can override it like this
config.aws_credentials[:endpoint] = 'my.custom.s3.service.com'
Custom options for AWS URLs
If you have a custom uploader that specifies additional headers for each URL, please try the following example:
class MyUploader < Carrierwave::Uploader::Base
# Storage configuration within the uploader supercedes the global CarrierWave
# config, so either comment out `storage :file`, or remove that line, otherwise
# AWS will not be used.
storage :aws
# You can find a full list of custom headers in AWS SDK documentation on
# AWS::S3::S3Object
def download_url(filename)
url(response_content_disposition: %Q{attachment; filename="#{filename}"})
end
end
Configure the role for bucket access
The IAM role accessing the AWS bucket specified when configuring CarrierWave
needs to be given access permissions to that bucket. Apart from the obvious permissions required depending on what you want to do (read, write, delete…), you need to grant the s3:PutObjectAcl
permission (a permission to manipulate single objects´ access permissions) lest you receive an AccessDenied
error. The policy for the role will look something like this:
PolicyDocument:
Version: '2012-10-17'
Statement:
- Effect: Allow
Action:
- s3:ListBucket
Resource: !Sub 'arn:aws:s3:::${BucketName}'
- Effect: Allow
Action:
- s3:PutObject
- s3:PutObjectAcl
- s3:GetObject
- s3:DeleteObject
Resource: !Sub 'arn:aws:s3:::${BucketName}/*'
Remember to also unblock ACL changes in the bucket settings, in Permissions > Public access settings > Manage public access control lists (ACLs)
.
Migrating From Fog
If you migrate from fog
your uploader may be configured as storage :fog
,
simply comment out that line, as in the following example, or remove that
specific line.
class MyUploader < Carrierwave::Uploader::Base
# Storage configuration within the uploader supercedes the global CarrierWave
# config, so adjust accordingly...
# Choose what kind of storage to use for this uploader:
# storage :file
# storage :fog
storage :aws
# More comments below in your file....
end
Another item particular to fog, you may have url(query: {'my-header': 'my-value'})
.
With carrierwave-aws
the query
part becomes obsolete, just use a hash of
headers. Please read [usage][#Usage] for a more detailed explanation about
configuration.
Contributing
In order to run the integration specs you will need to configure some
environment variables. A sample file is provided as .env.sample
. Copy it over
and plug in the appropriate values.
cp .env.sample .env
- Fork it
- Create your feature branch (
git checkout -b my-new-feature
) - Commit your changes (
git commit -am 'Add some feature'
) - Push to the branch (
git push origin my-new-feature
) - Create new Pull Request