A Ruby on Rails plugin that integrates AWS services with your application using
the latest version of AWS SDK For Ruby.
Installation
Add this gem to your Rails project’s Gemfile:
gem 'aws-sdk-rails', '~> 5'
This gem also brings in the following AWS gems:
aws-sdk-core
You will have to ensure that you provide credentials for the SDK to use. See the
latest AWS SDK for Ruby Docs
for details.
If you’re running your Rails application on Amazon EC2, the AWS SDK will
check Amazon EC2 instance metadata for credentials to load. Learn more:
IAM Roles for Amazon EC2
The AWS SDK is configured to use the built-in Rails logger for any
SDK log output. The logger is configured to use the :info log level. You can
change the log level by setting :log_level in the
Aws.config hash.
Aws.config.update(log_level: :debug)
Rails 5.2+ Encrypted Credentials
If you are using Encrypted Credentials,
the credentials will be decrypted and loaded under the :aws top level key:
Encrypted Credentials will take precedence over any other AWS Credentials that
may exist in your environment (e.g. credentials from profiles set in ~/.aws/credentials).
If you are using ActiveStorage
with S3, then you do not need to specify your credentials in your storage.yml
configuration because they will be loaded automatically.
AWS SDK eager loading
An initializer will eager load the AWS SDK for you. To enable eager loading,
add the following to your config/application.rb:
config.eager_load = true
ActiveSupport Notifications for AWS SDK calls
ActiveSupport::Notifications
instrumentation is enabled by default for all AWS SDK calls. Events are
published for each client operation call with the following event name:
<operation>.<serviceId>.aws. For example, S3’s :put_object has an event name
of: put_object.S3.aws. The service name will always match the namespace of the
service client (e.g. Aws::S3::Client => ‘S3’). The payload of the event is the
request context.
You can subscribe to these events as you would for other
ActiveSupport::Notifications:
ActiveSupport::Notifications.subscribe('put_object.S3.aws') do |name, start, finish, id, payload|
# process event
end
# Or use a regex to subscribe to all service notifications
ActiveSupport::Notifications.subscribe(/S3[.]aws/) do |name, start, finish, id, payload|
# process event
end
Elastic Beanstalk ActiveJob processing
Elastic Beanstalk worker environments
can be used to run ActiveJob without managing a worker process. To do this,
configure the worker
to read from the correct SQS queue that you want to process jobs from and set
the AWS_PROCESS_BEANSTALK_WORKER_REQUESTS environment variable to true in
the worker environment configuration. The
SQS Daemon
running on the worker sends messages as a POST request to http://localhost/.
The ElasticBeanstalkSQSD middleware will forward each request and parameters to their
appropriate jobs. The middleware will only process requests from the SQS daemon
and will pass on others and so will not interfere with other routes in your
application.
To protect against forgeries, daemon requests will only be processed if they
originate from localhost or the Docker host.
Running Jobs Async
By default the ElasticBeanstalkSQSD middleware will process jobs synchronously
and will not complete the request until the job has finished executing. For
long running jobs (exceeding the configured nginix timeout on the worker) this
may cause timeouts and incomplete executions.
To run jobs asynchronously, set the AWS_PROCESS_BEANSTALK_WORKER_JOBS_ASYNC
environment variable to true in your worker environment. Jobs will be queued
in a ThreadPoolExecutor and the request will return a 200 OK immediately and the
SQS message will be deleted and the job will be executed in the background.
By default the executor will use the available processor count as the the
max_threads. You can configure the max threads for the executor by setting
the AWS_PROCESS_BEANSTALK_WORKER_THREADS environment variable.
When there is no additional capacity to execute a task, the middleware
returns a 429 (too many requests) response which will result in the
sqsd NOT deleting the message. The message will be retried again once its
visibility timeout is reached.
Periodic (scheduled) tasks will also be run asynchronously in the same way.
Elastic beanstalk queues a message for the periodic task and if there is
no capacity to execute the task, it will be retried again once the message’s
visibility timeout is reached.
Periodic (scheduled) jobs
Periodic (scheduled) tasks
are also supported with this approach. Elastic
Beanstalk workers support the addition of a cron.yaml file in the application
root to configure this. You can call your jobs from your controller actions
or if you name your cron job the same as your job class and set the URL to
/, the middleware will automatically call the job.
AWS SDK for Ruby Rails Plugin
A Ruby on Rails plugin that integrates AWS services with your application using the latest version of AWS SDK For Ruby.
Installation
Add this gem to your Rails project’s Gemfile:
This gem also brings in the following AWS gems:
aws-sdk-coreYou will have to ensure that you provide credentials for the SDK to use. See the latest AWS SDK for Ruby Docs for details.
If you’re running your Rails application on Amazon EC2, the AWS SDK will check Amazon EC2 instance metadata for credentials to load. Learn more: IAM Roles for Amazon EC2
Features
ActionDispatch DynamoDB Session Storage
See https://github.com/aws/aws-actiondispatch-dynamodb-ruby
ActionMailer delivery with Amazon Simple Email Service
See https://github.com/aws/aws-actionmailer-ses-ruby
ActionMailbox ingress with Amazon Simple Email Service
See https://github.com/aws/aws-actionmailbox-ses-ruby
ActiveJob SQS adapter
See https://github.com/aws/aws-activejob-sqs-ruby
AWS Record Generators
See https://github.com/aws/aws-record-rails
AWS SDK uses the Rails logger
The AWS SDK is configured to use the built-in Rails logger for any SDK log output. The logger is configured to use the
:infolog level. You can change the log level by setting:log_levelin the Aws.config hash.Rails 5.2+ Encrypted Credentials
If you are using Encrypted Credentials, the credentials will be decrypted and loaded under the
:awstop level key:Encrypted Credentials will take precedence over any other AWS Credentials that may exist in your environment (e.g. credentials from profiles set in
~/.aws/credentials).If you are using ActiveStorage with
S3, then you do not need to specify your credentials in yourstorage.ymlconfiguration because they will be loaded automatically.AWS SDK eager loading
An initializer will eager load the AWS SDK for you. To enable eager loading, add the following to your
config/application.rb:ActiveSupport Notifications for AWS SDK calls
ActiveSupport::Notifications instrumentation is enabled by default for all AWS SDK calls. Events are published for each client operation call with the following event name:
<operation>.<serviceId>.aws. For example, S3’s:put_objecthas an event name of:put_object.S3.aws. The service name will always match the namespace of the service client (e.g. Aws::S3::Client => ‘S3’). The payload of the event is the request context.You can subscribe to these events as you would for other
ActiveSupport::Notifications:Elastic Beanstalk ActiveJob processing
Elastic Beanstalk worker environments can be used to run ActiveJob without managing a worker process. To do this, configure the worker to read from the correct SQS queue that you want to process jobs from and set the
AWS_PROCESS_BEANSTALK_WORKER_REQUESTSenvironment variable totruein the worker environment configuration. The SQS Daemon running on the worker sends messages as a POST request tohttp://localhost/. The ElasticBeanstalkSQSD middleware will forward each request and parameters to their appropriate jobs. The middleware will only process requests from the SQS daemon and will pass on others and so will not interfere with other routes in your application.To protect against forgeries, daemon requests will only be processed if they originate from localhost or the Docker host.
Running Jobs Async
By default the ElasticBeanstalkSQSD middleware will process jobs synchronously and will not complete the request until the job has finished executing. For long running jobs (exceeding the configured nginix timeout on the worker) this may cause timeouts and incomplete executions.
To run jobs asynchronously, set the
AWS_PROCESS_BEANSTALK_WORKER_JOBS_ASYNCenvironment variable totruein your worker environment. Jobs will be queued in a ThreadPoolExecutor and the request will return a 200 OK immediately and the SQS message will be deleted and the job will be executed in the background.By default the executor will use the available processor count as the the max_threads. You can configure the max threads for the executor by setting the
AWS_PROCESS_BEANSTALK_WORKER_THREADSenvironment variable.When there is no additional capacity to execute a task, the middleware returns a 429 (too many requests) response which will result in the sqsd NOT deleting the message. The message will be retried again once its visibility timeout is reached.
Periodic (scheduled) tasks will also be run asynchronously in the same way. Elastic beanstalk queues a message for the periodic task and if there is no capacity to execute the task, it will be retried again once the message’s visibility timeout is reached.
Periodic (scheduled) jobs
Periodic (scheduled) tasks are also supported with this approach. Elastic Beanstalk workers support the addition of a
cron.yamlfile in the application root to configure this. You can call your jobs from your controller actions or if you name your cron job the same as your job class and set the URL to/, the middleware will automatically call the job.Example:
and in your controller:
Will execute cron
SomeJobevery minute andSomeJobevery 12 hours via the/scheduledendpoint.