-
-
Notifications
You must be signed in to change notification settings - Fork 101
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add The Ability To Set Lifecycle Rules To DMS Tasks #20
Comments
This is a critical feature from my perspective, I've got a task settings file which seems to be getting re-ordered by AWS ahead of the diff, which means terraform attempts to shut down the task and replace the task settings on every apply. This alone basically makes the service unusable for any purpose requiring high uptime in a continuous deployment environment. Add to that the fact that terraform can easily time out and fail when attempting to stop the task in order to update the settings and it comes to a point where DMS is no longer fit for purpose because of how terraform (and more specifically this module) is handling changes to task settings. I will likely have to rewrite the terraform without the module at this stage to work around this issue. |
unfortunately, we cannot use variables inside lifecycle rules hashicorp/terraform#3116 |
This issue has been automatically marked as stale because it has been open 30 days |
This issue was automatically closed because of stale in 10 days |
Just running into this...this is painful. Aside from the manual stop/start I'd like to ignore task setting changes for a couple items (Cloudwatch log group and stream) but every time I run TF it's resetting them. I'm not sure how we will operationalize this.... would have to manually stop DMS tasks, run any terraform changes then restart them. |
Having just recently put DMS into production via Terraform I would strongly recommend just not using this library. It's not particularly complicated to set it up from scratch by comparison and you'll be able to control the lifecycle settings. |
hashicorp/terraform-provider-aws#21512 |
I'm regretting it a bit now - I might pull the module locally. Due to lifecycle being non-dynamic are you putting lifecycle settings hard-coded in the module? Or are you just listing out resources without iterators and using lifecycle settings as needed per-resource? |
@mkohlmyr / @casey-robertson-paypal / @alordthorsen - would you mind sharing what you are specifying (or looking to specify) in the lifecycle ignore block, please? While we don't have a native way to variablize what settings are ignored by Terraform, we have supported "hacky" workarounds for common cases. I can't say that we can support every setting, but maybe there is a common group or a few that will make a difference. The reason we can only support a few is just overhead - we have to duplicate the resource (here, the task) for each scenario. In the mean time, please feel free to use the module for reference and/or fork it to suit your immediate needs. If I can collect more information, this will help determine what changes we could make to make this work better |
The cloudwatch use case was the one that came to mind - I'll need to dedicate some time to try again. I had given up and just let it create the trails. I thought at first I'd like to ignore the start/stop status of the job but realized some of those messages or errors were coming from the API - i.e. you simply aren't allowed to update certain resources while jobs are running. This is an AWS limitation, not a module issue. Looking forward to hearing the use cases of the others too. |
This issue has been automatically marked as stale because it has been open 30 days |
@bryantbiggs The issue link regarding lifecycle interpolation was just closed, saying it is possible, depending on the scenario. Does that allow for new solutions to this issue?
|
Also have experienced a lot of annoyance, having tasks spuriously restarted because of the inability to specify the lifecycle. |
Is your request related to a new offering from AWS?
Is this functionality available in the AWS provider for Terraform? See CHANGELOG.md, too.
lifecycle
block I found was in aws-provider:0.1.2 (2017). I think it was added in 0.6.0 (2015) from the terraform change log.Is your request related to a problem? Please describe.
I will start off by saying that my company is currently experimenting with DMS and is looking at the costs and performance characteristics of the service on our systems. By default we don't want tasks to start in the "start" state.
Currently there's no way to pass a lifecycle configuration to the produced
aws_dms_replication_task
's that are produced by this module. The effect of this is that when tasks are turned on manually (like what I'm doing right now). The effect of this is that you get plans likeWhere terraform is going to turn off a task that has been turned on.
The net effect of this is that if you want to have a system where
You need to accept that tasks will be stopped on each apply and you need to manually re-start them currently.
Describe the solution you'd like.
Add the ability to pass a lifecycle rule. Specifically being able to say
Describe alternatives you've considered.
Manually accept restarting tasks. =P
The text was updated successfully, but these errors were encountered: