CELR Decommission Process
Date
Jul 3, 2024
Attendees:
APHL/AIMS | Peraton | CDC/MVPS |
---|---|---|
Geo Miller: -- | Erroll Rosser: | Tricia Aden |
Gretl Glick: | Kristin Peterson | Teresa Jue |
Vanessa Holley |
| Megan Mueller |
Laura Carlton |
|
|
Dari Shirazi |
|
|
Eddie |
|
|
John Reaves |
|
|
Goals
Discussion topics
Item | Notes |
---|---|
Next Steps/Action Items | Agenda/Objectives:
|
| · This is a good question to ask; as there is some effort to identify, archive, disable, and delete the resources. · High Level: o Decommission Mirth configurations for CELR
o Archive and Remove CELR AWS components o Archive and Remove CELR services deployed to k8s o Archive and remove Redshift cluster ('onboard' environment only) · More specific for each active environment: o Disable and archive mirth destination to AIMS hosted CELR o Identify all S3 buckets used for CELR o Archive S3 bucket data (discuss when we might also delete) o Delete CELR pods from k8s cluster o Delete route53 entries for CELR k8s services (portal, etc) o Archive and remove CELR services docker images o Delete any infrastructure specific to CELR used by the CELR services in k8s § We will need to review if any k8s clusters can be downsized or removed as a result of this work o Delete CELR (named as alice) cloudformation stacks o Archive and delete ECS/Fargate tasks for S3/Redshift integration o Archive redshift data (discuss when we might also delete) o Turn off redshift cluster o Archive gitlab repositories o Final review of AIMS environments to validate we have not missed items not listed here. Can AIMS reprocess 22 ½ million messages from 2022 (and some from 2023; they will be archiving all of the data)? · I assume this is asking for a retroactive COPY from the CELR output to redshift and a final UNLOAD from Redshift for HHS pickup. o If this is the case, the S3 bucket CELR outputs to does have old data that could be copied to redshift § I would like to avoid doing this work, as it will not be easy to coordinate and takes a long time. Additionally, doing it is likely to cause more work if new questions come up after that copy and how the data changes. The theories about the time ranges they want recopied to Redshift could be wrong and have us doing a lot of back and forth investigating. The team already copied all of the data before redshift out of AIMS, I'd highly suggest that be used moving forward. § It takes a very long time to ingest large blocks of historical time ranges of data into redshift, and becomes something we can't catch up with incoming data streams. We'd have to at a minimum stop exporting to HHS for a few days while we copy data into redshift. The best idea I have here is that we could do this only after we've turned off any new data coming into the data lake. § We will need very specific dates or date ranges that are desired for a re-copy. § An easier idea is just to provide the CELR output bucket data (before Redshift), and let the downstream teams take it from there, but I believe the team already did this. o If this is not the case, we cannot reprocess messages further upstream due to data retention policies |
|
|
Action items
Quick decisions not requiring context or tracking
For quick, smaller decisions that do not require extra context or formal tracking, use the “Add a decision…” function here.
Decisions requiring context or tracking
For decisions that require more context (e.g., documentation of discussion, options considered) and/or tracking, use the decision template to capture more information.