Atlas
- A brief description of what you are looking to do
- How you think this will help
- Why this matters to you
138 results found
-
Terraform Serverless VPC Endpoint configuration
Create the equivalent of mongodbatlasprivatelinkendpoint but for serverless.
8 votes -
Enable setting encryption at rest details for project
Please allow us to set the encryption at rest KMS details for the project when we create the project
1 voteThis is available and supported.
https://www.mongodb.com/docs/atlas/atlas-operator/stable/ak8so-encryption-at-rest-customer-keys/
-
Configure --jsonFormat=canonical flag in export policy.
JSON does not support all data types that are available in BSON. This means that when using JSON there will be a so called "loss of fidelity" of the information.
However, using the --jsonFormat=canonical flag in a mongoexport command will preserve all available BSON data types, so the "loss of fidelity" issue can be completely avoided.Now we plan to export our cloud backups to an AWS S3 bucket. To do this, we would like to set up an export policy to automatically export the snapshots. We could already do this via the API. However, the data is output in…
2 votes -
Datadog integration for US5
Hi I am using US5 datadog, and as I learned that Atlas only supports US1, it would be great if the integration with US5 is also added.
1 voteThe US5 Datadog site is available in Atlas.
-
MS Teams alert support in terraform provider
It is possible to configure MS Teams alerts in the atlas UI, however terraform support is still missing. It would be great to have that option
4 votesThis is now out in version 1.4.2.
-
Disable data explorer and other features in project settings
The following features cannot be disabled in the terraform provider (project settings):
Real Time Performance Panel
Data Explorer
Performance Advisor and Profiler
Schema Advisor2 votesThe project settings can now be set via the project resource in the provider. Thank you!
-
Add autoExport snapshot to AWS S3 Bucket on mongodbatlas_cloud_backup_schedule
By company policy, we have to export our snapshots automatically to an AWS S3 Bucket.
I started following https://www.mongodb.com/docs/atlas/backup/cloud-backup/export/ and implemented on terraform due to the high number of projects, and clusters that we need to backup.
However, looks like the terraform provider doesn't support "autoExportEnabled" from https://www.mongodb.com/docs/atlas/reference/api/cloud-backup/schedule/modify-one-schedule/ on https://registry.terraform.io/providers/mongodb/mongodbatlas/latest/docs/resources/cloud_backup_schedule terraform resource.
Best regards,
Wagner Sartori Junior9 votesThis is now out in version 1.4.2.
-
Granular permissions via roles / hashicorp vault
We are using the Hashicorp Vault Atlas plugin in order to generate credentials for Atlas.
We are able to generate roles on the Atlas end and then use those roles to provision vault users.
However, I don't see a way to restrict those roles to just certain resources/clusters.
So the user can access all the deployments in a project.
It should be possible to restrict roles to certain resources only.1 voteHi,
This has already been added some time ago. You need to specify the scopes you want to include, here's the code - https://github.com/hashicorp/vault-plugin-database-mongodbatlas/blob/master/mongodbatlas.go#L206. It should be an array like roles, but with the resource name and then if it's a cluster or data lake (see scopes here: https://www.mongodb.com/docs/atlas/reference/api/database-users-create-a-user/). I hope that helps!
Best,
Melissa
-
Comprehensive Backup Ransomware Protection
MongoDB Atlas needs a modern, comprehensive, secure ransomware protection strategy for its customers. Simply providing the ability to backup a database, and encrypt that database with "bring your own key" is not enough. Below I highlight what I believe are key components of a comprehensive strategy (or at least a good start).
Immutable and Verifiable Backups
Once backups are created, Atlas should provide a facility to ensure the backup remains immutable. Further, Atlas should provide verification that a backup continues to be untouched / unmodified for its entire lifecycle.
Deletion Protection
Atlas should provide enhanced deletion protection for backups. Any…
3 votesHello,
I am pleased to announce that we have released our backup feature called Backup Compliance Policy, that protects your backups from being deleted by any user, ensuring WORM and full immutability (can not be edited/modified or deleted) for backups automatically in Atlas.
Backup Compliance Policy allows organizations to configure a project-level policy to prevent the deletion of backups before a predefined period, guarantee all clusters have backup enabled, ensure that all clusters have a minimum backup retention and schedule policy in place, and more.
With these controls, you can more easily satisfy data protection requirements (e.g., AppJ, DORA, immutable / WORM backups, etc.) without the need for manual processes.
Please note that the Backup Compliance Policy can not be disabled without MongoDB support once enabled so please make sure to read our documentation thoroughly before enabling.
In addition to Backup Compliance Policy, organizations can also utilize our multi-region…
-
Disable Specific API's
For certain API's, like the ability to Delete a backup, have the ability for an Owner to disable this API call entirely, to prevent bad actors from being able to destroy a system or even a good actor from unintentionally destroying a system. If a customer has a policy that no backups shall be deleted ever, have the ability to disable this API across the board.
2 votesHello,
I am pleased to announce that we have released our backup feature called Backup Compliance Policy, that protects your backups from being deleted by any user, ensuring WORM and full immutability (can not be edited/modified or deleted) for backups automatically in Atlas. This applies to any method of deleting backups, regardless of wheter it is through the UI or the API.
Backup Compliance Policy allows organizations to configure a project-level policy to prevent the deletion of backups before a predefined period, guarantee all clusters have backup enabled, ensure that all clusters have a minimum backup retention and schedule policy in place, and more.
With these controls, you can more easily satisfy data protection requirements (e.g., AppJ, DORA, immutable / WORM backups, etc.) without the need for manual processes.
Please note that the Backup Compliance Policy can not be disabled without MongoDB support once enabled so please make sure…
-
Return private endpoints for peered network from mongo-db prometheus discovery endpoint
We are using VPC peering to connect with Mongo Atlas. With the recent account about, prometheus integration. We added scrape config to mongo-db discovery API. However, scraping times out. Upon checking further it is found that discovery API returns public endpoints not private ones. Hence connection is failing. Is there a way that discovery API can send private endpoints.
10 votesI'm happy to announce that the Prometheus integration does now support VPC peering. This can be configured in the Prometheus configuration modal in the user interface when using the HTTP SD discovery method.
More information on how to configure this can also be found here: https://www.mongodb.com/docs/atlas/reference/api/third-party-integration-settings-discovery/#request-query-parameters
-
ReadOnly DATA API
Current DATA API feature looks promising. However there is no way provide access controls around it. If you have access to API key then you can potentially do both read-write to cluster. We did some PoC recently but couldnt promote to prod because of this problem. If we get a readonly Data API access that would be super helpful.
3 votesHi all this should be completed now.
-
Allow to assign API Key to Project via Terraform by referencing public key
Currently you can only assign an API Key to an Atlas Project via Terraform by referencing the ID of the API Key. Unfortunately, the ID is not exposed through the UI, only through the Atlas API. This is not very ideal for customers/users that are creating API Keys through the UI.
Adding support for referencing the API Key in Terraform via the public key instead of the ID would fix this. Or alternatively, exposing the API Key ID in Atlas.
3 votesWe released support for Data Source: mongodbatlas_api_keys which exposes the api_key_id paramter as part of v1.8.0. Thank you for the feedback.
For more details see: https://registry.terraform.io/providers/mongodb/mongodbatlas/latest/docs/data-sources/api_keys#api_key_id
-
A metric of availability of a cluster mongodb atlas for building my SLO
Hello
As other databases provider , their databases services (like postgresql ....) give metrics on availability of a the service (ex: postgres.canconnect) . I would like to have some metrics on the availibity of a cluster mongodb atlas for building my SLO:
ex: mongodb.canconnectThanks,
15 votes -
Add Prometheus as a Supported Third-Party Integration Settings type
Great work releasing the new Prometheus Integration functionality!
Ideally, we'd like to use Terraform to codify our interface with the Prometheus Integration, similar to how we leverage the existing Third-Party Integration Settings types.
8 votesThis is now out in version 1.4.2 of the provider.
-
Allow Management of Identity Providers through Terraform
The necessary API endpoints appear to exist, it would be great to expose them in the Terraform provider for Atlas to avoid having a split between Infrastructure-as-Code and having to browse the web interface.
2 votesThis is now out in version 1.4.2
-
Please add datadog US3 site also for the integration with MongoDB atlas
Please add datadog US3 site also for the integration with MongoDB atlas
8 votesThe US3 Datadog site is available.
-
Terraform integration to manage project's API keys
Hello team,
we're still missing the capability to create project's API keys via terraform and have to deal with some python scripts to automatically provision new keys in dynamic way(when new projects come up).
I know this question has been already asked and your idea is to use Vault, but probably it's not good idea to install this heavy tool for the single application...and we can't use Vault in dynamic way because it doesn't contain 'MongodbAtlas Secret Engine' available via terraform provider.
Please consider to add possibility of creation API keys via terraform and let us decide whether it's security…
30 votes -
AWS CFN MongoDB::Atlas::<> in all AWS regions
Please publish the MongoDB::Atlas::<> resources types into all AWS regions (currently only in us-east-1).
Some of us have restrictions in place that deny the direct use of us-east-1.
Thanks,2 votes -
Support tags on Mongo databases
I have been supporting multiple projects/clusters with databases from different business units.
I'd like the ability to tag databases using a key value pair - similar to GCP, Azure, AWS. Tags will help us organise and understand what's running in our clusters and ensuring a more efficient/granular billing process.
5 votesI'm excited to announce that we recently released tags on database deployments. You can now better organize and manage Atlas resources by applying tags to categorize database deployments based on your unique business needs.
https://www.mongodb.com/docs/atlas/tags/
Check it out today! You can apply tags to new or existing clusters or serverless instances.
- Don't see your idea?