Atlas
- A brief description of what you are looking to do
- How you think this will help
- Why this matters to you
1258 results found
-
Allow "Project Data Access Read Only" to retrieve restore links.
Allow "Project Data Access Read Only" to retrieve restore links. Currently, to retrieve a restore link from an Atlas cluster you must be a Project Owner.
In the current version to take the link to the snapshot we need to use Restore Jobs part of the API. Problem is that the same POST endpoint is used for the restore job which can do changes in the system and needs Owner level of the access and to generate link to the stream with the snapshot data. This second type of the job should not need so big access level as should…
4 votes -
Separate Atlas metric graphs of the outgoing traffic from the internal/replication traffic in Atlas cluster
This is a request to separate Atlas metric graphs of the outgoing traffic from the internal/replication traffic in Atlas cluster
4 votes -
Terminate existing connections from IPs when removed from the whitelist
When removing entries from an Atlas Project's IP whitelist, existing connections from those IPs are maintained, and thus those hosts can continue to perform operations until the connection times out. I'm aware such connections can be dropped by doing a Test Failover or pausing and unpausing the cluster, but in some scenarios this is not ideal if even possible. This also seems like a security gap: When access is revoked from a particular IP or CIDR block, any existing client connections originating from that IP or CIDR should be terminated and the hosts behind that IP or CIDR should be…
4 votes -
Limit certain users to create only specific cluster sizes
After discovering over-spending forced to remove Atlas permissions for all developers. Would be nice if I could limit them to create only M0 / M2 / M5 clusters as sandbox.
4 votes -
Description field for VPC peering page
Add a description field to easily identity the peering items in Peerings tab. Example: "Peering for Stage Kubernetes cluster apps"
4 votes -
Give an option to pause the cluster as a parameter, at the moment it is auto activating after 30 days of cluster pause which is not required
Give an option to pause the cluster as a parameter for specific duration of time, at the moment it is auto activating after 30 days of cluster pause which is not required for all the users and manually we have to keep an eye on this at the moment.
4 votes -
Atlas - Preferences - My Date Format - allow 24hr
We can specify date format but the time format is limited to 12hr clock. Please allow us to specify 24hr so the activity log and rest of the UI is a little easier to read. Thanks
4 votes -
Terraform lifecycle ignore_changes tags
It would be nice if tags would not be set of list and will be a map, like tags for Azure resources.
In that case, you can ignore some tags by name. Like thislifecycle {
ignore_changes = [
tags["costcenter"],
tags["environment"],
tags["projectcode"]
]
}https://github.com/mongodb/terraform-provider-mongodbatlas/issues/2006
3 votes -
"Manage your own encryption keys deactivated" event
We recently discovered that there is no event when deactivating the option "Manage your own encryption keys" within the cluster configuration.
While there are hints towards this within the CLUSTERUPDATESTARTED and AUTOMATIONCONFIGPUBLISHED_AUDIT events, there is no clear warning that the encryption keys of the cluster are affected.
As an unwanted change of encryption at rest keys for data can have severe impact on data accessibility and security, I would propose adding a "Manage your own encryption keys deactivated" event to the Activity Feed.
3 votes -
Push Logs to AWS S3 bucket: Terraform
Push Logs to AWS S3 bucket, This feature is helpful, It gives a flexibility for large organizations who need to segregate logs considering Inventory, Security and Observability to organize logs to a respective destinations from S3.However the terraform provider does have api's to fulfill this feature https://www.mongodb.com/docs/atlas/push-logs/ . If it can available in terms of terraform resources, This will help our organization to enable logging in terraform IAC to simplify the logging and deployments
3 votesHi Jai, thanks for sharing feedback. This item is already under development and should be published to Terraform Registry over next few weeks!
-
API - Version 2
We saw that the api version is now in v2 for some resources (as clusters) - https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Clusters/
We tried to change some App services functions from v1 to v2, but ended resulting in some errors (or needing to add more parameters than the original one - in version 1).
Using version 1, we only inform what we need to change ("instanceSize") plus the providerName.to use version 2, we need to inform all this parameters (if it's a replica-set. If it is a sharding we also need to inform the numShards):
"replicationSpecs":[{"regionConfigs":[{"electableSpecs":{"instanceSize":"M10","nodeCount":"3"},"priority":"7","providerName":"GCP","regionName":"CENTRAL_US"}]}]}What I need to change is only the…
3 votes -
MongoAtlas Orga - Maximum number API keys exceeded
We have detected in our organisation that the limit of API keys has been reached. This is currently affecting our project teams in terms of resource distribution, so we are asking for an increase in the quoter in the short term.
Problem: We can't create keys and also delete them. The problem is that the key does not have an owner. Only org owner can delete this key. They did but this is only a reference deletion.
Alerting: Condition for alerting is not there
What we want to achieve:
- Transparency for the org owners about what the limit level…3 votes -
Support for push-based logging integration with Google Cloud Storage
As of now, MongoDB Atlas supports log export functionality exclusively with AWS S3. In addition to the existing AWS S3 bucket integration, would be good to have the capability of pushing logs to GCP Cloud Storage for organisations who depend on GCP for cloud infrastructure.
3 votes -
Atlas DNS name more than 23 characters from the cluster name.
Allow users to have longer DNS name from the cluster name. Currently, Atlas cuts the first 23 characters from the cluster name and the users can't check the cluster name inside the mongo shell.
Example:
Cluster name longer than 23 characters : xxxx-us-east-1-multi-use-1
DNS name : xxxx-us-east-1-multi-us.xsj8t.mongodb.net3 votes -
AWS gp3 Decouple IOPs from Disk Size
With AWS EBS gp3 volumes, IOPs and throughput can be provisioned separately from EBS storage size. While Atlas now uses gp3 volumes and provides a base throughput of 3,000 IOPs, higher throughputs are still directly tied to disk size.
We run IO-intensive workloads that require high throughput, and we have to severely over-size disks to get the needed throughput. We don't require the extra-low latency of provisioned IOPs (which is much more expensive than over-provisioning storage).
According to the linked AWS documentation below, if the M60 instance size is running on an ec2 m5.4xlarge instance type, it can support a…
3 votes -
Manage federated database views via Terraform
The mongodbatlasfederateddatabase_instance resource allows managing tables in the federated instance, but it does not allow us to manage views. Please update the provider so we can also create and manage views
3 votes -
Support for GCP Saudi Arabia/Dammam me-central2 for Atlas
GCP announced availability of services in me-central2. Can MongoDB please support this region in GCP?
3 votes -
Retrieve latest/max oplog timestamp via the Atlas Administration API
At the moment the latest/max oplog timestamp can be retrieved via the Atlas UI in the backups section for point in time restores. This feedback post is to make this information available as part of the atlas administration api.
For example, in the UI it states:
You can only restore up to this oplog timestamp: 1692580549 and this increment: 1.
Would be good to have this value available to be retrieved via the API for restores requiring oplog timestamp values in this format.
3 votes -
Enable database users separate read/write permissions based on cluster
We have multiple clusters in the same project, and we need to limit write access to certain CLUSTERS while leaving read only access to any other cluster. The problem is that we cannot do this by DATABASE name as there are many of them, as well as dynamically addition of new ones in clusters, so it will become obsolete and very hard to maintain.
Trying to use the restriction does not help as it limits all access, including read only, to those clusters.
What we'd like to see is that we can limit write access by cluster (not database/collection) and…3 votes -
Datadog integration for AP1 region
I am using Datadog region AP1 (Japan) and I learned that Atlas doesn't support the region for the integration.
It would be really great and helpful if the integration supports the AP1 region.
3 votes
- Don't see your idea?