Atlas
- A brief description of what you are looking to do
- How you think this will help
- Why this matters to you
471 results found
-
Atlas Console option to start the resynchronization of a primary node
It would be great if during an automatic maintenance process, considering the restart of all nodes you were able to find a button in the Atlas Console to start the restart of the primary node, so that way you can select this process in a less operative window and avoid a severe business impact in your organization.
1 vote -
Within Atlas, create service to allow project owners to schedule shutdown/start up times of cluster to reduce hosting costs
Request to provide an enhancement to the Atlas service to allow project owners to schedule shutdown/start up times of clusters(or databases) and provide some notification of a successful shutdown and startup. This should be a feature with the Atlas UI and API.
This will allow teams that have non-critical environments to effectively "turn off" and freeze the state of a cluster in order to reduce hosting cost during off hours. This can provide a better value proposition for existing customers and a selling point for new customers, without having to introduce a third party tool that requires additional configuration and/or…
2 votes -
UI/UX: make site scaling for large screens
UI/UX: make site scaling for large screens, now text fonts are too small
1 vote -
Atlas DNS name more than 23 characters from the cluster name.
Allow users to have longer DNS name from the cluster name. Currently, Atlas cuts the first 23 characters from the cluster name and the users can't check the cluster name inside the mongo shell.
Example:
Cluster name longer than 23 characters : xxxx-us-east-1-multi-use-1
DNS name : xxxx-us-east-1-multi-us.xsj8t.mongodb.net3 votes -
Atlas SQL Interface: Support queries based on BinData (UUID)
We have a need to write reports that query MongoDB via the Atlas SQL Interface using BinData (a UUID with BinData type 4) as selection criteria (see support ticket Case 01313460: Query bindata using atlas sql interface). We were told that this is currently not supported, although querying data based on a unique identifier is a most basic and essential function as far as DB interfaces are concerned.
Example query from ticket: SELECT * FROM bn_events WHERE _id = CAST('6b0be834-c03e-47bc-9c7d-4726a1e8c69a' AS BINDATA)
This is a serious limitation that may prevent us from being able to use Atlas SQL Interface for…
1 vote -
Use wildcard databases in user role privileges (like collections)
Use wildcard databases in user role privileges (like collections)
12 votes -
Allow multiple VPC CIDR ranges in Peering connection
Allow multiple VPC CIDR for a single Peering connection from Atlas to AWS.
Our AWS account has two CIDR ranges for a single VPC. Atlas peering connection only allows connection to a single CIDR range.
Today, I need to create a separate private link connection to be able to connect to the second CIDR range.30 votes -
To Include sub dashboards Format in Atlas Charts
To Integrate related dashboards as sub-dashboards under a single main dashboard to improve organization and navigation
1 vote -
(RBAC) Remove CREATE_COLLECTION grant from INSERT permission
In according to the least privilege principle we would like to have a better granularity when defining roles. This feature would help us to have uncontrolled collection creations by applications
2 votes -
Always show an option for both SRV Connection String and Standard Connection String
There are more than a few examples where users DNS does not resolve SRV Connections strings. They are often directed to select an old version / legacy connection string.
Two issues with this are:
1. Sometimes outdated/deprecated options are provided (ssl=true)
2. Serverless does not provide this.1 vote -
UUID Schema Validation
It would be very handy to have the ability to specify a schema validator using bson type UUID (binData subtype 4) instead of the more generic binData subtype 0.
11 votes -
Azure VNet terminology needs updating
Setting up the Peering Connection (Project Network) has antiquated terminology and instructions.
Microsoft Entra ID is the new name for Azure AD. The names Azure Active Directory, Azure AD, and AAD are replaced with Microsoft Entra ID. Microsoft Entra is the name for the product family of identity and network access solutions. Microsoft Entra ID is one of the products within that family. https://learn.microsoft.com/en-us/entra/fundamentals/new-name#:~:text=Microsoft Entra ID is the,the products within that family.
The name change took place on July 15, 2023. https://learn.microsoft.com/en-us/azure/active-directory/fundamentals/new-name
Currently:
Directory/Tenant ID
This is a 32-character alphanumeric string. Find this in the Azure portal in Azure Active…
1 vote -
Customer feedback around the UI of Atlas - unable to edit multiple documents at once
The customer wanted me to provide product feedback around the inability to edit multiple documents at once from the Atlas UI. The customer uses a 3rd party tool to do this instead and thought it would be better and more intuitive if this was built right into the Atlas UI.
1 vote -
Stop scamming your users
How about you stop scamming your users with deceptive instruction on your download pages.
1 vote -
AWS gp3 Decouple IOPs from Disk Size
With AWS EBS gp3 volumes, IOPs and throughput can be provisioned separately from EBS storage size. While Atlas now uses gp3 volumes and provides a base throughput of 3,000 IOPs, higher throughputs are still directly tied to disk size.
We run IO-intensive workloads that require high throughput, and we have to severely over-size disks to get the needed throughput. We don't require the extra-low latency of provisioned IOPs (which is much more expensive than over-provisioning storage).
According to the linked AWS documentation below, if the M60 instance size is running on an ec2 m5.4xlarge instance type, it can support a…
4 votes -
Restore a snapshot from one organization to another organization via API call
I have tried to create automation to restore the snapshot from One organization cluster to other organization cluster. Example Prod cluster from Organization A -> Stage cluster from Organization B.
currently this method is not possible to restore a snapshot from one organization to another organization via API call as API key cannot be shared over the scope of different organization.( As per mongo atlas support Engineer)
Please enable this feature to automate the snapshot restore instead of using UI for restore.
1 vote -
Analytics node with NVMe option
We currently have use mongodb mainly to correct data and visualize with chart features. Some chart require heave disk read operation and need more performance. We need fast storage access only for analytics nodes. If we can select NVMe option only for analytics nodes, we use it but we don't want to switch whole clusters to it since main cluster don't need higher cpu/memory/stroage/iops. We just want to use NVMe only with analytics nodes. Some restriction such as min storage size for main cluster is acceptable but don't want to upgrade to higher tier of main cluster.
1 vote -
Push Logs to AWS S3 bucket: IAM role Max_Session expect 12hours
With respect to the feature "Push Logs to AWS S3 bucket " Atlas expect max-session-duration for IAM cross account role is : 43200 (12HRS) by default.
Some organizations like us have limitations on max-session. We reached out MongoDB Support, They said " The Atlas team has confirmed that currently, we do not support max session durations of less than 12 hours. However, there is an internal ticket for improvement, where session duration can be internally configurable via MongoDB's internal support UI. Case ref: 01276442This will provide a flexibility for the organizations to have max-session configured according to their internal…
1 vote -
Enable the choices of using NVMe SSDs in MS Azure from tier M60 to M40, M50, comparable to AWS.
Enable the choices of using NVMe SSDs in MS Azure from tier M60 to M40, M50, comparable to AWS.
1 vote -
Export logs to S3 at the organization level rather than project level
Currently if we have to export logs to S3 we need to enable log export to S3 at project level. If I multiple projects then same things needs to be done, there should be an option to send logs to S3 on organization level when setup should send the logs to S3 bucket with path name (projectname+'/'+clustername).
1 vote
- Don't see your idea?