Skip to content

Atlas

Share your idea. In order to help prioritize, please include the following information

  1. A brief description of what you are looking to do
  2. How you think this will help
  3. Why this matters to you

  • Hot ideas
  • Top ideas
  • New ideas
  • My feedback

181 results found

  1. Ability to "Mass Kill" slow running queries

    Currently, Atlas has a "Kill Op" option which is useful to kill single long-running queries.

    When upgrading to MongoDB 7.0, we were faced with a situation where the Slot-Based Query Engine (SBE) was causing 1000s of queries to execute slowly, we wanted to kill them all, but it was more than a human could do by clicking "Kill Op" 1-by-1. Hence a "Mass Kill" feature which kills queries longer than X seconds (X is configurable) would have helped us greatly in an outage scenario. We ultimately rebooted our cluster to kill queries, then manually implemented a script which did this…

    7 votes
    How important is this to you?
  2. Last Database User Access

    My team has found ourselves in the position of needing to iterate through the Database Access History Logs in order to discover whether or not a database user is still being used for the past t time periods.

    We need to have this functionality in order to guarantee a seamless credential rotation.

    With the current API this operation will require iterating over each cluster's access logs for the past t time periods. If I have "n" clusters and "t" time periods this is an O(n * t) runtime. I'm not really great at O() notation, but if t is greater…

    7 votes
    How important is this to you?
  3. Alert for WiredTiger Cache

    Hi,

    Can you please create an alert for WiredTiger metrics, such as used cache?

    We had several cluster instances going over the 5% of used cached (dirty data) and would like to be notified when it happens.

    Regards,
    Sergei

    This is needed in order to determine whether

    7 votes
    How important is this to you?
  4. Export Aggregation Results as Metrics to Prometheus

    Add support for exporting MongoDB aggregation results as Prometheus metrics. This would allow users to track custom queries and dynamic data, enabling more granular and meaningful monitoring and alerting in Prometheus and Grafana.

    6 votes
    How important is this to you?
  5. Integrar la alerta de Replication lag de Atlas en el API de Prometheus

    Se solicita incluir métricas en el API de prometheus para poder implementar la alarma de Replication lag

    6 votes
    How important is this to you?
  6. Trigger execution details

    It would be very useful to have some details provided in Atlas to trace the execution of the trigger, like the _id and operation type at a minimum. When a trigger executes the only information we see in Atlas is [
    "Sent to AWS EventBridge"
    ].

    6 votes
    How important is this to you?
  7. Improve "Host has index suggestions"" alert by including the hostname

    The "Host has index suggestions" alert notification does not contain any reference to the host/deployment that triggered it making it hard to figure out where it came from. A hostname reference would be very helpful.

    6 votes
    How important is this to you?
  8. Atlas activity feed

    We want option to export Atlas activity feed in a file to export and work with it. Parse it etc ...

    Even better, make a collection with it.

    6 votes
    How important is this to you?
  9. Allow Atlas to send additional metrics to New Relic via integration

    This is a request to add some additional metrics to those that are currently sent from Atlas to New Relic when the integration is enabled.

    The specific metrics of interest here are:
    - Disk Space (%) Used (or Free)
    - DB Storage
    - Query Targeting

    Having pointed out the above metrics specifically, I would also say: the closer that the metrics sent to New Relic via the Atlas integration can resemble those available in the Atlas cluster Metrics page, the better.

    6 votes
    How important is this to you?
  10. Add replica set name to cluster info main page

    It would be nice to get quick access to the name of the replica set from the cluster overview page.

    6 votes
    How important is this to you?
  11. Atlas metrics granularity after 48 hours

    For metrics older than 48 hours, the data is presented in 1-hour intervals. This level of granularity is often too coarse for a thorough examination of past events and trends. Such a broad view can obscure smaller yet significant details critical for understanding and resolving performance issues that occurred in the past.

    Suggested Improvement:

    having a smaller granularity value for historical metrics beyond the 48-hour timeframe. Providing data in smaller intervals would greatly enhance our ability to conduct in-depth analyses and diagnose past performance issues accurately. This would be particularly beneficial for conducting detailed investigations of historical data and identifying…

    5 votes
    How important is this to you?
  12. Persist setProfilingLevel Setting in Atlas

    Currently the profiling level resets to 0 every time the mongod process restarts. You can persist this with a self-hosted mongo by editing operationProfiling.mode in the conf file, but this is not possible in Atlas.

    5 votes
    How important is this to you?
  13. CPU Steal % on Azure

    The CPU Steal % metric and alert seems to be currently implemented only for AWS. Please implement it also on Azure, as there is no easy way currently to get alerted if the underlying burstable VM (e.g. for M20) runs out of credits ...

    5 votes
    How important is this to you?
  14. Monitoring for WiredTiger data handles

    Add process measurements for monitoring WiredTiger data handles.

    From Atlas documentation https://docs.atlas.mongodb.com/reference/atlas-limits/#collection-and-index-limits :

    While there is no hard limit on the number of collections in a single cluster, the performance of a cluster might degrade if it serves a large number of collections and indexes. Larger collections have a greater impact on performance.

    but now there is no way to get this information except db.serverStatus()

    5 votes
    How important is this to you?
  15. track collection size

    I would like to be able to track the size of our collections (i.e. # of records, total disk size) over time. It would be ideal if there was an automated way to do this with Atlas, perhaps with Charts ?

    5 votes
    How important is this to you?
  16. Add horizontal scroll bar to metrics tab

    I can't see metrics for all of my nodes while in the metrics tabs unless I make the window very large. It would be much easier to analyze the health of my cluster if I could review metrics for all the nodes by scrolling horizontally, rather than having to switch between a grouping of nodes.

    5 votes
    How important is this to you?
  17. Use 24-hour time format in Metrics date pickers

    Timestamps in Metrics plots are displayed in 24-hour time format, but the date/time filter inputs require 12-hour time (using am/pm).

    screenshot: https://p-37FYgJ.b1.n0.cdn.getcloudapp.com/items/L1upe7Wv/Image%202020-06-25%20at%204.28.22%20PM.png?v=ec3b0551ea4035a207a90ab83817bba6

    12-hour time is not widely used around the world, and it is cumbersome to use in the date pickers especially give that other parts of the UI use 24-hour time.

    Can we update these date pickers to use 24-hour time instead?

    5 votes
    How important is this to you?
  18. Allow NewRelic integration to use EU APIs instead of default US APIs for clusters in EU region

    The NewRelic integration in MongoDB is using the default NewRelic API that is in non-EU region. There is a separate API endpoint for EU region, that is described at https://docs.newrelic.com/docs/using-new-relic/welcome-new-relic/get-started/introduction-eu-region-data-center#.

    This is causing 'newRelicLicenseKey' failed to authenticate error for the MongoDB Atlas cluster setup in EU region.

    5 votes
    How important is this to you?
  19. Allow to create Dashboard

    Please provide feature to create a dashboard for metrics like CPU, memory consumption etc for a cluster or more than one cluster in single graphs like for cpu one graph, for memory one graph like that just like monitoring tools provide for example zabbix screens and dashboards, stackdriver dashboards, dynatrace dashboards.

    This is to have a glance on our entire architecture instead of checking each and every cluster every time. This will be particularly useful during load tests or business peak period where we need to check more than one cluster performance.

    5 votes
    How important is this to you?
  20. Metric reporting private endpoint state

    On Mongo Atlas platform we are able to see the status of both Atlas Private Endpoint and Azure Private Endpoint. It would be helpful to have these statuses available as a metric on the prometheus integration.

    4 votes
    How important is this to you?
  • Don't see your idea?

Atlas

Categories

Feedback and Knowledge Base