Ops Tools
462 results found
-
Document list of Alert conditions that post to the Activty Feed but not Alerts page
Document list of Alert conditions that post to the Activty Feed but not Alerts page
It seems like some "non-actionable" alert/event conditions (example: Host has Restarted) post to the Activity Feed but not the Alerts Page.
Notifications are sent.
I cannot find a list of these that post only to the Activity Feed so it would be nice to have them documented.2 votes -
Ops Manager and Backup infrastracture Disaster Recovery support with K8s Operator
We have carried out tests with MongoDB v1.5.5 K8s Operator and Ops Manager 4.2.18 with Backup infrastructure (S3 Snapshots) in an Openshift 3.11 environment (MongoDB Support case attached).
In this case, a "Disaster Recovery" simulation has been carried out. However, several components created by the Operator had to be restored to obtain a similar state to the one before the "disaster".
Furthermore, it is very likely that the S3 Snapshots will be lost if the process is not completed in a certain manner.
It would be great to have an official approach to deploy/restore an OM resource using MongoDB K8s…
4 votesThere is no current supported mechanism for backing up Ops Manager in a way that guarantees the data. As Ops Manager is itself a backup tool, it's challenging to maintain the integrity of the data in DR scenarios.
For this reason we recommend multi-site high availability for OM and AppDB. This is already possible when running OM on hardware of in VMs, but not currently supported in Kubernetes (unless a Kubernetes cluster is spanning sites).
Later this year (2023) we hope to support OM deployments across multiple Kubernetes clusters - as we already support (in beta) for Replica Sets (full release in April 2023 with Sharded cluster support in May/June 2023). Doing so will reduce the criticality of a OM/AppDB backup solution within Kubernetes.
-
Encrypt Password LDAP Query User
The LDAP Query user's password appears in plain text in mongod.config file. The ability to mask its password in automation config file using credentialstool would mitigate a security risk.
9 votes -
Allow Ops Manager users to move/migrate backup job snapshots from one S3 bucket to a different S3 bucket
Ops Manager users with S3 blockstores may need to move snapshots and backup jobs to a new S3 bucket. For MongoDB blockstores, this is accomplished using a groom.
Move Blocks to a Different Blockstore
https://docs.opsmanager.mongodb.com/current/core/administration-interface/#groom-priority-pageThis feature request is to provide the same feature to groom backup snapshots/jobs to a new bucket for S3 blockstores.
42 votes -
Using Ops Manager, I want to get an alert when a collection is a certain size or near no storage so that I can do an investigation
As a DBA using Ops Manager, I want to get an alert when a collection is a certain size or near no storage so that I can do an investigation to see if there is abnormal data growth. As of right now, I can only monitor cluster data size.
2 votes -
Using Ops Manager, Tell me useful information such as "given the current 60 day growth trends, you will run out of space in 2 months"
As a DBA using Ops Manager, I want the system to tell me useful information such as "given the current 60 day growth trends, you will run out of space in 2 months" so that I can do better capacity planning.
2 votes -
I want to auto scale vertically and horizontally nodes cpu or disk in Ops Manager
As a DBA using Ops Manager, I want to auto scale vertically and horizontally nodes cpu or disk when certain thresholds are met. As of right now, there is no auto scale functionality in Ops Manager.
1 vote -
Restore a single database within a MongoDB cluster and not store the all the databases in a MongoDB
As a DBA using Ops Manager, I want to restore a single database within a MongoDB cluster and not store the all the databases in a MongoDB cluster so that my restore only impacts a part of my cluster. As of right now, it restores the entire cluster with all database in it.
2 votes -
Only download to /mms/mongodb-releases/ those versions actually in use in a deployment
/mms/mongodb-releases/ is continually updated with the most current minor release of each major version which can fill the partition.
This request is for "Only download to /mms/mongodb-releases/ those versions actually in use in a deployment"
eg if I am only using 4.2.20 in my actual deployments only download the most recent 4.2.x and not 3.6.x
of course if I do decide to deploy a 3.6.x deployment this FR would need to automatically be bypassed and download my 3.6.x and the most recent 3.6.x.
This request is similar to but not the same as another Feature Request:
Do not download EOL…3 votes -
Send `Monitoring is down` and `Backup is down` alerts for each MongoDB Agent (Monitoring/Backup Module) and include hostname information in
What is the problem that needs to be solved?
Monitoring is down
andBackup is down
alerts needs to be sent for each individual MongoDB Agent (Monitoring/Backup Module) which become down,Monitoring is down
andBackup is down
alerts should include hostname information in them.Why is it a problem? (the pain) Customer can't easily identify which MongoDB Agent (Monitoring/Backup Module) become down without hostname information available in the alert (in multi-project environment this becomes operational pain for the customer).
12 votes -
Support rolling rebuild index
In some particular situation indexes need to be rebuild.
It would be nice to support a rolling index rebuild like Ops Manager does for the standard create index.3 votes -
Add ability to have systemLog redirected to stdout (just have to remove systemLog.destination and path)
To be able to have MongoDB logs redirect to stdout and this having it into GKE CloudLogs, we should NOT configure a systemLog.destination nor a systemLog.path.
In 0.6.0 release, systemLog.destination and path are hardcoded and cannot be nullable.
see automationconfigbuilder.go at line 208:
...
process.SetSystemLog(SystemLog{
Destination: "file",
Path: path.Join(DefaultAgentLogPath, "/mongodb.log"),
})
...1 vote -
mongomirror compatibility with SRV strings
It would be ideal if the mongomirror utility could accept SRV-based connection strings for the source and destination clusters. The inability to do this can cause pain for customers. For example, this does not work:
--destination "mongodb+srv://mgo-aura-dgs-prdsrv-tk-pl-0.yrmiy.mongodb.net"
So users are stuck doing this:
--destination "atlas-nkaylx-shard-0/pl-0-ap-northeast-1.yrmiy.mongodb.net:1036,pl-0-ap-northeast-1.yrmiy.mongodb.net:1037,pl-0-ap-northeast-1.yrmiy.mongodb.net:1038"
Two issues with that:
1. It's painful looking up the host/port and atlas defined replica set name
2. The destination info is unreadable by most humans. The service name (i.e. SRV record connstring) includes the name of the cluster that the customer defined, and not the random has that Atlas generates2 votes -
Configure MongoDB Automation Agent collecting stats on some collection to not trigger alerts
We just had a support case about some alerts being raised on our cluster because the MongoDB Automation Agent collecting stats on some collection doing queries without index triggers "Scanned Objects / Returned" ratio has went over 1000.
It would be really nice to at least not raise alerts when it's the mongodb automation agent that triggered it. Were monitoring our alerts a lot and these are false positive we can't do anything about it seems other than create all the indexes it needs, which might change over time. We have no guarantee of which index it needs.
Another alternative…
9 votes -
index review before sharding with OPS Manager
Currently, when you use OPS Manager to shard a collection, it automatically creates a foreground index that exactly matches the sharding key, when such index doesn't exist. When the sharding key column(s) is(are)already prefixing an existing index, it's not sufficent.
This is dangerous in Live environments because the whole database is blocked for a long time (sharded collections are usually big collections).
So, several features could exist :
- before continuing with sharding, OPS Manager warns that it needs to create this foreground index first. You can stop if you don't agree (and create this index by yourself first).
-…5 votes -
avoid generating alert with error message if one oplog node in replicaset got rebooted.
currently if one of the node in appdb/oplogdb goes down for any reasons (for example, linux patch rebooting the node), ops manager generates alert
"Ops Manager was unable to connect to this database and run the
ping
command. The database could be down, unreachable, or running with authencation and Ops Manager does not have adequate permissions."there are still 2 other running nodes in replicaset. so this alert is misleading and generates false alarms.
2 votes -
Add compound indexes support for Ops Manager managed Sharding
What is the problem that needs to be solved? Ops Manager Automation does not take into account compound indexes (https://docs.mongodb.com/manual/core/index-compound/), e.g. if we have
{ a: 1, b: 1 }
index already exist Ops Manager will still create{ a: 1 }
index fora
Shard Key.Why is it a problem? (the pain) This creates unnecessary indexes with performance impact on the MongoDB Server process.
2 votes -
Request Logs - Add timeframe to list of options
While diagnosing historical issues "Request Logs" does not allow for selectable timeframes. As such, for these issues, logs must be manually gathered. Adding an optional timeframe will assist in diagnosing issues much faster.
3 votes -
Provide recurring/daily reporting on backup status from Ops Manager
Ops Manager should generate a recurring/daily report of the status of all backups. This report should include at least a list of successful snapshots, a list of unsuccessful snapshots (over the configured reporting period), and the latest successful snapshot for each deployment being backed up. Additionally, this report may include resource availability such as storage available for future snapshots.
40 votes -
S3 Snapshot Store Speed Test
It is often quite difficult to diagnose latency/bandwidth/generally slow S3 storage. It would be useful if Ops Manager could run a short test to show:
- How fast a single large object can be PUT and GET
- Measure parallel PUTs and GETs against test objects
- How much latency there is between Ops Manager and S3
1 vote
- Don't see your idea?