Database
287 results found
-
1 vote
-
Change how document size is calculated
Disclaimer: I'm aware that this is a pretty big feature request.
Currently a document has a max size of 16 mb. I'm not suggesting changing that. However, I'm suggesting changing how that size is calculated.
I'm designing a database in which documents have a tree structure. The document will have an array called "nodes", which can hold any number of child nodes. Those nodes can also have arrays of nodes, and so on. It's a simple tree structure.
For major users of my application, that outer document could get pretty big. It's not a problem right now (no users, no…
1 vote -
Multi-version-intermediate upgrade package
When upgrading the MongoDB server community edition from a very old version to the newest version one cannot skip installing intermediate versions, e. g. version 3.6 -> 4.4 does not work because the database files cannot be auto-migrated and the FeatureCompatibilityVersion is too low on 3.6 (or not existent).
Instead we have to install 3.6 -> 4.0 -> 4.2 -> 4.4 and execute db.adminCommand( { setFeatureCompatibilityVersion: "..." } ) appropriately in between installing each new version.
I propose creating an additional installer package for Linux which takes care of all that to migrate the internal database structure from the installed…
1 vote -
Have an API to have the History of Primary Nodes
Have an API to have the History of Primary Nodes,from the time the Replica set is initiated.I know we can have alerts for Primary switch over,but if we want to analyse the data of elections/primary node for say about last 1 year , an API would help.
1 vote -
1 vote
-
Please create a build in role that grants developers all permissions needed to add/edit schema validations short of adminAll.
Please create a built-in role that grants developers all permissions needed to add/edit schema validations short of dbAdminAnyDatabase.
1 vote -
Add option to $dayOfWeek to choose between Monday and Sunday
Hi!
I was wondering if you could add an optional parameter to $dayOfWeek that allows you to choose on which day you want the functionality to start counting.
Thanks!
1 vote -
Add regex support in pipeline operator `replaceAll`
It would be very nice to have something like this possible:
{ $replaceAll: { input: "$text", find: "/[;,.]/g", replacement: "." } }
Many thanks !
1 vote -
URL Decode `readPreferenceTags`
If a url encoded value is set to readPreferenceTags it will be ignored causing for instance connecting to analytics node impossible from tooling that correctly encodes the urls.
It can look like this for instance:
readPreferenceTags=nodeType:ANALYTICS
becomesreadPreferenceTags=nodeType%3AANALYTICS
1 vote -
the profile output in association with the $comment query operator need to have consistency across operations.
I have observed the different profile result with regard to $comment query operator in association with find and update operation respectively as follows;
find operation shows a comment on the command.filter.$comment and the command.comment field in the system.profile collection.
op: 'query',
ns: 'db101.Bets',
command: {
find: 'Bets',
filter: {
_id: ObjectId("61a9db4b3bd34e4f68fb9abc"),
'$comment': 'test-dba'
},
comment: 'test-dba',
lsid: { id: UUID("43ebee67-3184-4ede-9cee-ecca7457861a") },
'$db': 'db101'
},update operation shows a comment only on the command.q.$comment field in the system.profile collection.
op: 'update',
ns: 'db101.Bets',
command: {
q: {
_id: ObjectId("61a9db4b3bd34e4f68fb9abc"),
'$comment': 'test-dba'
},
u: { '$set': { odds: 0.5 } },
multi:…
1 vote -
Support for CentOS Stream 8 in MongoDB OPS Manager version 5.x
Per the Server Support Matrix https://www.mongodb.com/try/download/ops-manager support for CentOS Stream 8 is not al.
We would like to see the currently supported MongoDB OPS Manager version 5.x available on the CentOS Stream 8 distribution.1 vote -
Please report mixed-type numeric _id fields in $merge stage error
Posting this idea at the request of one of the Jira users. You can find more technical details about this in the Jira issue:
https://jira.mongodb.org/browse/SERVER-61613
The gist of it is that I may have two collections,
b2
andb3
, that are not distinguishable in the Mongo Shell, like this:db.b2.find() [ { _id: 1, created: ISODate("2021-11-18T23:16:33.149Z") }, { _id: 2, created: ISODate("2021-11-18T23:16:33.149Z") } ] db.b3.find() [ { _id: 1, created: ISODate("2021-11-18T22:53:02.113Z") }, { _id: 2, created: ISODate("2021-11-18T22:53:02.113Z") } ] ``` When I merge each into a collection `pg` with this syntax:
db.pg.aggregate([{$merge: {into: "b3", whenMatched: "merge", whenNotMatched: "fail"}} ]);…
1 vote -
multiple centersphere as a geometry for geowithin
im looking at one of my queries that a system regularly runs and some times we look for records that are within up to 250 different centerspheres, i wornder if we could enable, like geowithin has the ability to support multiple poligons, we could enable multiple centerspheres ,
$match: {{'location':{
"$geoWithin" : {
"$centerSphere" : [[ 14.4321, -9.4321], 2.5232135647961246e-05]
}
}}, {'location':{
"$geoWithin" : {
"$centerSphere" : [[ 14.4321, -9.4321], 2.5232135647961246e-05]
}
}}, ...}
we could do :
```
$match: {'location':{
"$geoWithin" : {
"$centerSpheres" : [
[[ 14.4321, -9.4321], 2.5232135647961246e-05] ,
[[ 14.4321, -9.4321], 2.5232135647961246e-05],
...
]} }}
1 vote -
There is a specific collection that I need more performance than others. Is there a way to assign more ram/memory to a specific collection?
There is a specific collection that I need more performance than others. Is there a way to assign more ram/memory to a specific collection?
1 vote -
$populate stage
Please provide a $populate stage that allows to resolve single referenced documents.
Internally it could use the combination of $lookup and $unwind:Related:
https://stackoverflow.com/questions/37793844/mongodb-how-to-resolve-dbref-on-client-side
1 vote -
Reserve connections to admin users
When the max number of connections is reached, no one can login to the database until some connections are closed or a failover is triggered (killing all the connections).
The admin users should have a few reserved connections so that they can login to the database and take actions, like kill some connections.
1 vote -
Mongo replicaset init sync issue
When we have large size mongo replica set, we may have to take out each nodes for maintenance for few hrs or days.
In that case we have to increase size of Oplog to keep several days of transactions to resync nodes after maintanance.
If the RS headroom falls below oplog window, then we have blow away data and do init sync to add node back after maintanence.
The problem with very large size Oplog, will slow down any Change Stream process. Also it will occupy un neccasary space within collections.
Can mongo offer alternative way to constantly dump oplog…
1 vote -
Shard Drain/ Removal issue
If we have more shards, and if want to remove few shards (more than 1), we use below command.
db.adminCommand( { removeShard : "Shardname" } )
Ex, if I have Shard 1, 2, 3, 4, 5. want to remove Shard 2 & 5.
I want to remove one shard at a time to minimize impact to users , then want to remove, say Shard 2, then want to remove Shard 5.
If we do this, some chunks from Shard#2 also get moved to "Shard#5", which is suppose to removed later. This causes Shard#5 chunk size increase. Then takes more time.
…
1 vote -
Implement read-only fields and documents in database
I haven't tracked down this functionality yet, so excuse me if it is already implemented.
As a developer and system administrator I came to a dilemma how to prevent myself to make changes to documents.
I can limit myself as a programmer, but as a administrator I can always logon to console and make changes ''by hand''.
I was thinking of read-only field type, which once set, could not be updated nor removed from document (although backup and restore is the first problem which comes to mind). This limit should be set on a database level. Read-only field could typically…
1 vote -
The ability to perform a quick rollback or rewind of the database
This is along the lines of a flashback, to a previous point in time say from a very impacting change to data.. A large delete or data modification event. This avoids the need to take a complete outage for hours and hours restoring TBs of data and re-applying change logs. Would be great if this could be done to the granularity of a single or multiple collections too. This could use the oplogs present locally or in the oplog store.
1 vote
- Don't see your idea?