Database
298 results found
-
URL Decode `readPreferenceTags`
If a url encoded value is set to readPreferenceTags it will be ignored causing for instance connecting to analytics node impossible from tooling that correctly encodes the urls.
It can look like this for instance:
readPreferenceTags=nodeType:ANALYTICS
becomesreadPreferenceTags=nodeType%3AANALYTICS
1 vote -
the profile output in association with the $comment query operator need to have consistency across operations.
I have observed the different profile result with regard to $comment query operator in association with find and update operation respectively as follows;
find operation shows a comment on the command.filter.$comment and the command.comment field in the system.profile collection.
op: 'query',
ns: 'db101.Bets',
command: {
find: 'Bets',
filter: {
_id: ObjectId("61a9db4b3bd34e4f68fb9abc"),
'$comment': 'test-dba'
},
comment: 'test-dba',
lsid: { id: UUID("43ebee67-3184-4ede-9cee-ecca7457861a") },
'$db': 'db101'
},update operation shows a comment only on the command.q.$comment field in the system.profile collection.
op: 'update',
ns: 'db101.Bets',
command: {
q: {
_id: ObjectId("61a9db4b3bd34e4f68fb9abc"),
'$comment': 'test-dba'
},
u: { '$set': { odds: 0.5 } },
multi:…
1 vote -
Support for CentOS Stream 8 in MongoDB OPS Manager version 5.x
Per the Server Support Matrix https://www.mongodb.com/try/download/ops-manager support for CentOS Stream 8 is not al.
We would like to see the currently supported MongoDB OPS Manager version 5.x available on the CentOS Stream 8 distribution.1 vote -
Please report mixed-type numeric _id fields in $merge stage error
Posting this idea at the request of one of the Jira users. You can find more technical details about this in the Jira issue:
https://jira.mongodb.org/browse/SERVER-61613
The gist of it is that I may have two collections,
b2
andb3
, that are not distinguishable in the Mongo Shell, like this:db.b2.find() [ { _id: 1, created: ISODate("2021-11-18T23:16:33.149Z") }, { _id: 2, created: ISODate("2021-11-18T23:16:33.149Z") } ] db.b3.find() [ { _id: 1, created: ISODate("2021-11-18T22:53:02.113Z") }, { _id: 2, created: ISODate("2021-11-18T22:53:02.113Z") } ] ``` When I merge each into a collection `pg` with this syntax:
db.pg.aggregate([{$merge: {into: "b3", whenMatched: "merge", whenNotMatched: "fail"}} ]);…
1 vote -
multiple centersphere as a geometry for geowithin
im looking at one of my queries that a system regularly runs and some times we look for records that are within up to 250 different centerspheres, i wornder if we could enable, like geowithin has the ability to support multiple poligons, we could enable multiple centerspheres ,
$match: {{'location':{
"$geoWithin" : {
"$centerSphere" : [[ 14.4321, -9.4321], 2.5232135647961246e-05]
}
}}, {'location':{
"$geoWithin" : {
"$centerSphere" : [[ 14.4321, -9.4321], 2.5232135647961246e-05]
}
}}, ...}
we could do :
```
$match: {'location':{
"$geoWithin" : {
"$centerSpheres" : [
[[ 14.4321, -9.4321], 2.5232135647961246e-05] ,
[[ 14.4321, -9.4321], 2.5232135647961246e-05],
...
]} }}
1 vote -
There is a specific collection that I need more performance than others. Is there a way to assign more ram/memory to a specific collection?
There is a specific collection that I need more performance than others. Is there a way to assign more ram/memory to a specific collection?
1 vote -
$populate stage
Please provide a $populate stage that allows to resolve single referenced documents.
Internally it could use the combination of $lookup and $unwind:Related:
https://stackoverflow.com/questions/37793844/mongodb-how-to-resolve-dbref-on-client-side
1 vote -
Reserve connections to admin users
When the max number of connections is reached, no one can login to the database until some connections are closed or a failover is triggered (killing all the connections).
The admin users should have a few reserved connections so that they can login to the database and take actions, like kill some connections.
1 vote -
Mongo replicaset init sync issue
When we have large size mongo replica set, we may have to take out each nodes for maintenance for few hrs or days.
In that case we have to increase size of Oplog to keep several days of transactions to resync nodes after maintanance.
If the RS headroom falls below oplog window, then we have blow away data and do init sync to add node back after maintanence.
The problem with very large size Oplog, will slow down any Change Stream process. Also it will occupy un neccasary space within collections.
Can mongo offer alternative way to constantly dump oplog…
1 vote -
Shard Drain/ Removal issue
If we have more shards, and if want to remove few shards (more than 1), we use below command.
db.adminCommand( { removeShard : "Shardname" } )
Ex, if I have Shard 1, 2, 3, 4, 5. want to remove Shard 2 & 5.
I want to remove one shard at a time to minimize impact to users , then want to remove, say Shard 2, then want to remove Shard 5.
If we do this, some chunks from Shard#2 also get moved to "Shard#5", which is suppose to removed later. This causes Shard#5 chunk size increase. Then takes more time.
…
1 vote -
Implement read-only fields and documents in database
I haven't tracked down this functionality yet, so excuse me if it is already implemented.
As a developer and system administrator I came to a dilemma how to prevent myself to make changes to documents.
I can limit myself as a programmer, but as a administrator I can always logon to console and make changes ''by hand''.
I was thinking of read-only field type, which once set, could not be updated nor removed from document (although backup and restore is the first problem which comes to mind). This limit should be set on a database level. Read-only field could typically…
1 vote -
The ability to perform a quick rollback or rewind of the database
This is along the lines of a flashback, to a previous point in time say from a very impacting change to data.. A large delete or data modification event. This avoids the need to take a complete outage for hours and hours restoring TBs of data and re-applying change logs. Would be great if this could be done to the granularity of a single or multiple collections too. This could use the oplogs present locally or in the oplog store.
1 vote -
`Blocking` Flag for Explain Output
It's not always obvious which aggregation stages are blocking or non-blocking when viewing
explain
output for an aggregation command. It would be helpful to include ablocking
flag for each stage of anaggregate
command in itsexplain
output.1 vote -
restricted mode for database
database in restricted mode can very helpful .. so we can carry out , lock users out, rebuild indexes, compact and other admin tasks.
1 vote -
Allow changing config values without restart
It would be great if configuration changes could be effected without needed to restart nodes.
For example, audit filter, enabling/disabling different security mechanisms.
This would be especially useful for Atlas and clients with large clusters where restarts will cause a performance deficit due to cold cache after node restarts.
1 vote -
Provide straightforward syntax for 1-to-1 joins in aggregation
The syntax for joins that bring back multiple documents from foreign collections is very straightforward and yields exactly what one would expect, but simple joins that are bread and butter in SQL require very convoluted and expensive to run syntax.
Consider a product database that has
products
,categories
andreviews
collections. Each product has a unique category and may have multiple reviews. Getting all reviews in an aggregation is very straightforward (top stage), but getting categories, similar to SQL, is as convoluted as it gets (bottom stage).
…db.products.aggregate( [ // // Document aggregates naturally aggregate foreign documents // into
1 vote -
MultiTenant Abstraction
Just as a time-series collection in MongoDB 5.0 abstracts the underlying implementation of the bucketing pattern, customers who implement a multi-tenant model through separate databases per tenant run into the issue of too many dhandles and would benefit from an abstraction of the implementation of a collection-with-discriminator-field - they would be able to meet their internal compliance requirement of separate databases per customer, and would have minimal changes to their code.
1 vote -
Retrywrite error fixed in Mongodb 4.4
We have test the retry write in Atlas by using mongodb java driver in Mongodb 4.4 cluster. The error is similar to SERVER-53624(https://jira.mongodb.org/browse/SERVER-53624). The supporter respond the error will fix in Mongodb 5.0. We hope this error can also fixed in Mongodb 4.4
1 vote -
Make targeted query to a specific shard without using the shard key as part of query.
As of right now you need to use the shard key as part of the query to make a target query to a specific shard. Would like the ability to make targeted query to a specific shard without using the shard key as a part of the query.
Maybe one way of doing this is using index metadata to avoid scatter gather query and using that index meta data to instead make targeted queries on sharded clusters.
1 vote -
Easy Paging with offset
Paging is a common functionality for REST APIs. When implementing paging queries for large datasets skip and limit is often not an option. When using query filters with non unique fields, such as creation date there can be problems such as duplicate entities on subsequent pages. It would be great to be able to additionaly pass in an offset document reference, that would be used as a start in case the filter does not lead to a unique starting point.
so for example
db.test.find({"creationDate" : {$gte:ISODate('2021-08-27T07:25:00Z')}, {"offset":<ObjectId>).sort({"creationDate":1}).limit(20)1 vote
- Don't see your idea?