Database
302 results found
-
Document known breaking change in JavaScript Engine
If a new version of MongoDB will include a new version of MozJS (ex https://jira.mongodb.org/browse/SERVER-29286) and this version will break any existing scripts it would be ideal if this was documented.
For example, with MongoDB 4.2 "value".contains() no longer worked in scripts and had to be converted to "value".includes()
2 votes -
VPC between ATLAS/MongoDB/AWS and Heroku/AWS
Would be a great great feature if we can create a VPC between ATLAS/MongoDB/AWS and Heroku/AWS.
Thanks.
Adailton Silva2 votes -
Add option to repair a single collection
Add support for repairing a single collection instead of what is currently available to repair an entire database via db.repairDatabase()
2 votes -
Aggregate $accumulator 'Combine' stage for sharded collections
After the ‘accumulate’ stage has run against each document on a shard it would be extremely useful to run a script on the shard against the state to reduce the state down to something smaller before passing the final state over the network back to mongos for the ‘merge’. (Over in the Elastic camp they call this the ‘combine_script’).
Scenario:
I have a customer orders database sharded by the Customer ID (meaning all data relating to any specific customer is kept on the same shard - so customers don’t need to be ‘merged’ between shards). I can write an accumulator…2 votes -
Assign MongoDB roles to LDAP users
I would like to create a softlab and therefore give the possibility to a user to create collections in a database which is dedicated to him.
Currently, to do that I need to :
1. create a single-member LDAP group for each user
2. map this group to a MongoDB role authorizing the user's database.Create a group with a single member is conceptually useless. I want to avoid step 1.
My feature request : add the ability to assign MongoDB roles to LDAP users as well as groups.
Regards,
Jerome2 votes -
Support for Transaction Savepoints (Nested Transactions)
The ability to partially rollback transactions to a specified point.
2 votes -
Allow collection collation to be editable
Collation of a collection can be set at creation time only. It would be useful to edit these fields to avoid creating an entire new collection and copying the collection over.
2 votes -
Add Event Stream Features (Apache Kafka, NATS Streaming)
I thought it would be great if mongodb can support an event streaming(event bus) feature.
Existing popular event streaming services(AWS Kinesis, NATS Streaming, Apache Kafka) can persist data which is somewhat like a database.Its great for debugging, data logging for later uses such as machine learning. Since I can see the full flow of the data and its changes.
But there are two downside with most even streaming services.
1. Its very difficult to query the data.
2. "Eventual Consistency" issue when dealing programmatic errors(bugs). Most event streams have this nice feature, which keeps sending the same event to…2 votes -
Performance problem with compound index using doubles, with range filtering
It looks like there may be a problem when using doubles in a compound index and using range-test filters against all of them in a query.
The problem we see from the explain plans is that far more keys are being examined that there should be and this leads to poor performance.
If we switch to using a similar setup but with integers instead, we don’t see the problem.See the second part of this ticket for full details
https://support.mongodb.com/case/00659614
We have a document that looks like this –
{
"id" : ObjectId("5e2ab24eca314f10b486d827"),
"attributes" : [
{
"attributeCode" :…2 votes -
Too many keys in 2dsphere index when used on geoJson within arrays
Mongo support have asked me to raise this bug as an improvement idea because they see it as an edge case.
The issue is here and includes more details:https://support.mongodb.com/case/00659614
In summary, if you have a geoJson sub document that sits within an array, and you index it using the 2dspehere index, you get keys duplicated for every element in the array.
For example, if you have geoJson that represents a big 4,500 vertex polygon, and you put it within an array element, if the array has only one element I see 20 keys generated, which is good. But if…
2 votes -
Making Elections Faster.
Election is an expensive operation. We can avoid the elections if we can enforce some priority-based ordering on who shall become primary if the current primary fails.
Specifically, we can keep a node with second highest priority in the data center of the most preferred primary. Usually, the network failures within the data centers are less probable than across the data centers. So, it is reasonable to believe that if the second preferred member loses the contact with current primary, it is the primary node which has failed and not the network. As a supplement, we can also have redundant…2 votes -
Option Renaming Time Series Collections
Ability to rename TS collection . We need it for migration history data to TS without downtime during conversion.
1 vote -
Support retrieving all connections info from clients to mongodb
Support retrieving all connections info from clients to mongodb, not just those running commands returned by currentOp(). OS command netstat can return all ip info, but it does not contain the relation ino between ip and database such database name and user. It's a common function in relation database, e.g. 'show processlist' command in mysql. And it's useful when we try to migrate to a new MongoDB cluster, we need to find out all the clients connected to the original mongodb.
1 vote -
Add support for all type of joins like Postgres has and improve performance
$lookup is a performance killer. Joins are crucial parts in every OLTP system. $lookup is the equivalent to join in SQL, however $lookup is slow, doesn't support hash joins or other efficient join algorithm implemented in Postgres for example.
Seems that if mongo won't add support, their DB puts behind Postgres.
1 vote -
Multiple unique single indexes in sharded collections
MongoDB sharded collection can only have one unique index which is the shard key. However, in real applications, one might need more than one single unique index. "Proxy" collections suggested in https://www.mongodb.com/docs/manual/tutorial/unique-constraints-on-arbitrary-fields/ has 3 issues:
The most important one is that it does not enforce a one-to-one relationship between the main document and the proxy document. So if I follow the example in the link, two different emails can have the same parentid. You can add a unique single index on parentid but that will not work if the proxy collection is sharded.
If I want to add…
1 vote -
Is there any limitation on size of query in pipeline (character in query, not data)
I have an aggregate query, when the character in pipeline over 21066 character, the result return wrong format that the pipeline expected. So I wonder is there any limitation on size of query in pipeline?
My case is:
Query: [
"$match":{"profileID": {"$in": ['', .....]}},
{"$group":{"_id":"$profileID","materialLevel2":{"$sum":{"$cond":[{"$and":[{"$eq":["$expenseL2","true"]},{"$eq":["$expenseL3","false"]},{"$eq":["$expenseL4","false"]}]},1,0]}},"materialLevel3":{"$sum":{"$cond":[{"$and":[{"$eq":["$expenseL3","true"]},{"$eq":["$expenseL4","false"]}]},1,0]}},"materialLevel4":{"$sum":{"$cond":[{"$eq":["$expenseL4","true"]},1,0]}}}}
]The true result is:
[{
"_id" : "000.03.50.H29",
"materialLevel2" : 1.0,
"materialLevel3" : 1.0,
"materialLevel4" : 2.0
},...]But when the length of query greater than 21066 character, the result of query is:
[{
"profileID" : "00.00.00.PTT",
"expenseL2" : "true",
"expenseL3" : "true",
"expenseL4" : "false"
},....]Thanks!
1 vote -
Support concise correlated queries for $graphLookUp
Similar to $lookUp, support concise correlated queries for $graphLookUp.
It would allow adding a $project stage to reduce overhead.
1 vote -
Capture pre-images for ONLY specified change events
Currently, pre-images are captured for all updates, replacements and deletes on a given collection.
Ask: Enable capture of pre-images for ['DELETED', ...] documents only.
Context: A customer uses a collection for ingestion, where 16-20M documents are ingested every 20 minutes. Documents are indexed with Elastic Search upon insertion, and are removed from Elastic Search upon deletion. However, the ingestion process requires subsequent operations: an average of 30 updates are made on each document, and the customer doesn't want to pay the penalty of generating pre- and post-images on updates.
Impact: Updating a collection of 80GB (20M docs of 4kb each) 30…
1 vote -
MongoDB Audit Log Missing Failed and Successful Command Execution Flag
When you initially create a database account…..this action will be logged in the database audit logs as configured, however when you try to re-attempt to create the account again….a normal error message will be displayed as shown below:
db.createUser( {user: "test111_3",pwd: "emad123",roles: [ { role: "readWrite", db: "admin" } ]})
When examining the audit logs the two entries are identical in results which shouldn’t be the case….I think the flag “result” when command executed failed should have a different value to distinguish successfully executed commands from failed executed commands:
{ "atype" : "createUser", "ts" : { "$date" : "2024-08-30T13:04:59.535+03:00" },…
1 vote -
upgrade src/third_party/gperftools ≥ 2.10
gperftools new version support new architecture
1 vote
- Don't see your idea?