Database
296 results found
-
Capture pre-images for ONLY specified change events
Currently, pre-images are captured for all updates, replacements and deletes on a given collection.
Ask: Enable capture of pre-images for ['DELETED', ...] documents only.
Context: A customer uses a collection for ingestion, where 16-20M documents are ingested every 20 minutes. Documents are indexed with Elastic Search upon insertion, and are removed from Elastic Search upon deletion. However, the ingestion process requires subsequent operations: an average of 30 updates are made on each document, and the customer doesn't want to pay the penalty of generating pre- and post-images on updates.
Impact: Updating a collection of 80GB (20M docs of 4kb each) 30…
1 vote -
Allow to decrease time series granularity and custom bucketing values
In our IoT use case, we are leveraging MongoDB’s time series functionality. Due to high write volume, we need to adjust the timeseries.granularity and bucketMaxSpanSeconds parameters to manage the write load. However, after increasing the bucketMaxSpanSeconds, we need to run the system for several days to observe stability. If the value is set too high, MongoDB does not support decreasing the bucketing value, and we are forced to create a new collection instead.
It would greatly simplify our testing process and increase flexibility in adapting to business changes if MongoDB allowed the decrease of bucketMaxSpanSeconds after it has been increased.
2 votes -
Making operations that hang indefinitely more visible
Hi Team,
It would be useful for some kind of timeout message to be implemented if an operation is waiting on commitQuorum - say wait an hour and then return an error message with "Can't perform operation, required number of commitQuorum members not available to process this operation" or something similar. I don't think a command should hang indefinitely, and it can also help both customers and support for the analysis of this kind of issues.
Thanks in advance
2 votes -
Metadata (Created Date) for database accounts
We frequently have audit questions about when a given user was created in our Mongo databases. Other DBMSs (Oracle, for instance) have a 'CREATED' field in a metadata table (i.e. DBA_USERS) that shows when a user was created, and this would be very helpful when responding to these audits.
2 votes -
connection
Hi,
it will be greatly appreciated to have an option to control the connections number when choosing an instance.
M30 is just what we need except for the connections number
we have around ~2.5k and max is 3k it would help a lot to have an option to increase just the connections number, so we wont need to move to M403 votes -
Is there any limitation on size of query in pipeline (character in query, not data)
I have an aggregate query, when the character in pipeline over 21066 character, the result return wrong format that the pipeline expected. So I wonder is there any limitation on size of query in pipeline?
My case is:
Query: [
"$match":{"profileID": {"$in": ['', .....]}},
{"$group":{"_id":"$profileID","materialLevel2":{"$sum":{"$cond":[{"$and":[{"$eq":["$expenseL2","true"]},{"$eq":["$expenseL3","false"]},{"$eq":["$expenseL4","false"]}]},1,0]}},"materialLevel3":{"$sum":{"$cond":[{"$and":[{"$eq":["$expenseL3","true"]},{"$eq":["$expenseL4","false"]}]},1,0]}},"materialLevel4":{"$sum":{"$cond":[{"$eq":["$expenseL4","true"]},1,0]}}}}
]The true result is:
[{
"_id" : "000.03.50.H29",
"materialLevel2" : 1.0,
"materialLevel3" : 1.0,
"materialLevel4" : 2.0
},...]But when the length of query greater than 21066 character, the result of query is:
[{
"profileID" : "00.00.00.PTT",
"expenseL2" : "true",
"expenseL3" : "true",
"expenseL4" : "false"
},....]Thanks!
1 vote -
Named MongoDB Connections
When a service is acting erroneously and generating hundreds or thousands of connections, it's currently difficult to determine which service is doing so when you have 30+ services connecting to MongoDB.
My proposal is that we should be optionally able to specify a non-unique name for the connection in the MongoDB URL (possibly after a #), which would allow DB administrators to see how many of each named connection was connected at any given time, and also other metrics (operations/s per name, average/max query execution time per name, etc.).
Example URL:
mongodb://user:****@clustername.abcd.mongodb.net:27017/dbName?authSource=admin#myServiceName2 votes -
log redaction customization
If we have customizations on the log redaction feature then we can add our preference where we have to apply log redaction. As of now according to my opinion it is getting applied on the all log fields which may create a issue while troubleshooting the any issue.
2 votes -
user authentication
Hy
It would be extremely useful to be able to create users who can only connect to the database from specific networks or even specific IP addresses, similar to what is possible with MySQL.
For example, using the following commands:
CREATE USER 'user_name'@'10.214.3.0' IDENTIFIED BY 'password';
GRANT ALL PRIVILEGES ON shorturl.* TO 'user_name'@'10.214.3.0';
You can create a user who can access the database only from the network with the IP address 10.214.3.0.
I would like to know if it is possible to achieve similar functionality in mongodb as well. This would be very useful for my purposes, as I want…
3 votes -
Support concise correlated queries for $graphLookUp
Similar to $lookUp, support concise correlated queries for $graphLookUp.
It would allow adding a $project stage to reduce overhead.
1 vote -
2 votes
-
MongoDB Audit Log Missing Failed and Successful Command Execution Flag
When you initially create a database account…..this action will be logged in the database audit logs as configured, however when you try to re-attempt to create the account again….a normal error message will be displayed as shown below:
db.createUser( {user: "test111_3",pwd: "emad123",roles: [ { role: "readWrite", db: "admin" } ]})
When examining the audit logs the two entries are identical in results which shouldn’t be the case….I think the flag “result” when command executed failed should have a different value to distinguish successfully executed commands from failed executed commands:
{ "atype" : "createUser", "ts" : { "$date" : "2024-08-30T13:04:59.535+03:00" },…
1 vote -
upgrade src/third_party/gperftools ≥ 2.10
gperftools new version support new architecture
1 vote -
Providing connection details with username for real time connection monitoring.
Currently, mongd.log provides connection details like remote IP, source port with authentication information but doesn't provide the connection is in active state or not.
serverStatus() only provides number of current and active connections.Example:
10.0.0.100:12345 - username active
10.0.0.101:12346 - username idle2 votes -
Support hex conversion and binary operators in aggregation framework
The aggregation framework lacks function for HEX <-> Number conversion and binary operations. Such functions are common in almost every programming language.
Typically and most required operations would be:
- convert Int/Long to Hex
- convert Hex to Int/Long
- Create BinData from Hex string
- Get Hex string of BinData
- Binary AND
- Binary OR
- Binary XOR
- Binary NOT (i.e. complement)
Some of these functions are available in mongo shell or in Bitwise Query Operators but they do not exist in aggregation framework.
For details see https://jira.mongodb.org/browse/SERVER-55386
This request becomes more urgent, because starting in MongoDB 8.0, server-side JavaScript functions ($accumulator, $function, $where) are…
1 vote -
$sort should allow 0 as argument meaning "no sort"
The $sort operator and aggregation stage should allow 0 (possibly also null) as an argument for the field sort order, meaning, "Don't sort."
This allows a variable to be passed in that can conditionally skip the sort operation, in addition to the present specification, in which a variable can only choose ascending or descending sort.1 vote -
Easier way to troubleshoot storage use size discrepancy across nodes in the same replica set
While initial sync may potentially help on this topic, it would be great if the product has any easier way to identify the cause of significantly different storage use size (so to give a better confidence if initial sync is going to help).
1 vote -
Feature to perform Machine Learning predictive analysis and classification in MongoDB
I want to bring the machine learning compute and predictive analysis into MongoDB atlas. Instead of ETL my data out of Atlas to achieve this, I will reduce my architectural complexity by having an aggregation operator that does this on my documents stored in Atlas.
1 vote -
Export Backup Snapshots to GCP Bucket
As Atlas user, i want to Export Backup Snapshots to a GCP bucket because i don't have an AWS subscription.
2 votes -
Parallelize unionWith
Today $unionWith aggregation command is executed sequentially. EG first we query collection A and then collection B and then the union occurs.
The process should be parallelized so the query part will run in parallel while the union will be done as best effort tree merge try to speed up the overall Elapsed Time of the query1 vote
- Don't see your idea?