Database
302 results found
-
MongoDB should have a database backup system commit base for large self-management databases, like git commit.
There is already a mongodb backup tool mongodump but it is not a good solution for backing up large databases. We thought of making a CMS Platform with mongodb. Billions of websites can be created there. For such a large platform, the mongodump backup tool does not seem sufficient.
We want a solution where Mongodb will manage a different backup directory their data will not be deleted even run the delete command. like blockchain. If we run the delete command it will delete the data from the main database but the secondary backup database will declare it as a delete…
3 votes -
Add support for Atlas Search via Stable API
In the aggregation pipeline, the $search stage is very helpful in developing my organization's application's functionalities. But using the Stable API or Versioned API, as it was once called, I am unable to use the $search stage as it is a part of Atlas Search and not a core MongoDB Server function. I would like to request that the support for Atlas Search be added to Stable API. This will allow my organization to seamlessly upgrade the server without breaking functionality. Other workarounds are not possible for us. Please do consider this as a feature for current or future versions…
2 votes -
Option Renaming Time Series Collections
Ability to rename TS collection . We need it for migration history data to TS without downtime during conversion.
1 vote -
Making operations that hang indefinitely more visible
Hi Team,
It would be useful for some kind of timeout message to be implemented if an operation is waiting on commitQuorum - say wait an hour and then return an error message with "Can't perform operation, required number of commitQuorum members not available to process this operation" or something similar. I don't think a command should hang indefinitely, and it can also help both customers and support for the analysis of this kind of issues.
Thanks in advance
3 votes -
Capture pre-images for ONLY specified change events
Currently, pre-images are captured for all updates, replacements and deletes on a given collection.
Ask: Enable capture of pre-images for ['DELETED', ...] documents only.
Context: A customer uses a collection for ingestion, where 16-20M documents are ingested every 20 minutes. Documents are indexed with Elastic Search upon insertion, and are removed from Elastic Search upon deletion. However, the ingestion process requires subsequent operations: an average of 30 updates are made on each document, and the customer doesn't want to pay the penalty of generating pre- and post-images on updates.
Impact: Updating a collection of 80GB (20M docs of 4kb each) 30…
1 vote -
Support retrieving all connections info from clients to mongodb
Support retrieving all connections info from clients to mongodb, not just those running commands returned by currentOp(). OS command netstat can return all ip info, but it does not contain the relation ino between ip and database such database name and user. It's a common function in relation database, e.g. 'show processlist' command in mysql. And it's useful when we try to migrate to a new MongoDB cluster, we need to find out all the clients connected to the original mongodb.
1 vote -
Allow to decrease time series granularity and custom bucketing values
In our IoT use case, we are leveraging MongoDB’s time series functionality. Due to high write volume, we need to adjust the timeseries.granularity and bucketMaxSpanSeconds parameters to manage the write load. However, after increasing the bucketMaxSpanSeconds, we need to run the system for several days to observe stability. If the value is set too high, MongoDB does not support decreasing the bucketing value, and we are forced to create a new collection instead.
It would greatly simplify our testing process and increase flexibility in adapting to business changes if MongoDB allowed the decrease of bucketMaxSpanSeconds after it has been increased.
2 votes -
Add support for all type of joins like Postgres has and improve performance
$lookup is a performance killer. Joins are crucial parts in every OLTP system. $lookup is the equivalent to join in SQL, however $lookup is slow, doesn't support hash joins or other efficient join algorithm implemented in Postgres for example.
Seems that if mongo won't add support, their DB puts behind Postgres.
1 vote -
Metadata (Created Date) for database accounts
We frequently have audit questions about when a given user was created in our Mongo databases. Other DBMSs (Oracle, for instance) have a 'CREATED' field in a metadata table (i.e. DBA_USERS) that shows when a user was created, and this would be very helpful when responding to these audits.
2 votes -
Multiple unique single indexes in sharded collections
MongoDB sharded collection can only have one unique index which is the shard key. However, in real applications, one might need more than one single unique index. "Proxy" collections suggested in https://www.mongodb.com/docs/manual/tutorial/unique-constraints-on-arbitrary-fields/ has 3 issues:
The most important one is that it does not enforce a one-to-one relationship between the main document and the proxy document. So if I follow the example in the link, two different emails can have the same parentid. You can add a unique single index on parentid but that will not work if the proxy collection is sharded.
If I want to add…
1 vote -
connection
Hi,
it will be greatly appreciated to have an option to control the connections number when choosing an instance.
M30 is just what we need except for the connections number
we have around ~2.5k and max is 3k it would help a lot to have an option to increase just the connections number, so we wont need to move to M403 votes -
4 votes
-
Is there any limitation on size of query in pipeline (character in query, not data)
I have an aggregate query, when the character in pipeline over 21066 character, the result return wrong format that the pipeline expected. So I wonder is there any limitation on size of query in pipeline?
My case is:
Query: [
"$match":{"profileID": {"$in": ['', .....]}},
{"$group":{"_id":"$profileID","materialLevel2":{"$sum":{"$cond":[{"$and":[{"$eq":["$expenseL2","true"]},{"$eq":["$expenseL3","false"]},{"$eq":["$expenseL4","false"]}]},1,0]}},"materialLevel3":{"$sum":{"$cond":[{"$and":[{"$eq":["$expenseL3","true"]},{"$eq":["$expenseL4","false"]}]},1,0]}},"materialLevel4":{"$sum":{"$cond":[{"$eq":["$expenseL4","true"]},1,0]}}}}
]The true result is:
[{
"_id" : "000.03.50.H29",
"materialLevel2" : 1.0,
"materialLevel3" : 1.0,
"materialLevel4" : 2.0
},...]But when the length of query greater than 21066 character, the result of query is:
[{
"profileID" : "00.00.00.PTT",
"expenseL2" : "true",
"expenseL3" : "true",
"expenseL4" : "false"
},....]Thanks!
1 vote -
Named MongoDB Connections
When a service is acting erroneously and generating hundreds or thousands of connections, it's currently difficult to determine which service is doing so when you have 30+ services connecting to MongoDB.
My proposal is that we should be optionally able to specify a non-unique name for the connection in the MongoDB URL (possibly after a #), which would allow DB administrators to see how many of each named connection was connected at any given time, and also other metrics (operations/s per name, average/max query execution time per name, etc.).
Example URL:
mongodb://user:****@clustername.abcd.mongodb.net:27017/dbName?authSource=admin#myServiceName2 votes -
log redaction customization
If we have customizations on the log redaction feature then we can add our preference where we have to apply log redaction. As of now according to my opinion it is getting applied on the all log fields which may create a issue while troubleshooting the any issue.
2 votes -
user authentication
Hy
It would be extremely useful to be able to create users who can only connect to the database from specific networks or even specific IP addresses, similar to what is possible with MySQL.
For example, using the following commands:
CREATE USER 'user_name'@'10.214.3.0' IDENTIFIED BY 'password';
GRANT ALL PRIVILEGES ON shorturl.* TO 'user_name'@'10.214.3.0';
You can create a user who can access the database only from the network with the IP address 10.214.3.0.
I would like to know if it is possible to achieve similar functionality in mongodb as well. This would be very useful for my purposes, as I want…
3 votes -
Support concise correlated queries for $graphLookUp
Similar to $lookUp, support concise correlated queries for $graphLookUp.
It would allow adding a $project stage to reduce overhead.
1 vote -
MongoDB Audit Log Missing Failed and Successful Command Execution Flag
When you initially create a database account…..this action will be logged in the database audit logs as configured, however when you try to re-attempt to create the account again….a normal error message will be displayed as shown below:
db.createUser( {user: "test111_3",pwd: "emad123",roles: [ { role: "readWrite", db: "admin" } ]})
When examining the audit logs the two entries are identical in results which shouldn’t be the case….I think the flag “result” when command executed failed should have a different value to distinguish successfully executed commands from failed executed commands:
{ "atype" : "createUser", "ts" : { "$date" : "2024-08-30T13:04:59.535+03:00" },…
1 vote -
upgrade src/third_party/gperftools ≥ 2.10
gperftools new version support new architecture
1 vote -
Providing connection details with username for real time connection monitoring.
Currently, mongd.log provides connection details like remote IP, source port with authentication information but doesn't provide the connection is in active state or not.
serverStatus() only provides number of current and active connections.Example:
10.0.0.100:12345 - username active
10.0.0.101:12346 - username idle2 votes
- Don't see your idea?