Connectors (BI, Kafka, Spark)
-
Enable custom DRDL upload for Atlas BI Connector
We cannot upload custom DRDL to Atlas BI connector. This forces us to roll our own BI connector
22 votes -
Bi connector - store schema information
We have the BI connector installed/configured via OPS manager on one of our deployments.
When we have to restart the node, the BI connector takes a long time to start as it has to rebuild the schema.
Hence we would like to file an enhancement request to be able to store the schema information and then retrieve it again after a restart.9 votes -
Allow separate whitelist for Atlas BI Connector
Currently, whitelisting is only possible at the Project level. We would like to allow whitelisting of the BI connector instance separately from the project.
The Users/IPs that connect to mongo are completely different for BI connector vs the actual DB
7 votes -
Enable Write Permission In BI
Hope BI enable write permission because a lot of system especially Windows base using ODBC
6 votes -
Add X509 as a Authentication Method
The current authentication methods lack in their ability to control password complexity requirements (SCRAM-SHA-*) or require exposing on-premise LDAP servers to the public internet. The addition of X509 is worthwhile given that MongoDB Atlas already supports X509 as a user authentication mechanism.
5 votes -
BI Connector username and password in yaml file
Currently BI Connector has username and password in clear. this presents problem in terms of security. BI connector has to have a way to encrypt the password and then we store the encrypted password which BI connector can decrypt and use it. it can be any standard encrypt decrypt algorithm but the secret key can be generated by the person using this.
5 votes -
BI Connector Atlas - View current schema creation/update status
In Atlas the is no way to see the BI connector logs, so, the request is to be able to see the Current schema creation/update status or even better, be able to see the log.
5 votes -
Notification and alerts when BI Connector fails
When a BI Connector fails, no alerts are sent to admins, only the notification on the page is available indicating that it will "restart in 5 minutes".
5 votes -
Support Prepared Statements in BI Connector
The Power Query Editor in Microsoft leverages a prepared statement for query/filtering.
Error message:
This command is not supported in the prepared statement protocol yetThis is a feature request to support prepared statements in the mongosqld BI Connector service.
4 votes -
Enable CSFLE decryption in BI Connector
Enable CSFLE decryption in BI Connector, in order to provide encrypted fields in MongoDB decrypted for ODBC exposition.
CSFLE is described here: https://www.mongodb.com/docs/manual/core/security-client-side-encryption/.
For exemple, CSFLE would then be an option triggerable at BI Connector Launch, that should come with KMS link or keyring if enabled.3 votes -
Ignore heartbeats-mongodb topic by default
As per KAFKA-208, SMTs can't be applied to the heartbeats-mongodb topic. Users should not have to configure each connector to ignore this topic. Please either ignore this topic by default or provide a command-line switch so it can be ignored.
3 votes -
Multiple clusters in a single mongosqld configuration.
The limitation of supporting multiple clusters in a single mongosqld configuration is now supported or not.
3 votes -
Get schema validation "feedback" in Kafka Mongo Sink Connector
Objective :
We want to be able to validate that data matches some requirements. We would like to to perform this data validation by adding a JSON schema in Mongo (such as it is described here : https://docs.mongodb.com/manual/core/schema-validation/).
Problem is that current implementation of the current Mongo DB Kafka Sink connector does not implement the required elements to benefit from features brought by this KIP : https://cwiki.apache.org/confluence/display/KAFKA/KIP-610%3A+Error+Reporting+in+Sink+Connectors
So if we define such a validation on Mongo, if a message has a value that does not match the definition, it would not go in the dead letter queue, and the…
3 votes -
Kafka source connector once only semantics
Added as a suppport case here : https://support.mongodb.com/case/00634630
When using the connector as a Source, i.e we capture change streams from the Source Mongo DB and stream that to a Kafka endpoint.
Imagine these are updates on financial transactions in mongodb and they are NOT tolerant to
1) missed data and
2) duplicated data
in that order.So, we need to make sure that the Change Streams that we are observing(matching) on, are delivered once and exactly once to the Kafka pipeline. (Blog on the same : https://www.confluent.io/blog/exactly-once-semantics-are-possible-heres-how-apache-kafka-does-it/). If exactly-once semantics are enabled, it makes commits transactional by default.
…
3 votes -
Kafka connector to support Kafka Schema Registry
One of the issues that our team has been talking about is when getting data from MongoDB, via a Kafka connector, and sending it through to Kafka we try to enforce schemas in Kafka but that schema is not enforced on the MongoDB data. This leads to developers needing to make sure they let the Data Engineering team know when their schema evolves so we can accommodate that change in the Avro schema. Our thought is to potentially have the developers use the Confluent Schema Registry to serialize their data to Avro prior to writing it to MongoDB. This would…
2 votes -
Can't fetch data on MongoDB ODBC via BI connector
Can't fetch data on MongoDB ODBC via BI connector. Test connections are successful but couldn't fetch the data from database servers, only getting 'information_schema' and 'mysql' as by default databases
2 votes -
MongoDB Sink Connector CDC default handler
I would like to have a default CDC handler that can process data produced from MongoDB Source Connector without Debezium https://docs.mongodb.com/kafka-connector/master/kafka-sink-cdc#cdc-handler-configuration
2 votes -
Deploy MongoDB BI Connector product using the MongoDB Kubernetes Operator
We would like to have the ability to deploy and run MongoDB BI connector as a container under MongoDB Kubernetes Operator. Currently there is no support on such deployments.
2 votes -
Two tiered model for authentication
The BI Connector facilitates large scale (1000s) of "personal data marts" by acting as a controlled go-between enduser tools like Tableau and a "main" data collection. It is not practical or even desirable to have pass-thru authentication of all these users to the backend mongodb database. Instead, the BI connector could use a special collection in a mongodb instance (not necessarily the target!) to hold SHA(password), name, and YAML equivalent. When started, the mongosqld would verify command line inputs of SHA(password) and name and context, etc. and if OK, would exec an appropriately password-protected endpoint at 3307 with the config…
1 vote -
Allow per-field length declaration for varchar and char types in BI Connector
Usually in defining schemas for SQL databases, you can specify a max length for the size of a char or varchar column. It'd be nice to have the ability to do that in a schema that's passed to the mongosqld BI Connector process.
The only option now is to specify a max varchar size that applies to all varchar fields. It'd be nice to be able to define this on a per field basis.
This is an issue for a customer I'm working with because of the way their BI tool allocates memory for temporary objects created when bridging the…
1 vote
- Don't see your idea?