Data Federation and Data Lake
57 results found
-
6 votes
-
Make mongodump work with Online Archive
Update mongodump so that it can be used against an Online Archive.
2 votes -
readPreference=Secondary for Federated Data Store
We're using Online Archives to ensure our cluster data size stays manageable, while at the same time enabling our data extraction process to have access to older date on an exception basis.
Not being able to set the read preference on our mongoexport connection string for our federated data source ( https://www.mongodb.com/docs/atlas/app-services/mongodb/read-preference/) is a significant issue for our use case.
12 votes -
Support Geo Queries on Object Storage
I'd like to be able to query using the Geo functionality inside of MongoDB Query Language on data stored in Object Storage.
Maybe using a format like: https://github.com/opengeospatial/geoparquet
3 votes -
Filtered Data Lake Ingestions
Our immediate need is that our applications are multi-tenant, so it would be very useful if we could create tenant-specific data lakes, by setting particular constraints in the ingestion configuration (ex. only ingest the documents with tenantId = 'specificTenantId').
However, the usefulness of filtered data lake ingestions can be multifaceted. The ingestion could be done only for archived=false documents, documents with status=ACTIVE, etc.2 votes -
Data Uploading process Is Little bit Difficult for new users. Upload a demo vedio of Uploading.
Overall I Found one of the Interesting Software and Friendly use
1 vote -
Atlas Data Explorer to support using Aggregation Builder against Atlas Data Lake
You can use the Atlas Data Explorer and Aggregation Builder in the MongoDB Atlas web dashboard on regular collections and views. Unfortunately there appears to be no way to use them against a Data Lake within the web dashboard, either directly or while constructing new Data Sources for Charts. Attempting to use Aggregation Builder on a Data Lake while defining a Data Source forwards to a URL that returns 404.
It would be great if the same functionality was available for Data Lake as well.
3 votes -
Add eu-north-1 as a option for AWS hosting
Sweden is a very innovative country with many startups and scaleups and AWS is used very often for hosting of services and data. Sweden is also very strict on rules where and how to store data and that is why AWS has eu-north-1 as a location to choose for storing data (which is in Sweden). Currently Data Lake doesn't support that option, the closest one is Germany. It would be great to support eu-north-1 as well, so that we don't have to live with the unnecessary latency.
3 votes -
On-line Archive survives region outage
I understand that even with a geo-replicated cluster if that cluster is configured with an online archive and there's a region outage, access to the online archive data is lost. It is still unclear to me if queries against collections configured with online would fail in this scenario. In any case, it would make sense to me to enable the S3 bucket backing the on-line archive to itself be replicated using "Amazon S3 Cross-Region Replication (CRR)"
3 votes -
Connect Atlas Data Lake to my self managed cloud object storage (S3)
I'd like to be able to connect Atlas Data Lake to my self managed cloud object storage (S3 Compatible) in my data center or private cloud
2 votes -
Import and Export archiving rules
Ability to import and export archiving rules to be able to restore them if/when we need to restore the cluster. Also useful when replicating prod clusters to our stage environment
3 votes -
Ability to use GUID field as a partition field for online archive
Hi,
Today there is no way to partition the archive data based on a field that is of type GUID (legacy GUID). For example, I tried selecting a field which had
Binary('0TfYLb3Qg0WT2mZu0wbq8Q==', 3)
as the value but I got an error saying that the field is not supported to be a partition field. It makes sense to do this because archived data is usually old and at that time most people were using legacy guids as opposed to object ids.3 votes -
Connect an ODBC Driver to Atlas Data Lake
I'd like to use an ODBC driver to connect to Atlas Data Lake's SQL support.
2 votes -
Specify when you'd like Online Archive to migrate data
I'd like the ability to specify when a migration from my Atlas cluster to my Online Archive to take place.
2 votes -
Support Online Archives in Charts
We use Atlas Charts and would like to keep the data moved to Online Archive accessible for reporting/visualizations purposes.
3 votes -
Add support to $out to S3 for Standard JSON
I'd like to be able to use $out but output to Standard JSON instead of Extended JSON as the tool I'm using needs to consume standard JSON.
3 votes -
Connect Data Lake to Self Managed MongoDB Clusters
Connect your Atlas Data Lake to Self Managed MongoDB clusters in private data centers, self managed in a public cloud, or locally hosted.
(This would also allow certain cloud services like Charts and a component of Realm Scheduled Triggers to work with these clusters.)
2 votes -
Cross Project Access to Atlas Clusters from Data Lake
I would like by Data Lake in Project A to be able to query data in a Cluster in Project B.
17 votes -
Add Incremental Materialized Views
Add the ability to create a view where the result is pre-computed and is updated incrementally as more data becomes available.
6 votes -
Ability to "rehydrate" Atlas cluster from online archive
Consider an archive scenario when a user of a given app has not logged into the app in [x] number of weeks/months, so all their data is moved to Online Archive. Once they log back into the app again, their "cold" data should now be considered "hot" and be moved back into Atlas. While we can use $out to copy data back to Atlas, there is no current way to remove the "rehydrated" data from S3 once it's been copied back to Atlas
5 votes
- Don't see your idea?