Data Federation and Data Lake
40 results found
-
Support Google Cloud Storage
I'd like to use the Atlas Data Lake with data I've stored in Google Cloud Storage.
47 votes -
Support Microsoft Azure Blob Storage
I'd like to use the Atlas Data Lake with data I've stored in Microsoft Azure Blob Storage.
39 votes -
Connect Power BI to Atlas Data Lake
I'd like to create visualizations by connecting Power BI to Atlas Data Lake.
16 votes -
readPreference=Secondary for Federated Data Store
We're using Online Archives to ensure our cluster data size stays manageable, while at the same time enabling our data extraction process to have access to older date on an exception basis.
Not being able to set the read preference on our mongoexport connection string for our federated data source ( https://www.mongodb.com/docs/atlas/app-services/mongodb/read-preference/) is a significant issue for our use case.
12 votes -
Allow GridFS to use Atlas and object storage (via ADL) when connecting to the cloud MDB
Many users of MongoDB store metadata in MDB and PDFs and other files in object storage. With GridFS already built into drivers, it seems like a nice change would allow ADL to federate GridFS functionality across Atlas and the file in object storage
12 votes -
On-Prem Data Lake Options
I'd like to run an instance of MongoDB's Data Lake product on my own infrastructure.
12 votes -
Access Online Archive in Customer Data Lake
I've started using Online Archive for my data and would not like to add this as a source to my Atlas Data Lake so I can query my cluster + archive + s3 data all together.
10 votes -
S3 alternative provider support
A lot of providers support the same API of AWS. I think it will be simple to integrate them !
9 votes -
Process files and move them when processed
I use AWS SES for incoming emails. It drops text files into an s3 bucket and then I need to parse these into json documents. There did not seem to be any means of setting this up within data-lake. I also need to move documents from the inbox folder to the processed folder once they have been added to the database. Was hoping data lake could manage all my ingest etl for me and all I need to do is worry about talking to the database once set up, at this point I think I will just implement the above…
7 votes -
Cross Project Access to Atlas Clusters from Data Lake
I would like by Data Lake in Project A to be able to query data in a Cluster in Project B.
6 votes -
5 votes
-
Ability to "rehydrate" Atlas cluster from online archive
Consider an archive scenario when a user of a given app has not logged into the app in [x] number of weeks/months, so all their data is moved to Online Archive. Once they log back into the app again, their "cold" data should now be considered "hot" and be moved back into Atlas. While we can use $out to copy data back to Atlas, there is no current way to remove the "rehydrated" data from S3 once it's been copied back to Atlas
5 votes -
Support AWS IAM for Data Federation Authentication
We would like to be able to connect to the Federated Database Instance using AWS IAM for Authentication just like you can for Atlas Clusters.
4 votes -
build atlas on prem
Allow atlas to control on prem instance
4 votes -
Online Archive
Hi Team - With regards to Atlas Data lake and using Online Archive customer request to be able to have (Time + Query) i.e. anything that is older than 60 days that match X query.
4 votes -
Support Geo Queries on Object Storage
I'd like to be able to query using the Geo functionality inside of MongoDB Query Language on data stored in Object Storage.
Maybe using a format like: https://github.com/opengeospatial/geoparquet
3 votes -
Atlas Data Explorer to support using Aggregation Builder against Atlas Data Lake
You can use the Atlas Data Explorer and Aggregation Builder in the MongoDB Atlas web dashboard on regular collections and views. Unfortunately there appears to be no way to use them against a Data Lake within the web dashboard, either directly or while constructing new Data Sources for Charts. Attempting to use Aggregation Builder on a Data Lake while defining a Data Source forwards to a URL that returns 404.
It would be great if the same functionality was available for Data Lake as well.
3 votes -
Add eu-north-1 as a option for AWS hosting
Sweden is a very innovative country with many startups and scaleups and AWS is used very often for hosting of services and data. Sweden is also very strict on rules where and how to store data and that is why AWS has eu-north-1 as a location to choose for storing data (which is in Sweden). Currently Data Lake doesn't support that option, the closest one is Germany. It would be great to support eu-north-1 as well, so that we don't have to live with the unnecessary latency.
3 votes -
Import and Export archiving rules
Ability to import and export archiving rules to be able to restore them if/when we need to restore the cluster. Also useful when replicating prod clusters to our stage environment
3 votes -
Ability to use GUID field as a partition field for online archive
Hi,
Today there is no way to partition the archive data based on a field that is of type GUID (legacy GUID). For example, I tried selecting a field which had
Binary('0TfYLb3Qg0WT2mZu0wbq8Q==', 3)
as the value but I got an error saying that the field is not supported to be a partition field. It makes sense to do this because archived data is usually old and at that time most people were using legacy guids as opposed to object ids.3 votes
- Don't see your idea?