Data Federation and Data Lake
40 results found
-
Support Online Archives in Charts
We use Atlas Charts and would like to keep the data moved to Online Archive accessible for reporting/visualizations purposes.
3 votes -
Add Incremental Materialized Views
Add the ability to create a view where the result is pre-computed and is updated incrementally as more data becomes available.
3 votes -
Add support for Text format files
I have a custom log format that I'd like to be able query. I imagine I would describe the format of the text files to Atlas Data Lake and then be able to query them.
3 votes -
Simplify interface for query commands
User friendly data filtering, queries for updating or deleting data from collections.
3 votes -
Support the PDF File Format
I would like to be able to query PDF files using Atlas Data Lake.
3 votes -
Add last modified timestamp to Data Federation Provenance for S3
It would be great to have the last modified timestamp of a file in S3 returned with the provenance functionality in Atlas Data Federation.
2 votes -
Filtered Data Lake Ingestions
Our immediate need is that our applications are multi-tenant, so it would be very useful if we could create tenant-specific data lakes, by setting particular constraints in the ingestion configuration (ex. only ingest the documents with tenantId = 'specificTenantId').
However, the usefulness of filtered data lake ingestions can be multifaceted. The ingestion could be done only for archived=false documents, documents with status=ACTIVE, etc.2 votes -
On-line Archive survives region outage
I understand that even with a geo-replicated cluster if that cluster is configured with an online archive and there's a region outage, access to the online archive data is lost. It is still unclear to me if queries against collections configured with online would fail in this scenario. In any case, it would make sense to me to enable the S3 bucket backing the on-line archive to itself be replicated using "Amazon S3 Cross-Region Replication (CRR)"
2 votes -
Connect Atlas Data Lake to my self managed cloud object storage (S3)
I'd like to be able to connect Atlas Data Lake to my self managed cloud object storage (S3 Compatible) in my data center or private cloud
2 votes -
Connect an ODBC Driver to Atlas Data Lake
I'd like to use an ODBC driver to connect to Atlas Data Lake's SQL support.
2 votes -
Specify when you'd like Online Archive to migrate data
I'd like the ability to specify when a migration from my Atlas cluster to my Online Archive to take place.
2 votes -
Connect Data Lake to Self Managed MongoDB Clusters
Connect your Atlas Data Lake to Self Managed MongoDB clusters in private data centers, self managed in a public cloud, or locally hosted.
(This would also allow certain cloud services like Charts and a component of Realm Scheduled Triggers to work with these clusters.)
2 votes -
M0 Support for Evaluation
Please provide M0 support for evaluation purposes.
2 votes -
The "Date field to archive on" option under Archiving Rule tab should also accept date in timestamp format.
The "Date field to archive on" option under Archiving Rule tab in Online Archive should also accept date field having timestamp format instead of only having date format.
2 votes -
Add support for XML
I would like to be able to query XML files using my Atlas Data Lake
2 votes -
Specify Delimiter for CSV Files
I need to specify the delimiter for my CSV files.
2 votes -
Schema inference
Schemaless is flexible but it has a big impact for the downstreams especially for data exchange and DW/AI.
It is a must-have effort to derive & infer the schema from the actual documents, so that we can understand/track/evolve/translate the document schema.
https://www.mongodb.com/blog/post/engblog-implementing-online-parquet-shredder is a great article.
I'd like to propose an additional feature in ADL/ADF to make schema inference as a 1st-class citizen with faster turnaround & less operation cost.
After the $out operation of ADL/ADF, please collect the Parquet schema from each data files and union/unify them into a single schema. This schema will be stored in a .schema.json…
1 vote -
Make mongodump work with Online Archive
Update mongodump so that it can be used against an Online Archive.
1 vote -
Data Uploading process Is Little bit Difficult for new users. Upload a demo vedio of Uploading.
Overall I Found one of the Interesting Software and Friendly use
1 vote -
Add support to $out to S3 for Standard JSON
I'd like to be able to use $out but output to Standard JSON instead of Extended JSON as the tool I'm using needs to consume standard JSON.
1 vote
- Don't see your idea?