Data Federation and Data Lake
5 results found
-
Process files and move them when processed
I use AWS SES for incoming emails. It drops text files into an s3 bucket and then I need to parse these into json documents. There did not seem to be any means of setting this up within data-lake. I also need to move documents from the inbox folder to the processed folder once they have been added to the database. Was hoping data lake could manage all my ingest etl for me and all I need to do is worry about talking to the database once set up, at this point I think I will just implement the above…
7 votes -
Simplified JSON support for $out to S3
The ability to $out to S3 from a federated database instance is a game-changer for those working with their own data warehouses and data lakes.
One improvement that would make it better would be to support simplified JSON for json exports. Currently, $out uses extended json v2, which may not be compatible for systems reading from the destination S3 bucket, which require simplified JSON (which aligns with other tools like kafka source connector). Technically, it is possible to make this conversion yourself with clever use of the $toString aggregation pipeline operator in stages preceding $out. However there are several challenges:…
6 votes -
Add support to $out to S3 for Standard JSON
I'd like to be able to use $out but output to Standard JSON instead of Extended JSON as the tool I'm using needs to consume standard JSON.
3 votes -
Add support for Text format files
I have a custom log format that I'd like to be able query. I imagine I would describe the format of the text files to Atlas Data Lake and then be able to query them.
3 votes -
Add support for XML
I would like to be able to query XML files using my Atlas Data Lake
2 votes
- Don't see your idea?