We need to store some data in an encrypted form. We normally use S3's SSE. It seems that we cannot write from Spark to S3 using SSE. Supporting that would be helpful.9 votes
This feature is now finished. We support SSE-S3 and SSE-KMS. This is available in Databricks 2.2, which is going out in this/next week. One can mount directories using dbutils.fs.mount by specifying the desired SSE-based encryption method.
Expand the potential number of file types that can be imported via the web UI when creating a table e.g. JSON. Currently it appears to be text delimited files only.1 vote
This has been implemented. The UI in the table import now supports JSON as well as JDBC.
You can print notebooks, and many browsers allow you to print to PDF.
I have a large CSV file with column headers in the first row. When I try to import it as a table all my columns are of type "STRING" and my first row contains headers.
Would be good to be able to specify that first line in a text file contains column names.2 votes
CSV files often use quotes to protect embedded commas and the like.
Please provide an option to do quote processing.2 votes
This works for local files uploaded to Databricks Cloud (DBC). In Q1 2015 we will also add support for it for S3 import.
Support importing data from external databases, similar to the way one could import s3 buckets - I would personally love support for MongoDB and other NoSQL stores. I'd also consider making it easy to fetch data from Freebase etc.4 votes
- Don't see your idea?