Product Feedback

  • Hot ideas
  • Top ideas
  • New ideas
  • My feedback
  1. Clusters created with python3 used pip3

    I created a cluster using python3, but pip failed to install a package made for python3, and pip3 is not present.

    1 vote
    Vote
    Sign in
    (thinking…)
    Sign in with: Facebook Google
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    1 comment  ·  External libraries / applications  ·  Flag idea as inappropriate…  ·  Admin →
  2. there is a detailed description of mlflow exceptions along with their resolutions on mlflow website.

    While trying out the various functionality of mlflow lifecycle, I experienced multiple mlflow exceptions like 'Could not find a registered tracking store ' etc. which was not explained anywhere on mlflow website or even in Azure Databricks documentation. It will be good to have a dedicated page explaining the reasons of the exceptions that usually occur along with solution to resolve them similar to what Microsoft does for Windows exceptions.

    1 vote
    Vote
    Sign in
    (thinking…)
    Sign in with: Facebook Google
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Flag idea as inappropriate…  ·  Admin →
  3. Upload library for all clusters through API

    If a library has been set to be installed on all clusters, islibraryforallclusters will be true. Right now, this can only be done through the home page. It would be great if this could be done through the API or CLI.

    3 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Facebook Google
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  REST API  ·  Flag idea as inappropriate…  ·  Admin →
  4. The job list page sorted by job id in descending order

    When I navigate to the job list page (https://###.cloud.databricks.com/#joblist) it would be nice if the jobs were sorted by job id in descending order so the latest job was at the top.

    I use Hadoop and HUE and they do this and I find it very less user friendly that the Databricks job list doesn't do this.

    Also, if I look at the details of a job and then navigate back to the job list the sort I had selected (job id desc) is lost and I have to resort.

    1 vote
    Vote
    Sign in
    (thinking…)
    Sign in with: Facebook Google
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    1 comment  ·  Navigation UI  ·  Flag idea as inappropriate…  ·  Admin →
  5. i am unable to create Cluster using trial version , its showing error "This account may not have enough cores to satisfy the request"

    I am trying to create cluster using trial account however getting error. I am unable to set the configuration as its allowing enter min and max workers. 1 worker is adding 4 cores and I have max 4 cores allocated. I cannot provide 0 as min value and max should be greater than min. So technically I am unable to create the cluster, please help.

    1 vote
    Vote
    Sign in
    (thinking…)
    Sign in with: Facebook Google
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    1 comment  ·  Flag idea as inappropriate…  ·  Admin →
  6. Can list what users have create cluster permission through api

    I would like to be able to list what users have create cluster permission through the rest api.

    3 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Facebook Google
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Flag idea as inappropriate…  ·  Admin →
  7. Style of widgets

    Customize the style of any widgets in azure Databricks is not possible

    3 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Facebook Google
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Notebooks  ·  Flag idea as inappropriate…  ·  Admin →
  8. use VSTS git repo with databricks on AWS

    We use VSTS for CI/CD and AWS for our cloud infrastructure. It would be nice to access VSTS from databricks on AWS.

    3 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Facebook Google
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Flag idea as inappropriate…  ·  Admin →
  9. Display Function in Notebooks could be used to generate multiple plots

    I would like to be able to generate a set of visuals from a single display function call or SQL query output without having to rerun the same chunk multiple times.

    13 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Facebook Google
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Notebooks  ·  Flag idea as inappropriate…  ·  Admin →
  10. Databricks could integrate better with a Python IDE

    Similar to the RStudio Server integration which has become availble recently, it would be great to have an integration with at least one Python IDE (PyCharm, Spyder, etc.). Is such an integration planned for the near future? The development of code in notebooks can be rather cumbersome for larger projects.

    26 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Facebook Google
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    1 comment  ·  Other  ·  Flag idea as inappropriate…  ·  Admin →
  11. Support of Shapely values prediction in MLlib boosted tree models.

    Shapely Feature Contributions at row low level for tree ensembles. This allows explainable at the row level of why things are being predicted. There is already a method that can be build into tree ensembles (like xboost) to calculated shapely very fast using the tree nodes
    https://www.groundai.com/project/consistent-feature-attribution-for-tree-ensembles/4

    1 vote
    Vote
    Sign in
    (thinking…)
    Sign in with: Facebook Google
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  External libraries / applications  ·  Flag idea as inappropriate…  ·  Admin →
  12. Ablity to Download pickled files to a local machine, created in a notebook.

    I need a simply way to download pickled model files created by notebooks and on the cluster driver. This doesn't appear to be possible with no public IP set.

    1 vote
    Vote
    Sign in
    (thinking…)
    Sign in with: Facebook Google
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Data import / export  ·  Flag idea as inappropriate…  ·  Admin →
  13. 3 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Facebook Google
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Flag idea as inappropriate…  ·  Admin →
  14. uploading notebook blocked scheduler

    Our team uses the CLI to upload notebooks to our data-bricks system and we have the scheduler schedule these notebooks as jobs. It happened the other day that the notebook was partially uploaded when it was scheduled to run. Whats more is it uploads command by command starting with the last. So in my case only the final command was in the notebook at the time.

    This could eventually cause some extremely strange and uncontrollable errors and devs shouldn't have to think about how a notebook could run forward and potentially every subset of that in case this happens. (This…

    3 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Facebook Google
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Notebooks  ·  Flag idea as inappropriate…  ·  Admin →
  15. Ability to Upload Python Wheel Libraries through API 2.0

    It would be great if we could just upload libraries through the 2.0 API into a Workspace folder rather than a DBFS folder. We will want to use this in our CI/CD pipeline to upload the newest version of our libraries.

    3 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Facebook Google
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  External libraries / applications  ·  Flag idea as inappropriate…  ·  Admin →
  16. Manage Databricks secrets API to improve the security violations

    credentials = dbutils.secrets.get(scope = "<scope-name>", key = "<key-name>")

    print(credentials)

    output : [REDACTED]

    But through iterating, it is possible to see the hidden secrets

    for singlekeycharacter in credentials:
    print(singlekeycharacter)

    Output:
    Q
    2
    $
    %
    5
    y

    U

    ;
    5


    In this the secrets are able to retrieved by iterating secrets variable.
    To address this issue, we have to restrict the iteration for the secrets assigned variable.

    1 vote
    Vote
    Sign in
    (thinking…)
    Sign in with: Facebook Google
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  External libraries / applications  ·  Flag idea as inappropriate…  ·  Admin →
  17. Implement REST API Documentation Service like Swagger

    Implement a REST API Documentation Service like Swagger.

    This will make it easier to use the Databricks APIs easier to use.

    Sometimes the available API documentation/Examples are not granular enough.
    There are features of APIs that are not covered by API Examples which can be difficult to implement as the API Documentation isn't as detailed as it could be.

    23 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Facebook Google
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    2 comments  ·  REST API  ·  Flag idea as inappropriate…  ·  Admin →
  18. aliases in URL

    The "databricks URL" seems to be just some kind of account ID.

    We have multiple accounts and it is almost impossible to tell which account I am logged in to on any given webpage. This makes it really easy to accidentally change clusters in "prod" and not in "dev"...

    Most SaaS products have some kind of account name next to the login data. (E.g. in your case that would be in little pull down in the upper right. ("Logged in as ...".) Also, just having a URL with a name and not a semi random string (dbc-0a0a0a0a-0000.cloud...) would help.

    1 vote
    Vote
    Sign in
    (thinking…)
    Sign in with: Facebook Google
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Flag idea as inappropriate…  ·  Admin →
  19. ADLS AD integration

    When an Azure Data Lake Store Gen 2 is mounted on the cluster, I'd like users to be able to mount only the ADLS file systems folders that they have access through Active Directory.

    At the moment if a user mounts a folder from ADLS, that folder is visible on the whole cluster.

    3 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Facebook Google
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    1 comment  ·  Other  ·  Flag idea as inappropriate…  ·  Admin →
  20. Manage folder permissions through API

    I would like to manage access (permissions) to folders through the API instead of the GUI. Such that I can automatically enforce folder structure/access.

    45 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Facebook Google
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Other  ·  Flag idea as inappropriate…  ·  Admin →
  • Don't see your idea?

Feedback and Knowledge Base