Most people on Azure are likely to also use VSTS as their git repository. VSTS support authorization tokens and git operations just like GIT itself so it should be an easy addition I would think.582 votes
We have now GA:d VSTS integration for Azure Databricks. Find out more about it on this documentation page:
It will really aid productivity if we can search and replace across all cells within a notebook.49 votes
Databricks notebooks supports ‘find and replace’ within a notebook. You can find documentation here:
The cell menu now allows copying and pasting whole cells, even across notebooks.
It would be great to have collapsible markdown headings just like the notebook extension "collapsible headings" in jupyter. Notebooks tend to become very large very quickly and this feature would greatly help us.72 votes
This is now a feature – please check it out! Cells that appear after cells containing Markdown headings can now be collapsed into the heading cell.
When a notebook is exported, the command cell titles are lost and cannot be imported.
This includes when using the built in github / bitbucket integration.
Perhaps the titles could be included within the MAGIC comments that proceed each cell in the source.4 votes
it would be easier to say to my colleagues "have a look at cell #15" rather than "scroll down and look for such and such title, after such chart".
it's ok if the cell numbers change even though i add and remove cells often. cell numbering is helpful when i'm asking someone to review something, which lasts about 15min most of the time. if i move the cells around, i'll inform them which cell again, that's not really a big issue for me.3 votes
Cell/command numbers are now supported. Under the File menu enable “Show Command Numbers”. In addition to cell/command numbers, you can now click on a command number, which changes the URL. You can, hence, share links to particular cells/commands.
Debugging in Databricks is plenty annoying as is. Letting us see line numbers should be a simple change that would make debugging much less frustrating. (Sometimes I've ended up pasting the contents of a cell into a text editor to find the line providing the error!)40 votes
Both line numbers and command numbers are now available. Under the VIEW menu you can show/hide them. Your preferences will be remembered. Command numbers can be clicked on, which gives you a permalink to that particular command. Such permalinks are great for sharing and pointing collaborators to specific commands in notebooks.
Sometimes you have logic in one cell which fails but other cells below it rely on it. It would be nice if you could do "run all down" meaning it runs the selected cell, plus all remaining cells in the notebook.
This would be nice in cases where you don't want to re-run the whole notebook, but just from the failed point on.5 votes
Every cell now has a run button with a dropdown letting you pick “Run All Above” and “Run All Below”.
It would be great to have "search and replace" ability in DB notebooks3 votes
Find and replace is released in August 2017, see more here: https://docs.databricks.com/release-notes/product/2017/august.html#notebook-find-and-replace
Define groups with a list of users, so we can set permission on a notebook/folder for that group.
Currently, whenever we add user to a team there is a lot of manual adding of the user to many notebook/folders.26 votes
Databricks released REST APIs to create and manage user groups. The API documentation is listed below:
"New client version available. Refresh page?" pop-up responds to 'OK'. Currently it pops again unless you hit 'cancel'.1 vote
Allow a repository branch to be specified with the GitHub integration. For more complicated Git workflows it isn't always best to make changes to the master branch. Being able to specify the a custom branch will allow users to create a dev branch for testing and merge it into the master branch when the notebook has been completed.28 votes
This feature has now been released in Databricks 2.11. You can select branches, create branches, and even submit pull requests.
For small, one-off tasks we use a shared default cluster. When we need to restart that cluster we do not want to kill any running tasks, so we need to go to Cluster UI, expand the cluster in question and open each notebook attached to that cluster to check whether it's running any command. This is tedious if we have a lot of notebooks attached to the cluster.
It would be cool if the UI showed whether the notebook is running any command or not.22 votes
This is now available.
1. Click on Clusters in the navigation bar.
2. Then click on the relevant cluster in the cluster list. This takes you to a detailed page about the selected cluster.
3. Now click on the “Notebooks” tab. The middle “status” column will either say “Idle” or “Running” depending on if the listed notebook is currently running or not.
Allow specifying default permissions for home directories of newly added users, when using ACLs. We would like to add an Everyone/Read permission by default.2 votes
This should now be possible.
We have a lot of libraries loaded and they fill the workspace pull-right. Also, the same library name can appear multiple times. It would be great if there were some more logical and informative display of the libraries.1 vote
The best practice we see from other customers is to organize the libraries into sub-folders. Otherwise it looks as if the workspace is “polluted” with a lot of libraries. We are simultaneously looking at improving our library management.
Pass arguments from the jobs to a notebook5 votes
It is possible to use the REST API to pass arguments to both notebook and JAR jobs.
There doesn't seem to be a way to get the current date in sql. now() returns a null pointer exception.3 votes
Please use unix_timestamp() instead. We’ll look into now().
If would be great if standard IPython notebook (as opposed to dbc format) could be imported to DBC. Why change the standard ipynb files? That allows users to test their ideas on any platform and then import to DBC1 vote
This feature is completed. You can now import and export to/from the Ipython notebook format.
Importing IPython nootbooks using drag and drop is confusing (move/ copy etc). Sometimes after dropping it just shows the markup with no err
Importing IPython nootbooks using drag and drop is confusing (move/ copy etc). Sometimes after dropping it just shows the markup with no err.Please improve this1 vote
You should be able to copy/paste cells in notebooks without drag ‘n drop. That’s especially useful for either (a) big cells with a lot of text in them or (b) cells which you want to move long distances.
My preferred method of offline editing would involve cloning a git repo, and pushing my changes when I got back online.13 votes
Databricks 1.4.2 which is released to every customer today should have GitHub integration for Professional Tier.
- Don't see your idea?