-
Notifications
You must be signed in to change notification settings - Fork 17
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Recognize py file as a notebook and use azure cluster as a kernel #578
Comments
Hi guys, Found this documentation and I have been trying to implement
Here the code: from databricks.connect import DatabricksSession
from pyspark.dbutils import DBUtils
spark = DatabricksSession.builder.getOrCreate()
dbutils = DBUtils(spark) # error in this line if I commented out that line and try to load a list of dictionaries in a pyspark df it fails with not mayor logs:
This code works fine in databricks UI notebook df = spark.createDataFrame(my_list_of_dicts)
df.limit(10) Environment: vscode: Databricks extension version: v0.3.11 |
Thumbs up for this feature request! 👍 I do use .py scripts or .ipynb notebooks depending on the project, and for both approaches GitHub copilot is amazing for accelerating code development. It would be great to be able to open Databricks .py scripts in VS code as a notebook, selecting a Databricks cluster as the compute environment. This would allow us to have a GitHub Copilot enabled Databricks environment, where we could be developing notebooks together with the copilot. |
+1 for the .py file which are the format used when notebook are synchroniser in a repository (Git). |
Is there any progress on this request? I have been using a lot databaricks and this will be a killer feature. |
We have some limited notebook support in experimental. You can enable it here https://docs.databricks.com/en/dev-tools/vscode-ext/dev-tasks/databricks-connect.html#additional-notebook-features-with-databricks-connect. You will still have to use the code lenses (the little gray buttons on top of cells), but it should popup a really nice window to show you outputs for the current cell. |
Describe the enhance
It will be a great feature if vscode can recognize the databricks notebook
py
files as a notebook and allow us to select the databricks cluster as a kernel.Even though we can run the notebook as
workflow on Databricks
it runs the entire notebook. What if I only want to run one cell at a time.How to reproduce
vscode already recognize
# COMMAND ----------
as a cell:but if we try to run it (or debug), there is not the databricks cluster:
Ideal
based on the first commented line
# Databricks notebook source
open the file as a notebook and allow me to select the databricks kernel:NOTE: Please add a tag to suggest features additional to the one for bugs :)
The text was updated successfully, but these errors were encountered: