Is it possible to use the Scala worksheet in IntelliJ as an alternative to a Jupyter notebook. I followed the solution mentioned here , however it runs the code locally and not on a remote cluster. My main challenge is that the IntelliJ IDE is running locally in my laptop and the spark cluster is in the cloud. How do I ask IntelliJ to use the remote SBT ?
Use IntelliJ idea's Scala worksheet as a Scala REPL for Apache Spark
743 Views Asked by Liberty_bird At
1
There are 1 best solutions below
Related Questions in SCALA
- Mocking AmazonS3 listObjects function in scala
- Last SPARK Task taking forever to complete
- How to upload a native scala project to local repo by sbt like using "maven install"
- Folding a list of OR clauses in io.getquill
- How to get latest modified file using scala from a folder in HDFS
- Enforce type bound for inferred type parameter in pattern matching
- can't write pyspark dataframe to parquet file on windows
- spark streaming and kafka integration dependency problem
- how to generate fresh singleton literal type in scala using macros
- exception during macro expansion: type T is not a class, play json
- Is there any benefit of converting a List to a LazyList in Scala?
- Get all records within a window in spark structured streaming
- sbt publishLocal of a project with provided dependencies in build.sbt doesn't make these dependencies visible to projects using the project as library
- Scala composition of partially-applied functions
- How to read the input json using a schema file and populate default value if column not being found in scala?
Related Questions in APACHE-SPARK
- Getting error while running spark-shell on my system; pyspark is running fine
- ingesting high volume small size files in azure databricks
- Spark load all partions at once
- Databricks Delta table / Compute job
- Autocomplete not working for apache spark in java vscode
- How to overwrite a single partition in Snowflake when using Spark connector
- Parse multiple record type fixedlength file with beanio gives oom and timeout error for 10GB data file
- includeExistingFiles: false does not work in Databricks Autoloader
- Spark connectors from Azure Databricks to Snowflake using AzureAD login
- SparkException: Task failed while writing rows, caused by Futures timed out
- Configuring Apache Spark's MemoryStream to simulate Kafka stream
- Databricks can't find a csv file inside a wheel I installed when running from a Databricks Notebook
- Add unique id to rows in batches in Pyspark dataframe
- Does Spark Dynamic Allocation depend on external shuffle service to work well?
- Does Spark structured streaming support chained flatMapGroupsWithState by different key?
Related Questions in BIGDATA
- How to make an R Shiny app with big data?
- Liquibase as SaaS To Configure Multiple Database as Dynamic
- how to visualize readible big datasets with matplotlib?
- Are there techniques to mathematically compute the amount of searching in greedy graph searching?
- Pyspark & EMR Serialized task 466986024 bytes, which exceeds max allowed: spark.rpc.message.maxSize (134217728 bytes)
- Is there a better way to create a custom analytics dashboard tailored for different users?
- Trigger a lambda function/url with Apache Superset
- How to download, then archive and send zip to the user without storing data in RAM and memory?
- Using bigmemory package in R to solve the Ram memory problem
- spark - How is it even possible to get an OOM?
- Aws Athena SQL Query is not working in Apache spark
- DB structure/file formats to persist a 100TB table and support efficient data skipping with predicates in Spark SQL
- How can I make this matching function faster in R? It currently takes 6-7 days, and this is not practical
- K-means clustering time series data
- Need help related to Data Sets
Related Questions in JUPYTER
- I am getting 'NoneType object is not subscriptable' error in web scraping method
- Jupyter notebook: " No module named 'google.colab'", after having updated Anaconda
- Jupyter Hub not fiinding my workspace and killing it (version 4.1.13)
- How to call `InteractiveShellEmbed`-based custom excepthook via `jupyter-console --existing`
- RuntimeError: PytorchStreamReader failed reading zip archive: failed finding central directory when running command to train yolov5 on custom data
- Same python code runs 20 times slower in Jupyter Notebook compared to command line
- How to render more than 1 dialogs on Jupyterlab?
- Jupyter Notebook not rendering report with pandas profiling
- jupyter notebook - Why `tornado.web.HTTPError: HTTP 403: Forbidden`?
- Error when trying to open csv in google colab; Jupyter works well
- Jupyter command `jupyter-bundlerextension` not found
- having issues installing libraries using the 'conda' command after installing anaconda
- When using a jupyter notebook in VSCode, the Python interpreter is ignored
- Unexpected Error with ipynb file in VSCode
- Spacy matcher is not finding any matches for counties
Related Questions in APACHE-TOREE
- Default lib jars folder for Apache Toree kernel
- Transpose in Scala
- Problem reading Kudu tables using Spark (Jupyer Notebook with Apache Toree - Scala Kernel )
- Unapplied methods are only converted to functions when a function type is expected
- How to read a text file from HDFS in Scala natively (without using Spark)?
- How to execute shell command in Apache Toree notebook
- Fail to create Spark session
- How to set a specific port for single-user Jupyterhub server REST API calls?
- "no libName in java.library.path" on System.loadLibrary("libName")
- Use IntelliJ idea's Scala worksheet as a Scala REPL for Apache Spark
- Jupyter Notebook (Scala, kernel - Apache Toree) with Vegas, Graph not showing data
- How to make Apache Toree work with Jupyter
- How to enable hive support for spark in notebook?
- How to show a dataframe in a good format in jupyter notebook using pyspark with apachee-toree kernel
- How to show the dataframe in pyspark in presentable manner.
Trending Questions
- UIImageView Frame Doesn't Reflect Constraints
- Is it possible to use adb commands to click on a view by finding its ID?
- How to create a new web character symbol recognizable by html/javascript?
- Why isn't my CSS3 animation smooth in Google Chrome (but very smooth on other browsers)?
- Heap Gives Page Fault
- Connect ffmpeg to Visual Studio 2008
- Both Object- and ValueAnimator jumps when Duration is set above API LvL 24
- How to avoid default initialization of objects in std::vector?
- second argument of the command line arguments in a format other than char** argv or char* argv[]
- How to improve efficiency of algorithm which generates next lexicographic permutation?
- Navigating to the another actvity app getting crash in android
- How to read the particular message format in android and store in sqlite database?
- Resetting inventory status after order is cancelled
- Efficiently compute powers of X in SSE/AVX
- Insert into an external database using ajax and php : POST 500 (Internal Server Error)
Popular Questions
- How do I undo the most recent local commits in Git?
- How can I remove a specific item from an array in JavaScript?
- How do I delete a Git branch locally and remotely?
- Find all files containing a specific text (string) on Linux?
- How do I revert a Git repository to a previous commit?
- How do I create an HTML button that acts like a link?
- How do I check out a remote Git branch?
- How do I force "git pull" to overwrite local files?
- How do I list all files of a directory?
- How to check whether a string contains a substring in JavaScript?
- How do I redirect to another webpage?
- How can I iterate over rows in a Pandas DataFrame?
- How do I convert a String to an int in Java?
- Does Python have a string 'contains' substring method?
- How do I check if a string contains a specific word?
If you are just trying out a tutorial in a more interactive/adhoc fashion to learn Spark programming and concepts, the link you mentioned in your post is how you'll do it. In that mode, you're basically simulating a single-node Spark cluster (i.e your local machine) that acts as both the Driver as well as the executor node (all in one).
However, that's not how you'll actually submit and run a Spark application on an ACTUAL Spark cluster in a more real world scenario. If that's what you're trying to do, you'll need to instead use one of the two deployment modes Spark offers: Client Mode and Cluster Mode. You'll use the "spark-submit" command line utility to submit your compiled Spark library (the fat JAR) as a Spark job. More details here.