I would like to set up tests on my transforms into Foundry, passing test inputs and checking that the output is the expected one. Is it possible to call a transform with dummy datasets (.csv file in the repo) or should I create functions inside the transform to be called by the tests (data created in code)?
Python unit tests for Foundry's transforms?
1k Views Asked by pietro At
1
There are 1 best solutions below
Related Questions in TESTING
- How does Robot's Telnet library work?
- Behat doesn't load extensions?
- Load additional CONFIG file with values
- rails controller test failing non-deterministicly wrt state leak (I think)
- Ordering tests using trial twisted
- Unexcepted failed Gavel/Dredd test
- How to use Jasmine and CucumberJS with Protractor
- Django login tests session problems
- How to mock specific RequireJs dependencies while unit testing
- Test case for WCF REST Service
- how to test this business logic
- Protractor - How to get first or last CHILD value
- Factory Not Registered in rspec but found in console
- Pick out certain lines from files
- Selenium stops running after click() function runs
Related Questions in PYSPARK
- dataframe or sqlctx (sqlcontext) generated "Trying to call a package" error
- Importing modules for code that runs in the workers
- Is possible to run spark (specifically pyspark) in process?
- More than expected jobs running in apache spark
- OutOfMemoryError when using PySpark to read files in local mode
- Can I change SparkContext.appName on the fly?
- Read ORC files directly from Spark shell
- Is there a way to mimic R's higher order (binary) function shorthand syntax within spark or pyspark?
- Accessing csv file placed in hdfs using spark
- one job takes extremely long on multiple left join in Spark-SQL (1.3.1)
- How to use spark for map-reduce flow to select N columns, top M rows of all csv files under a folder?
- Spark context 'sc' not defined
- How lambda function in takeOrdered function works in pySpark?
- Is the DStream return by updateStateByKey function only contains one RDD?
- What to set `SPARK_HOME` to?
Related Questions in PALANTIR-FOUNDRY
- How do I configure a Contour heatmap to display colors for small overall data?
- Embed a Contour board in an iframe in Object Explorer or Slate
- In Foundry Contour, how can I analyze a prior version of a dataset?
- How do I make a Foundry Slate container collapsible?
- How to decrypt AES-encrypted files with Data Connection?
- Does Foundry Data Connection support NFS transfer?
- How do I enforce a minimum test coverage percentage in my Foundry Code Repositories?
- Data Connection - downloading files from multiple URLs in one sync
- How to configure Code Workbook's timezone for CURRENT_TIMESTAMP?
- install JAR package related to pyspark into foundry
- How to access report parameters array in fusion spreadsheet
- How to create a dataset from a fusion sheet in foundry?
- How can you use Data Connection to sync a dataset with only a schema and no rows?
- How can I get the minimum value of multiple columns in contour?
- How do I transform the data set into a dictionary inside the repo. I am using pyspark within foundry
Related Questions in FOUNDRY-CODE-REPOSITORIES
- How do I enforce a minimum test coverage percentage in my Foundry Code Repositories?
- How do I transform the data set into a dictionary inside the repo. I am using pyspark within foundry
- How to create python libraries and how to import it in palantir foundry
- Is adding a column description via write_dataframe aware of the branch it is run on?
- how to access the data frame without my_compute_function
- Python unit tests for Foundry's transforms?
- How do I compute my Foundry 'latest version' dataset faster?
- How do I make my many-join / many-union datasets compute faster?
- Is there a way to populate column descriptions specific to data set?
- How to revert/roll back to an earlier commit in Foundry Code Repo
- How to add a column as a file name to a parsed dataset in Palantir Foundry?
- Pass a whole dataset contains multiple files to HuggingFace function in Palantir
- How to upload dataset without authentication in Palantir foundry
- How can I copy code from one Code Repository to another in Foundry?
- In Palantir Foundry, how should I get the current SparkSession in a Transform?
Trending Questions
- UIImageView Frame Doesn't Reflect Constraints
- Is it possible to use adb commands to click on a view by finding its ID?
- How to create a new web character symbol recognizable by html/javascript?
- Why isn't my CSS3 animation smooth in Google Chrome (but very smooth on other browsers)?
- Heap Gives Page Fault
- Connect ffmpeg to Visual Studio 2008
- Both Object- and ValueAnimator jumps when Duration is set above API LvL 24
- How to avoid default initialization of objects in std::vector?
- second argument of the command line arguments in a format other than char** argv or char* argv[]
- How to improve efficiency of algorithm which generates next lexicographic permutation?
- Navigating to the another actvity app getting crash in android
- How to read the particular message format in android and store in sqlite database?
- Resetting inventory status after order is cancelled
- Efficiently compute powers of X in SSE/AVX
- Insert into an external database using ajax and php : POST 500 (Internal Server Error)
Popular Questions
- How do I undo the most recent local commits in Git?
- How can I remove a specific item from an array in JavaScript?
- How do I delete a Git branch locally and remotely?
- Find all files containing a specific text (string) on Linux?
- How do I revert a Git repository to a previous commit?
- How do I create an HTML button that acts like a link?
- How do I check out a remote Git branch?
- How do I force "git pull" to overwrite local files?
- How do I list all files of a directory?
- How to check whether a string contains a substring in JavaScript?
- How do I redirect to another webpage?
- How can I iterate over rows in a Pandas DataFrame?
- How do I convert a String to an int in Java?
- Does Python have a string 'contains' substring method?
- How do I check if a string contains a specific word?
If you check your platform documentation under
Code Repositories->Python Transforms->Python Unit Tests, you'll find quite a few resources there that will be helpful.The sections on writing and running tests in particular is what you're looking for.
// START DOCUMENTATION
Writing a Test
Full documentation can be found at https://docs.pytest.org
Pytest finds tests in any Python file that begins with test_. It is recommended to put all your tests into a test package under the src directory of your project. Tests are simply Python functions that are also named with the test_ prefix and assertions are made using Python’s assert statement. PyTest will also run tests written using Python’s builtin unittest module. For example, in transforms-python/src/test/test_increment.py a simple test would look like this:
Running this test will cause checks to fail with a message that looks like this:
Testing with PySpark
PyTest fixtures are a powerful feature that enables injecting values into test functions simply by adding a parameter of the same name. This feature is used to provide a spark_session fixture for use in your test functions. For example:
// END DOCUMENTATION
If you don't want to specify your schemas in code, you can also read in a file in your repository by following the instructions in documentation under
How To->Read file in Python repository// START DOCUMENTATION
Read file in Python repository
You can read other files from your repository into the transform context. This might be useful in setting parameters for your transform code to reference.
To start, In your python repository edit setup.py:
This tells python to bundle the yaml and csv files into the package. Then place a config file (for example config.yaml, but can be also csv or txt) next to your python transform (e.g. read_yml.py see below):
You can read it in your transform
read_yml.pywith the code below:So your project structure would be:
This will output in your dataset a single row with one column "result" with content:
// END DOCUMENTATION