I understood the concept of automatic differentiation, but couldn't find any explanation how tensorflow calculates the error gradient for non differentiable functions as for example tf.where in my loss function or tf.cond in my graph. It works just fine, but I would like to understand how tensorflow backpropagates the error through such nodes, since there is no formula to calculate the gradient from them.
How does tensorflow handle non differentiable nodes during gradient calculation?
1.6k Views Asked by Natjo At
1
There are 1 best solutions below
Related Questions in PYTHON
- How to store a date/time in sqlite (or something similar to a date)
- Instagrapi recently showing HTTPError and UnknownError
- How to Retrieve Data from an MySQL Database and Display it in a GUI?
- How to create a regular expression to partition a string that terminates in either ": 45" or ",", without the ": "
- Python Geopandas unable to convert latitude longitude to points
- Influence of Unused FFN on Model Accuracy in PyTorch
- Seeking Python Libraries for Removing Extraneous Characters and Spaces in Text
- Writes to child subprocess.Popen.stdin don't work from within process group?
- Conda has two different python binarys (python and python3) with the same version for a single environment. Why?
- Problem with add new attribute in table with BOTO3 on python
- Can't install packages in python conda environment
- Setting diagonal of a matrix to zero
- List of numbers converted to list of strings to iterate over it. But receiving TypeError messages
- Basic Python Question: Shortening If Statements
- Python and regex, can't understand why some words are left out of the match
Related Questions in TENSORFLOW
- A deterministic GPU implementation of fused batch-norm backprop, when training is disabled, is not currently available
- Keras similarity calculation. Enumerating distance between two tensors, which indicates as lists
- Does tensorflow have a way of calculating input importance for simple neural networks
- How to predict input parameters from target parameter in a machine learning model?
- Windows 10 TensorFlow cannot detect Nvidia GPU
- unable to use ignore_class in SparseCategoricalCrossentropy
- Why is this code not working? I've tried everything and everything seems to be fine, but no
- Why convert jpeg into tfrecords?
- ValueError: The shape of the target variable and the shape of the target value in `variable.assign(value)` must match
- The kernel appears to have died. It will restart automatically. whenever i try to run the plt.imshow() and plt.show() function in jupyter notebook
- Pneumonia detection, using transfer learning
- Cannot install tensorflow ver 2.3.0 (distribution not found)
- AttributeError: module 'keras._tf_keras.keras.layers' has no attribute 'experimental'
- Error while loading .keras model: Layer node index out of bounds
- prediction model with python tensorflow and keras, gives error when predicting
Related Questions in BACKPROPAGATION
- Why doesn't this code work? - Backpropagation algorithm
- SymPy - Can't calculate derivative wrt expression, is there an alternative for intermediate expressions?
- Siamese Network Backpropagation
- Locally blocking gradient update for nested neural network
- Where exactly are TPUs used during machine learning?
- How to implement gradient op for a custom tensorflow op, for which the it is hard to derive a mathematical closed form formula for gradient?
- Does padded rows (fake inputs) affect backpropagation?
- Backpropagation for Two Different Neural Network Models with Combined Loss Functions
- Backpropagation through composition of neural networks where one has fixed parameters
- Neural Network backpropagation algorithm only partially training in python
- Neural Network works with ReLu bot not sigmoid
- Backpropagation and gradient descent with python
- Skipping backpropagation for certain element of certain data with stop_recording() in tf.GradientTape()
- How can I make a one neuron neural network?
- My one neuron neural network does not work with my dataset
Related Questions in AUTOMATIC-DIFFERENTIATION
- Optimizing an autograd + list comprehension python operation
- How to generate jacobian of a tensor-valued function using torch.autograd?
- How to implement Carleman Matrix in Haskell?
- Calculating two gradients in pytorch and reusing an intermediate gradient
- solving an ODE using neural networks
- How to preallocate using JacobianConfig for "Hessian of vector valued function" double Jacobian in Julia ForwardDiff package
- JAX `custom_vjp` for functions with multiple outputs
- JAX `vjp` fails for vmapped function with `custom_vjp`
- JAX `vjp` does not recognize cotangent argument with `custom_vjp`
- Failing to return gradients
- How can I get exhaust list of functions available to grad_fn for backward() function in Pytorch?
- Accurately finding the gradient of values from 2 vectors
- Back propagation in pytorch to calculate component wise derivatives of vector
- What is the correct way of defining a differentiable function with scalar arguments that returns a vector/array?
- How to compute in tensorflow the partial derivative of each observation from the output w.r.t each input observation
Trending Questions
- UIImageView Frame Doesn't Reflect Constraints
- Is it possible to use adb commands to click on a view by finding its ID?
- How to create a new web character symbol recognizable by html/javascript?
- Why isn't my CSS3 animation smooth in Google Chrome (but very smooth on other browsers)?
- Heap Gives Page Fault
- Connect ffmpeg to Visual Studio 2008
- Both Object- and ValueAnimator jumps when Duration is set above API LvL 24
- How to avoid default initialization of objects in std::vector?
- second argument of the command line arguments in a format other than char** argv or char* argv[]
- How to improve efficiency of algorithm which generates next lexicographic permutation?
- Navigating to the another actvity app getting crash in android
- How to read the particular message format in android and store in sqlite database?
- Resetting inventory status after order is cancelled
- Efficiently compute powers of X in SSE/AVX
- Insert into an external database using ajax and php : POST 500 (Internal Server Error)
Popular # Hahtags
Popular Questions
- How do I undo the most recent local commits in Git?
- How can I remove a specific item from an array in JavaScript?
- How do I delete a Git branch locally and remotely?
- Find all files containing a specific text (string) on Linux?
- How do I revert a Git repository to a previous commit?
- How do I create an HTML button that acts like a link?
- How do I check out a remote Git branch?
- How do I force "git pull" to overwrite local files?
- How do I list all files of a directory?
- How to check whether a string contains a substring in JavaScript?
- How do I redirect to another webpage?
- How can I iterate over rows in a Pandas DataFrame?
- How do I convert a String to an int in Java?
- Does Python have a string 'contains' substring method?
- How do I check if a string contains a specific word?
In the case of
tf.where, you have a function with three inputs, conditionC, value on trueTand value on falseF, and one outputOut. The gradient receives one value and has to return three values. Currently, no gradient is computed for the condition (that would hardly make sense), so you just need to do the gradients forTandF. Assuming the input and the outputs are vectors, imagineC[0]isTrue. ThenOut[0]comes fromT[0], and its gradient should propagate back. On the other hand,F[0]would have been discarded, so its gradient should be made zero. IfOut[1]wereFalse, then the gradient forF[1]should propagate but not forT[1]. So, in short, forTyou should propagate the given gradient whereCisTrueand make it zero where it isFalse, and the opposite forF. If you look at the implementation of the gradient oftf.where(Selectoperation), it does exactly that:Note the input values themselves are not used in the computation, that will be done by the gradients of the operation producing those inputs. For
tf.cond, the code is a bit more complicated, because the same operation (Merge) is used in different contexts, and alsotf.condalso usesSwitchoperations inside. However the idea is the same. Essentially,Switchoperations are used for each input, so the input that was activated (the first if the condition wasTrueand the second otherwise) gets the received gradient and the other input gets a "switched off" gradient (likeNone), and does not propagate back further.