I want to train a agent to place a polyomino (only one for example a square of the measure of 2x2) on a 9x9 field, that is either empty or already contains multiple OTHER (not the 2x2 square one) polyominos. So the observation space would not be continuous. Is this a proper use case for RL?
Can I use Reinforcment Learning for a problem that has a non continous observation space?
66 Views Asked by FlorianFusseder At
1
There are 1 best solutions below
Related Questions in ALGORITHM
- MCNP 6 - Doubts about cells
- Given partially sorted array of type x<y => first apperance of x comes before first of y, sort in average O(n)
- What is the algorithm behind math.gcd and why it is faster Euclidean algorithm?
- Purpose of last 2 while loops in the merge algorithm of merge sort sorting technique
- Dots and Boxes with apha-beta pruning
- What is the average and worst-case time complexity of my string searching algorithm?
- Building a School Schedule Generator
- TC problem 5-2:how to calculate the probability of the indicator random variable?
- LCA of a binary tree implemented in Python
- Identify the checksum algorithm
- Algorithm for finding a subset of nodes in a weighted connected graph such that the distance between any pair nodes are under a postive number?
- Creating an efficent and time-saving algorithm to find difference between greater than and lesser than combination
- Algorithm to find neighbours of point by distance with no repeats
- Asking code suggestions about data structure and algorithm
- Heap sort with multithreading
Related Questions in MACHINE-LEARNING
- Trained ML model with the camera module is not giving predictions
- Keras similarity calculation. Enumerating distance between two tensors, which indicates as lists
- How to get content of BLOCK types LAYOUT_TITLE, LAYOUT_SECTION_HEADER and LAYOUT_xx in Textract
- How to predict input parameters from target parameter in a machine learning model?
- The training accuracy and the validation accuracy curves are almost parallel to each other. Is the model overfitting?
- ImportError: cannot import name 'HuggingFaceInferenceAPI' from 'llama_index.llms' (unknown location)
- Which library can replace causal_conv1d in machine learning programming?
- Fine-Tuning Large Language Model on PDFs containing Text and Images
- Sketch Guided Text to Image Generation
- My ICNN doesn't seem to work for any n_hidden
- Optuna Hyperband Algorithm Not Following Expected Model Training Scheme
- How can I resolve this error and work smoothly in deep learning?
- ModuleNotFoundError: No module named 'llama_index.node_parser'
- Difference between model.evaluate and metrics.accuracy_score
- Give Bert an input and ask him to predict. In this input, can Bert apply the first word prediction result to all subsequent predictions?
Related Questions in REINFORCEMENT-LEARNING
- pygame window is not shutting down with env.close()
- Recommended way to use Gymnasium with neural networks to avoid overheads in model.fit and model.predict
- Bellman equation for MRP?
- when I run the code "env = gym.make('LunarLander-v2')" in stable_baselines3 zoo
- Why the reward becomes smaller and smaller, thanks
- `multiprocessing.pool.starmap()` works wrong when I want to write my custom vector env for DRL
- mat1 and mat2 must have the same dtype, but got Byte and Float
- Stable-Baslines3 Type Error in _predict w. custom environment & policy
- is there any way to use RL for decoder only models
- How do I make sure I'm updating the Q-values correctly?
- Handling batch_size in a TorchRL environment
- Application of Welford algorithm to PPO agent training
- Finite horizon SARSA Lambda
- Custom Reinforcement Learning Environment with Neural Network
- Restored Policy gives action that is out of bound with RLlib
Related Questions in MACHINE-LEARNING-MODEL
- How to code in Java in to train a xml file with positive and negative examples image for an image recognition model by using HOG Features
- Flask API with TensorFlow Lite model always predicts the same class, regardless of input image
- Model Trainer Issue on End-to-End ML Project - TypeError: initiate_model_training() missing 4 required positional arguments
- Model Trainer Issue on End-to-End ML Project - TypeError: __init__() got an unexpected keyword argument 'config'
- Model Trainer Issue on End-to-End ML Project - TypeError: __init__() got an unexpected keyword argument 'trained_model_file_path'
- "The size of byte buffer and the shape do not match" erro in android studio with TensorFlow lite model
- Unable to Find Option for Exporting Custom ML Model for Object Detection in Google Vision (GCP Vertex AI)
- Wrong type of credentials for creating tuning model in quickstart
- Problem with GLM regressor in reusing it for different input data
- How to detect target object's offset from a reference point?
- Runtime Error while transfer learning a model using learn.fit_one_cycle(32)
- Machine Learning - Random Forest Classification - AxisError Axis 1 is out of bounds for array of dimension 1
- How create blending ensemble properly and call as ussual Libary, how to integrate in below my model
- How can I save an Azure ML sklearn model to a specific blob storage container?
- ModuleNotFoundError: No module named 'segmentation_models'
Trending Questions
- UIImageView Frame Doesn't Reflect Constraints
- Is it possible to use adb commands to click on a view by finding its ID?
- How to create a new web character symbol recognizable by html/javascript?
- Why isn't my CSS3 animation smooth in Google Chrome (but very smooth on other browsers)?
- Heap Gives Page Fault
- Connect ffmpeg to Visual Studio 2008
- Both Object- and ValueAnimator jumps when Duration is set above API LvL 24
- How to avoid default initialization of objects in std::vector?
- second argument of the command line arguments in a format other than char** argv or char* argv[]
- How to improve efficiency of algorithm which generates next lexicographic permutation?
- Navigating to the another actvity app getting crash in android
- How to read the particular message format in android and store in sqlite database?
- Resetting inventory status after order is cancelled
- Efficiently compute powers of X in SSE/AVX
- Insert into an external database using ajax and php : POST 500 (Internal Server Error)
Popular # Hahtags
Popular Questions
- How do I undo the most recent local commits in Git?
- How can I remove a specific item from an array in JavaScript?
- How do I delete a Git branch locally and remotely?
- Find all files containing a specific text (string) on Linux?
- How do I revert a Git repository to a previous commit?
- How do I create an HTML button that acts like a link?
- How do I check out a remote Git branch?
- How do I force "git pull" to overwrite local files?
- How do I list all files of a directory?
- How to check whether a string contains a substring in JavaScript?
- How do I redirect to another webpage?
- How can I iterate over rows in a Pandas DataFrame?
- How do I convert a String to an int in Java?
- Does Python have a string 'contains' substring method?
- How do I check if a string contains a specific word?
Sure, why not? The simplest versions of reinforcement learning algorithms use a discrete state space (and, indeed, assume for convergence that the agent is able to visit each state sufficiently many times). Even if there are too many states and you have to replace the Q function by a learned approximation (probably a neural net), you can use a one-hot encoding for the input.