I have deployed an Alluxio cluster based on HDFS cluster. When I use Alluxio Native Java Api to write some files to Alluxio with setting writetype ASYNC_THROUGH, the files (even just having 1G) seem not be written to HDFS, keep a state of TO_BE_PERSISTED all the time after few days.
why alluxio files keep a state of TO_BE_PERSISTED all the time
373 Views Asked by Long.zhao At
1
There are 1 best solutions below
Related Questions in HADOOP
- Can anyoone help me with this problem while trying to install hadoop on ubuntu?
- Hadoop No appenders could be found for logger (org.apache.hadoop.mapreduce.v2.app.MRAppMaster)
- Top-N using Python, MapReduce
- Spark Driver vs MapReduce Driver on YARN
- ERROR: org.apache.hadoop.fs.UnsupportedFileSystemException: No FileSystem for scheme "maprfs"
- can't write pyspark dataframe to parquet file on windows
- How to optimize writing to a large table in Hive/HDFS using Spark
- Can't replicate block xxx because the block file doesn't exist, or is not accessible
- HDFS too many bad blocks due to "Operation category WRITE is not supported in state standby" - Understanding why datanode can't find Active NameNode
- distcp throws java.io.IOException when copying files
- Hadoop MapReduce WordPairsCount produces inconsistent results
- If my data is not partitioned can that be why I’m getting maxResultSize error for my PySpark job?
- resource manager and nodemanager connectivity issues
- ERROR flume.SinkRunner: Unable to deliver event
- converting varchar(7) to decimal (7,5) in hive
Related Questions in HDFS
- Can anyoone help me with this problem while trying to install hadoop on ubuntu?
- ERROR: org.apache.hadoop.fs.UnsupportedFileSystemException: No FileSystem for scheme "maprfs"
- How to optimize writing to a large table in Hive/HDFS using Spark
- Update hadoop hadoop-2.6.5 to haddop 3.x. Operation category WRITE is not supported in state standby
- Copy/Merge multiple HDFS files using Nifi Processor
- HDFS too many bad blocks due to "Operation category WRITE is not supported in state standby" - Understanding why datanode can't find Active NameNode
- distcp throws java.io.IOException when copying files
- ERROR flume.SinkRunner: Unable to deliver event
- Apache flume does not run hadoop 3.1.0 Flume 1.11
- Livy session to submit pyspark from HDFS
- ClickHouse Server Exception: Code: 210.DB::Exception: Fail to read from HDFS:
- Confluent HDFS Sink connector error while connecting HDFS to Hive
- Node Transitioned from NEW to UNHEALTHY and Attempting to remove non-existent node
- Error associated with Azure Datalake Gen2 and Hadoop connection
- How do I directly read files from HDFS using dask?
Related Questions in ALLUXIO
- Trino Hive connector can't synchronize the partition metadata automatically
- Alluxio - How to check log4j version used in Alluxio release (alluxio-2.8.1)
- How to get rid of warnings with MEM and SSD tiers
- How to check if folder is cached in Alluxio?
- How can I have Alluxio show all the not-yet-accessed files in the directory?
- setting up Alluxio cluster, path points to Java 8 but Alluxio thinks it's java 7?
- difference between WORKER_EVICTOR and WORKER_BLOCK_ANNOTATOR
- How can I maintain a list of constant masters and workers under conf/masters and conf/workers in a managed Scaling cluster?
- Is hadoop3.1.1 not supported?
- Mount HDFS Data on Alluxio
- Alluxio on kubernetes(EKS) supports workers only with DaemonSet kind? Or we can deploy workers using Deployment kind and specify number of workers?
- Is there any way I can avoid specifying the port in the alluxio fs uri?
- How do I get Alluxio POSIX to run with version 3 libfuse?
- Will journal be lost if there is no time to flush to the journal writer before jvm hangs?
- Is the s3 connection configurable with role arn alone?
Trending Questions
- UIImageView Frame Doesn't Reflect Constraints
- Is it possible to use adb commands to click on a view by finding its ID?
- How to create a new web character symbol recognizable by html/javascript?
- Why isn't my CSS3 animation smooth in Google Chrome (but very smooth on other browsers)?
- Heap Gives Page Fault
- Connect ffmpeg to Visual Studio 2008
- Both Object- and ValueAnimator jumps when Duration is set above API LvL 24
- How to avoid default initialization of objects in std::vector?
- second argument of the command line arguments in a format other than char** argv or char* argv[]
- How to improve efficiency of algorithm which generates next lexicographic permutation?
- Navigating to the another actvity app getting crash in android
- How to read the particular message format in android and store in sqlite database?
- Resetting inventory status after order is cancelled
- Efficiently compute powers of X in SSE/AVX
- Insert into an external database using ajax and php : POST 500 (Internal Server Error)
Popular # Hahtags
Popular Questions
- How do I undo the most recent local commits in Git?
- How can I remove a specific item from an array in JavaScript?
- How do I delete a Git branch locally and remotely?
- Find all files containing a specific text (string) on Linux?
- How do I revert a Git repository to a previous commit?
- How do I create an HTML button that acts like a link?
- How do I check out a remote Git branch?
- How do I force "git pull" to overwrite local files?
- How do I list all files of a directory?
- How to check whether a string contains a substring in JavaScript?
- How do I redirect to another webpage?
- How can I iterate over rows in a Pandas DataFrame?
- How do I convert a String to an int in Java?
- Does Python have a string 'contains' substring method?
- How do I check if a string contains a specific word?
I find the answer. I find the log "ERROR DefaultAsyncPersistHandler - Not all the blocks of file /test/sample_data/order_detail_titile_2.0.txt stored on the same worker", When using ASYNC_THROUGH, Alluxio request all file blockers in one worker. There is no problem where I use CACHE_THROUGH