I have setup of 1 Name Node, 2 Data Nodes, 1 Resource Manager and 2 Node Managers.All components are running as docker containers. Every time when I execute a spark submit (yarn cluster mode) from 2 machines (2 clients), job gets completed in a sequential manner. Job1 and Job2 both goes in Accepted state, Job1 turns to Running and Finished state and then Job2 gets picked and finishes its execution. Is there any way these jobs gets executed in parallel fashion? How does Application manager picks these tasks to give it to node manager?
Application manager in YARN setup
362 Views Asked by Aneri At
1
There are 1 best solutions below
Related Questions in HADOOP
- pcap to Avro on Hadoop
- schedule and automate sqoop import/export tasks
- How to diagnose Kafka topics failing globally to be found
- Only 32 bit available in Oracle VM - Hadoop Installation
- Using HDFS with Apache Spark on Amazon EC2
- How to get raw hadoop metrics
- How to output multiple values with the same key in reducer?
- Loading chararray from embedded JSON using Pig
- Oozie Pig action stuck in PREP state and job is in RUNNING state
- InstanceProfile is required for creating cluster - create python function to install module
- mapreduce job not setting compression codec correctly
- What does namespace and block pool mean in MapReduce 2.0 YARN?
- Hadoop distributed mode
- Building apache hadoop 2.6.0 throwing maven error
- I am using Hbase 1.0.0 and Apache phoenix 4.3.0 on CDH5.4. When I restart Hbase regionserver is down
Related Questions in APACHE-SPARK
- Spark .mapValues setup with multiple values
- Where do 'normal' println go in a scala jar, under Spark
- How to query JSON data according to JSON array's size with Spark SQL?
- How do I set the Hive user to something different than the Spark user from within a Spark program?
- How to add a new event to Apache Spark Event Log
- Spark streaming + kafka throughput
- dataframe or sqlctx (sqlcontext) generated "Trying to call a package" error
- Spark pairRDD not working
- How to know which worker a partition is executed at?
- Using HDFS with Apache Spark on Amazon EC2
- How to create a executable jar reading files from local file system
- How to keep a SQLContext instance alive in a spark streaming application's life cycle?
- Cassandra spark connector data loss
- Proper way to provide spark application a parameter/arg with spaces in spark-submit
- sorting RDD elements
Related Questions in DOCKER
- Docker, redirecting to virtualbox port
- Collect only from STDERR when using Docker syslog logging driver
- How can I create a docker image from the current system?
- Moving Docker Containers Around
- How can I test with serverspec that Jenkins is running in a jenkins docker container?
- How to deploy django 1.8 on Elastic Beanstalk using Docker
- Emulating `docker run` using the golang docker API
- Where are docker images and containers stored when we use it with Windows?
- docker compose, vagrant and insecure Repository
- Commit data in a mysql container
- oh-my-zsh installation returns non zero code
- Use custom docker binary in CoreOS
- Can I use docker image ubuntu 14.04 if my host is 12.04?
- Hide/obfuscate environmental parameters in docker
- How to add initial users when starting a RabbitMQ Docker container?
Related Questions in HADOOP-YARN
- Impala Resource Estimation for queries with Group by
- mapreduce job not setting compression codec correctly
- What does namespace and block pool mean in MapReduce 2.0 YARN?
- nmap does not show all open ports
- Spark resources not fully allocated on Amazon EMR
- Not able to format namenode in hadoop-2.6.0 multi node installation
- MapReduce job fails with ExitCodeException exitCode=255
- Spark GraphX memory out of error SparkListenerBus (java.lang.OutOfMemoryError: Java heap space)
- Yarn autodetect slaves failure
- how to ignore key-value pair in Map-Reduce if values are blank?
- What happens to orphaned Yarn Child processes?
- Setting mapreduce.job.jar with a jar in path containing space
- Hive index mapreduce memory errors
- YARN log aggregation on a per job basis
- Importtsv command gives : Container exited with a non-zero exit code 1 error
Related Questions in APPLICATIONMANAGER
- How long an application is used in BlackBerry
- Monitor Cassandra Database Remotely using Application Manager 14
- Android: Broadcast receiver for any app launch or if any app comes to foreground
- How to use ApplicationManager to load an object with out GAC'ing it?
- Application manager in YARN setup
- How do you call your enviroment variables set in application manager using cpanel and PHP
- How to detect if user force stop my service
- Is there any class on BlackBerry similar to DevicePolicyManager on Android?
- The name 'DefaultAuthenticationTypes' does not exist in the current context
- Unable To Start Opera Mini With The Given URL
- Why wouldn't application manager run trouble script?
Trending Questions
- UIImageView Frame Doesn't Reflect Constraints
- Is it possible to use adb commands to click on a view by finding its ID?
- How to create a new web character symbol recognizable by html/javascript?
- Why isn't my CSS3 animation smooth in Google Chrome (but very smooth on other browsers)?
- Heap Gives Page Fault
- Connect ffmpeg to Visual Studio 2008
- Both Object- and ValueAnimator jumps when Duration is set above API LvL 24
- How to avoid default initialization of objects in std::vector?
- second argument of the command line arguments in a format other than char** argv or char* argv[]
- How to improve efficiency of algorithm which generates next lexicographic permutation?
- Navigating to the another actvity app getting crash in android
- How to read the particular message format in android and store in sqlite database?
- Resetting inventory status after order is cancelled
- Efficiently compute powers of X in SSE/AVX
- Insert into an external database using ajax and php : POST 500 (Internal Server Error)
Popular Questions
- How do I undo the most recent local commits in Git?
- How can I remove a specific item from an array in JavaScript?
- How do I delete a Git branch locally and remotely?
- Find all files containing a specific text (string) on Linux?
- How do I revert a Git repository to a previous commit?
- How do I create an HTML button that acts like a link?
- How do I check out a remote Git branch?
- How do I force "git pull" to overwrite local files?
- How do I list all files of a directory?
- How to check whether a string contains a substring in JavaScript?
- How do I redirect to another webpage?
- How can I iterate over rows in a Pandas DataFrame?
- How do I convert a String to an int in Java?
- Does Python have a string 'contains' substring method?
- How do I check if a string contains a specific word?
The cluster setup is using YARN
Capacity Scheduler, which is default in most of the available Hadoop distributions. If multiple jobs are submitted by the same user, they enter the same user queue which follows FIFO. This is the default behaviour of capacity scheduler.Fair Schedulercan be configured to run jobs in parallel by sharing the available resources.Add this property to
yarn-site.xmlConfigure the fair scheduler
queuesin an allocation file,If this property is not configured, a queue per user will be created by default.