I am currently using wordcount application in hadoop as a benchmark. I find that the cpu usage is fairly nearly constant around 80-90%. I would like to have a fluctuating cpu usage. Is there any hadoop application that can give me this capability? Thanks a lot.
1
There are 1 best solutions below
Related Questions in HADOOP
- pcap to Avro on Hadoop
- schedule and automate sqoop import/export tasks
- How to diagnose Kafka topics failing globally to be found
- Only 32 bit available in Oracle VM - Hadoop Installation
- Using HDFS with Apache Spark on Amazon EC2
- How to get raw hadoop metrics
- How to output multiple values with the same key in reducer?
- Loading chararray from embedded JSON using Pig
- Oozie Pig action stuck in PREP state and job is in RUNNING state
- InstanceProfile is required for creating cluster - create python function to install module
- mapreduce job not setting compression codec correctly
- What does namespace and block pool mean in MapReduce 2.0 YARN?
- Hadoop distributed mode
- Building apache hadoop 2.6.0 throwing maven error
- I am using Hbase 1.0.0 and Apache phoenix 4.3.0 on CDH5.4. When I restart Hbase regionserver is down
Related Questions in CPU
- 1MiB = 1024KiB = 2^10. Nonetheless, why not use just 1000 byte instead 1024 to calculate size?
- What is the simplest Turing complete CPU instruction set which can execute code from ROM?
- How to get CPU utilization in % in terminal (mac)
- Avoiding CPU Contention
- Lots of cache miss, Sparse matrix multiplication
- CPU new features enabled in Linux kernel
- Are correct branch predictions free?
- NUMA support on which CPU? What are the current server configuration of this kind of CPU?
- How to deal with virtual address when trying to get memory access pattern statistics?
- On x86, does enabling paging cause an "unconditional jump" (since EIP is now a virtual address)?
- cpu load when setting textbox value
- CPU usage exceeding 100% in top command third line
- 32bit cpu: how much memory can it use?
- CMOS Scaling vs Die Shrink
- Meaning of cores and logical processors in intel icore
Related Questions in WORKLOAD
- How can i learn if am using all of my cores in the maximum level
- MPI blocks execution after send when different workloads associated to a processor are used
- Python Multi-threading CPU workload
- How do we add the Job Submit Plugin API in Slurm running on Ubuntu 20.04?
- How to create a kubernetes workload and expose it in same command line?
- Limits of workload that can be put into hardware accelerators
- How to set up Network on Rancher UI for multi workload app
- A Workload Scheduler's step doesn't proceed and remains queued status
- Database script encounterd "AWKDBE018E Cannot access required JDBC Driver folder" in Workload Scheduler
- Is there any way to pass user variables to a CA Workload Automation job?
- What units can be used to benchmark CPU usage? Percentage seems unhelpful
- Hadoop workload
- Throughput Shaping timer- Externalising
- How can I get excel files or CSV for real case of using thingsboard
- View Autosys job status as dependency diagram
Trending Questions
- UIImageView Frame Doesn't Reflect Constraints
- Is it possible to use adb commands to click on a view by finding its ID?
- How to create a new web character symbol recognizable by html/javascript?
- Why isn't my CSS3 animation smooth in Google Chrome (but very smooth on other browsers)?
- Heap Gives Page Fault
- Connect ffmpeg to Visual Studio 2008
- Both Object- and ValueAnimator jumps when Duration is set above API LvL 24
- How to avoid default initialization of objects in std::vector?
- second argument of the command line arguments in a format other than char** argv or char* argv[]
- How to improve efficiency of algorithm which generates next lexicographic permutation?
- Navigating to the another actvity app getting crash in android
- How to read the particular message format in android and store in sqlite database?
- Resetting inventory status after order is cancelled
- Efficiently compute powers of X in SSE/AVX
- Insert into an external database using ajax and php : POST 500 (Internal Server Error)
Popular Questions
- How do I undo the most recent local commits in Git?
- How can I remove a specific item from an array in JavaScript?
- How do I delete a Git branch locally and remotely?
- Find all files containing a specific text (string) on Linux?
- How do I revert a Git repository to a previous commit?
- How do I create an HTML button that acts like a link?
- How do I check out a remote Git branch?
- How do I force "git pull" to overwrite local files?
- How do I list all files of a directory?
- How to check whether a string contains a substring in JavaScript?
- How do I redirect to another webpage?
- How can I iterate over rows in a Pandas DataFrame?
- How do I convert a String to an int in Java?
- Does Python have a string 'contains' substring method?
- How do I check if a string contains a specific word?
I don't think there's a way to throttle or specify a range for hadoop to use. Hadoop will use the CPU available to it. When I'm running a lot of jobs, I'm constantly in the 90%+ range.
One way you can control the CPU usage is to change the maximum number of mappers/reducers each tasktracker can run simultaneously. This is done through the
mapred.tasktracker.{map|reduce}.tasks.maximumsetting in$HADOOP_HOME/conf/core-site.xml.It will use less CPU on that tasktracker when the number of mapper/reducers is limited.
Another way is to set the configuration value for
mapred.tasktracker.{map|reduce}.taskswhen setting up the job. This will force that job to use that many mappers/reducers. This number will be split across the available tasktrackers, so if you have 4 nodes and want each node to have 1 mapper you'd setmapred.tasktracker.map.tasksto4. It's also possible that if a node can run 4 mappers, it will run all 4, I don't know exactly how hadoop will split out the tasks, but forcing a number, per job, is an option.I hope that helps get you to where you're going. I still don't quite understand what you are looking for. :)