I have been working on this problem for two days and still have not find the way.
Problem: Our Spark installed via newest CDH 5 always complains about the lost of LzoCodec class, even after I install the HADOOP_LZO through Parcels in cloudera manager. We are running MR1 on CDH 5.0.0-1.cdh5.0.0.p0.47.
Try to fix: The configurations in official CDH documentation about 'Using the LZO Parcel' are also added but the problem is still there.
Most of the googled posts give similar advices to the above. I also suspect that the spark is trying to run against YARN that is not activated there; but I can not find the configuration in CMF or other posts about this topic.
Please give me some help if you know how to deal with it.
Class com.hadoop.compression.lzo.LzoCodec not found
hadoops to me too ,when I query hive whith emr.Given others's solutions ,I take a shor-cut. in core-site.xml
delete this:
and delete
com.hadoop.compression.lzo.LzoCodec,com.hadoop.compression.lzo.LzopCodec
in this: