日韩性视频-久久久蜜桃-www中文字幕-在线中文字幕av-亚洲欧美一区二区三区四区-撸久久-香蕉视频一区-久久无码精品丰满人妻-国产高潮av-激情福利社-日韩av网址大全-国产精品久久999-日本五十路在线-性欧美在线-久久99精品波多结衣一区-男女午夜免费视频-黑人极品ⅴideos精品欧美棵-人人妻人人澡人人爽精品欧美一区-日韩一区在线看-欧美a级在线免费观看

歡迎訪問 生活随笔!

生活随笔

當(dāng)前位置: 首頁 > 编程资源 > 编程问答 >内容正文

编程问答

[转]hadoop新手错误解决方法

發(fā)布時(shí)間:2023/12/2 编程问答 30 豆豆
生活随笔 收集整理的這篇文章主要介紹了 [转]hadoop新手错误解决方法 小編覺得挺不錯(cuò)的,現(xiàn)在分享給大家,幫大家做個(gè)參考.

解決Exception: org.apache.hadoop.io.nativeio.NativeIO$Windows.access0(Ljava/lang/String;I)Z 等一系列問題,ljavalangstring


??

一.簡介

? ?Windows下的 Eclipse上調(diào)試Hadoop2代碼,所以我們在windows下的Eclipse配置hadoop-eclipse-plugin-2.6.0.jar插件,并在運(yùn)行Hadoop代碼時(shí)出現(xiàn)了一系列的問題,搞了好幾天終于能運(yùn)行起代碼。接下來我們來看看問題并怎么解決,提供給跟我同樣遇到的問題作為參考。

?

? Hadoop2的WordCount.java統(tǒng)計(jì)代碼如下:

? ? ?

import java.io.IOException; import java.util.StringTokenizer;import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.Path; import org.apache.hadoop.io.IntWritable; import org.apache.hadoop.io.Text; import org.apache.hadoop.mapreduce.Job; import org.apache.hadoop.mapreduce.Mapper; import org.apache.hadoop.mapreduce.Reducer; import org.apache.hadoop.mapreduce.lib.input.FileInputFormat; import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;public class WordCount {public static class TokenizerMapperextends Mapper<Object, Text, Text, IntWritable>{private final static IntWritable one = new IntWritable(1);private Text word = new Text();public void map(Object key, Text value, Context context) throws IOException, InterruptedException {StringTokenizer itr = new StringTokenizer(value.toString());while (itr.hasMoreTokens()) {word.set(itr.nextToken());context.write(word, one);}}}public static class IntSumReducerextends Reducer<Text,IntWritable,Text,IntWritable> {private IntWritable result = new IntWritable();public void reduce(Text key, Iterable<IntWritable> values,Context context) throws IOException, InterruptedException {int sum = 0;for (IntWritable val : values) {sum += val.get();}result.set(sum);context.write(key, result);}}public static void main(String[] args) throws Exception {Configuration conf = new Configuration();Job job = Job.getInstance(conf, "word count");job.setJarByClass(WordCount.class);job.setMapperClass(TokenizerMapper.class);job.setCombinerClass(IntSumReducer.class);job.setReducerClass(IntSumReducer.class);job.setOutputKeyClass(Text.class);job.setOutputValueClass(IntWritable.class);FileInputFormat.addInputPath(job, new Path(args[0]));FileOutputFormat.setOutputPath(job, new Path(args[1]));System.exit(job.waitForCompletion(true) ? 0 : 1);} }

?

?

問題一.An internal error occurred during: "Map/Reducelocation status updater".java.lang.NullPointerException

?

?

? ?我們hadoop-eclipse-plugin-2.6.0.jar放到Eclipse的plugins目錄下,我們的Eclipse目錄是F:\tool\eclipse-jee-juno-SR2\eclipse-jee-juno-SR2\plugins,重啟一下Eclipse,然后,打開Window-->Preferens,可以看到Hadoop Map/Reduc選項(xiàng),然后點(diǎn)擊出現(xiàn)了An internal error occurredduring: "Map/Reduce location status updater".java.lang.NullPointerException,如圖所示:

? ?

?

? 解決:

? ?我們發(fā)現(xiàn)剛配置部署的Hadoop2還沒創(chuàng)建輸入和輸出目錄,先在hdfs上建個(gè)文件夾?。

? ?#bin/hdfs dfs -mkdir –p?/user/root/input

? ?#bin/hdfs dfs -mkdir -p ?/user/root/output

?我們在Eclipse的DFS Locations目錄下看到我們這兩個(gè)目錄,如圖所示:

??

問題二.Exception in thread "main" java.lang.NullPointerException atjava.lang.ProcessBuilder.start(Unknown Source) ??

運(yùn)行Hadoop2的WordCount.java代碼時(shí)出現(xiàn)了這樣錯(cuò)誤,

? ? ?

log4j:WARNPlease initialize the log4j system properly. log4j:WARN Seehttp://logging.apache.org/log4j/1.2/faq.html#noconfig for more info. Exception in thread "main" java.lang.NullPointerExceptionatjava.lang.ProcessBuilder.start(Unknown Source)atorg.apache.hadoop.util.Shell.runCommand(Shell.java:482)atorg.apache.hadoop.util.Shell.run(Shell.java:455)atorg.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715)atorg.apache.hadoop.util.Shell.execCommand(Shell.java:808)atorg.apache.hadoop.util.Shell.execCommand(Shell.java:791)at

?

?

?

分析:

? 下載Hadoop2以上版本時(shí),在Hadoop2的bin目錄下沒有winutils.exe

解決:

? 1.下載https://codeload.github.com/srccodes/hadoop-common-2.2.0-bin/zip/master下載hadoop-common-2.2.0-bin-master.zip,然后解壓后,把hadoop-common-2.2.0-bin-master下的bin全部復(fù)制放到我們下載的Hadoop2的binHadoop2/bin目錄下。如圖所示:

?????

? 2.Eclipse-》window-》Preferences 下的Hadoop Map/Peduce 把下載放在我們的磁盤的Hadoop目錄引進(jìn)來,如圖所示:

? ??

?

? 3.Hadoop2配置變量環(huán)境HADOOP_HOME 和path,如圖所示:

?

?問題三.Exception in thread "main"java.lang.UnsatisfiedLinkError:org.apache.hadoop.io.nativeio.NativeIO$Windows.access0(Ljava/lang/String;I)Z

?

? 當(dāng)我們解決了問題三時(shí),在運(yùn)行WordCount.java代碼時(shí),出現(xiàn)這樣的問題

? ??

log4j:WARN No appenders could be found forlogger (org.apache.hadoop.metrics2.lib.MutableMetricsFactory). log4j:WARN Please initialize the log4jsystem properly. log4j:WARN Seehttp://logging.apache.org/log4j/1.2/faq.html#noconfig for more info. Exception in thread "main"java.lang.UnsatisfiedLinkError:org.apache.hadoop.io.nativeio.NativeIO$Windows.access0(Ljava/lang/String;I)Zatorg.apache.hadoop.io.nativeio.NativeIO$Windows.access0(Native Method)atorg.apache.hadoop.io.nativeio.NativeIO$Windows.access(NativeIO.java:557)atorg.apache.hadoop.fs.FileUtil.canRead(FileUtil.java:977)atorg.apache.hadoop.util.DiskChecker.checkAccessByFileMethods(DiskChecker.java:187)atorg.apache.hadoop.util.DiskChecker.checkDirAccess(DiskChecker.java:174)atorg.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:108)atorg.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.confChanged(LocalDirAllocator.java:285)atorg.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrite(LocalDirAllocator.java:344)atorg.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:150)atorg.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:131)atorg.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:115)atorg.apache.hadoop.mapred.LocalDistributedCacheManager.setup(LocalDistributedCacheManager.java:131)

?

?分析:

? ? C:\Windows\System32下缺少hadoop.dll,把這個(gè)文件拷貝到C:\Windows\System32下面即可。

?解決:

? ? hadoop-common-2.2.0-bin-master下的bin的hadoop.dll放到C:\Windows\System32下,然后重啟電腦,也許還沒那么簡單,還是出現(xiàn)這樣的問題。

?

??我們在繼續(xù)分析:

? ? 我們在出現(xiàn)錯(cuò)誤的的atorg.apache.hadoop.io.nativeio.NativeIO$Windows.access(NativeIO.java:557)我們來看這個(gè)類NativeIO的557行,如圖所示:

? ?????

?

? ?Windows的唯一方法用于檢查當(dāng)前進(jìn)程的請求,在給定的路徑的訪問權(quán)限,所以我們先給以能進(jìn)行訪問,我們自己先修改源代碼,return true 時(shí)允許訪問。我們下載對應(yīng)hadoop源代碼,hadoop-2.6.0-src.tar.gz解壓,hadoop-2.6.0-src\hadoop-common-project\hadoop-common\src\main\java\org\apache\hadoop\io\nativeio下NativeIO.java 復(fù)制到對應(yīng)的Eclipse的project,然后修改557行為return true如圖所示:

??

?

? ?

??問題四:org.apache.hadoop.security.AccessControlException: Permissiondenied: user=zhengcy, access=WRITE,inode="/user/root/output":root:supergroup:drwxr-xr-x

?

? 我們在執(zhí)行運(yùn)行WordCount.java代碼時(shí),出現(xiàn)這樣的問題

? ??

2014-12-18 16:03:24,092 WARN (org.apache.hadoop.mapred.LocalJobRunner:560) - job_local374172562_0001 org.apache.hadoop.security.AccessControlException: Permission denied: user=zhengcy, access=WRITE, inode="/user/root/output":root:supergroup:drwxr-xr-xat org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkFsPermission(FSPermissionChecker.java:271)at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:257)at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:238)at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:179)at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6512)at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6494)at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:6446)at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInternal(FSNamesystem.java:4248)at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInt(FSNamesystem.java:4218)at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:4191)at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:813)

? ?

?

?分析:

? 我們沒權(quán)限訪問output目錄。

解決:

? ? 我們 在設(shè)置hdfs配置的目錄是在hdfs-site.xml配置hdfs文件存放的地方,我在hadoop偽分布式部署那邊有介紹過,我們在這邊在復(fù)習(xí)一下,如圖所示:

我們在這個(gè)etc/hadoop下的hdfs-site.xml添加

??<property>?

? ? ?<name>dfs.permissions</name>? ? ? ?<value>false</value>? ??</property>

設(shè)置沒有權(quán)限,不過我們在正式的 服務(wù)器上不能這樣設(shè)置。

?

? 問題五:File/usr/root/input/file01._COPYING_ could only be replicated to 0 nodes instead ofminRepLication (=1) There are 0 datanode(s) running and no node(s) are excludedin this operation

? ? ?如圖所示:

? ? ??

??分析: ?

? 我們在第一次執(zhí)行#hadoop namenode –format 完然后在執(zhí)行#sbin/start-all.sh?

在執(zhí)行#jps,能看到Datanode,在執(zhí)行#hadoop namenode –format然后執(zhí)行#jps這時(shí)看不到Datanode ,如圖所示:

? ? ??

? ?然后我們想把文本放到輸入目錄執(zhí)行bin/hdfs dfs -put/usr/local/hadoop/hadoop-2.6.0/test/* /user/root/input ?把/test/*文件上傳到hdfs的/user/root/input中,出現(xiàn)這樣的問題,

?解決:

? 是我們執(zhí)行太多次了hadoopnamenode –format,在創(chuàng)建了多個(gè),我們對應(yīng)的hdfs目錄刪除hdfs-site.xml配置的保存datanode和namenode目錄。

轉(zhuǎn)載于:https://www.cnblogs.com/julyme/p/4657404.html

總結(jié)

以上是生活随笔為你收集整理的[转]hadoop新手错误解决方法的全部內(nèi)容,希望文章能夠幫你解決所遇到的問題。

如果覺得生活随笔網(wǎng)站內(nèi)容還不錯(cuò),歡迎將生活随笔推薦給好友。