先申明一下这纯属于自己个人成长的历程有,希望写博客来记录这自己的成长,同时分享自己遇到的问题,和大家交流。首先因为最近的需求,开始着手安装windows下的SPARK。我用的是spark-2.2.0+hadoop-2.8.1+scala.
首先下载hadoop以及scala.mis。这两个的安装比较简单:第一:hadoop需要配置环境变量:下面是我的配置的图:
第一个是:HADOOP_HOME ;hadoop的解压目录,若没有这目录会出现一下错误:
15/06/11
15:35:50
ERROR Shell: Failed to locate the winutils binary in the hadoop binary path
java.io.IOException: Could not locate executable
null\bin\winutils.exe in the Hadoop binaries.
at org.apache.hadoop.util.Shell.getQualifiedBinPath(Shell.java:356)
at org.apache.hadoop.util.Shell.getWinUtilsPath(Shell.java:371)
at org.apache.hadoop.util.Shell.<clinit>(Shell.java:364)
at org.apache.hadoop.util.StringUtils.<clinit>(StringUtils.java:80)
at org.apache.hadoop.security.SecurityUtil.getAuthenticationMethod(SecurityUtil.java:611)
at org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:272)
at org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:260)
at org.apache.hadoop.security.UserGroupInformation.loginUserFromSubject(UserGroupInformation.java:790)
at org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:760)
at org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:633)
at org.apache.spark.util.Utils$$anonfun$getCurrentUserName$1.apply(Utils.scala:2001)
at org.apache.spark.util.Utils$$anonfun$getCurrentUserName$1.apply(Utils.scala:2001)
at scala.Option.getOrElse(Option.scala:120)
at org.apache.spark.util.Utils$.getCurrentUserName(Utils.scala:2001)
at org.apache.spark.SecurityManager.<init>(SecurityManager.scala:207)
at org.apache.spark.SparkEnv$.create(SparkEnv.scala:218)
at org.apache.spark.SparkEnv$.createDriverEnv(SparkEnv.scala:163)
at org.apache.spark.SparkContext.createSparkEnv(SparkContext.scala:269)
at org.apache.spark.SparkContext.<init>(SparkContext.scala:272)
at org.apache.spark.SparkContext.<init>(SparkContext.scala:154)
at SparkFromHbase$.main(SparkFromHbase.scala:15)
at SparkFromHbase.main(SparkFromHbase.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at com.intellij.rt.execution.application.AppMain.main(AppMain.java:134)
第二个是hadoopd的bin目录;
双击运行scala.mis后在环境变量添加scala的bin目录。如上图所示。
最后是spark-2.2.0的安装,只需下载解压包。把解压后spark的bin目录添加到环境变量中如上图所示;
SPARK的下载地址:http://spark.apache.org/downloads.html,找到适合自己的版本下载就好,这里我使用的是Pre-built的版本,意思就是已经编译了好了,下载来直接用就好。
最后因为我是添加在系统盘C的 所以只能已管理员的身份运行spark-shell;
看到下面两行就说明成功了。
第一次写博客,希望各位大神,大牛提出宝贵意见,以交流为主,只是安装了SPARK,后续会不定期更新SPARK的练习与运用,以及自己遇到的问题。谢谢大家!