资讯专栏INFORMATION COLUMN

快速搭建 Hadoop 环境

wangtdgoodluck / 3082人阅读

摘要:对于来说,最主要的是两个方面,一个是分布式文件系统,另一个是计算模型,下面讲解下我在搭建环境过程。测试环境共台测试机,台台版本角色地址一部署前的准备工作需要知道依赖和以上,必须安装。必须安装并且保证一直运行,以便用脚本管理远端守护进程。

对于Hadoop来说,最主要的是两个方面,一个是分布式文件系统HDFS,另一个是MapReduce计算模型,下面讲解下我在搭建Hadoop 环境过程。

Hadoop 测试环境

  1. 共4台测试机,1台namenode 3台datanode
  2. OS版本:RHEL 5.5 X86_64
  3. Hadoop:0.20.203.0
  4. Jdk:jdk1.7.0
  5. 角色 ip地址
  6. namenode 192.168.57.75
  7. datanode1 192.168.57.76
  8. datanode2 192.168.57.78
  9. datanode3 192.168.57.79


一 部署 Hadoop 前的准备工作

  1. 1 需要知道hadoop依赖Java和SSH
  2. Java 1.5.x (以上),必须安装。
  3. ssh 必须安装并且保证 sshd 一直运行,以便用Hadoop 脚本管理远端Hadoop守护进程。
  4. 2 建立 Hadoop 公共帐号
  5. 所有的节点应该具有相同的用户名,可以使用如下命令添加:
  6. useradd hadoop
  7. passwd hadoop
  8. 3 配置 host 主机名
  9. tail -n 3 /etc/hosts
  10. 192.168.57.75 namenode
  11. 192.168.57.76 datanode1
  12. 192.168.57.78 datanode2
  13. 192.168.57.79 datanode3
  14. 4 以上几点要求所有节点(namenode|datanode)配置全部相同


二 ssh 配置
ssh 详细了解

  1. 1 生成私匙 id_rsa 与 公匙 id_rsa.pub 配置文件
  2. [hadoop@hadoop1 ~]$ ssh-keygen -t rsa
  3. Generating public/private rsa key pair.
  4. Enter file in which to save the key (/home/hadoop/.ssh/id_rsa):
  5. Enter passphrase (empty for no passphrase):
  6. Enter same passphrase again:
  7. Your identification has been saved in /home/hadoop/.ssh/id_rsa.
  8. Your public key has been saved in /home/hadoop/.ssh/id_rsa.pub.
  9. The key fingerprint is:
  10. d6:63:76:43:e2:5b:8e:85:ab:67:a2:7c:a6:8f:23:f9 hadoop@hadoop1.test.com
  11. 2 私匙 id_rsa 与 公匙 id_rsa.pub 配置文件
  12. [hadoop@hadoop1 ~]$ ls .ssh/
  13. authorized_keys id_rsa id_rsa.pub known_hosts
  14. 3 把公匙文件上传到datanode服务器
  15. [hadoop@hadoop1 ~]$ ssh-copy-id -i ~/.ssh/id_rsa.pub hadoop@datanode1
  16. 28
  17. hadoop@datanode1"s password:
  18. Now try logging into the machine, with "ssh "hadoop@datanode1"", and check in:
  19. .ssh/authorized_keys
  20. to make sure we haven"t added extra keys that you weren"t expecting.
  21. [hadoop@hadoop1 ~]$ ssh-copy-id -i ~/.ssh/id_rsa.pub hadoop@datanode2
  22. 28
  23. hadoop@datanode2"s password:
  24. Now try logging into the machine, with "ssh "hadoop@datanode2"", and check in:
  25. .ssh/authorized_keys
  26. to make sure we haven"t added extra keys that you weren"t expecting.
  27. [hadoop@hadoop1 ~]$ ssh-copy-id -i ~/.ssh/id_rsa.pub hadoop@datanode3
  28. 28
  29. hadoop@datanode3"s password:
  30. Now try logging into the machine, with "ssh "hadoop@datanode3"", and check in:
  31. .ssh/authorized_keys
  32. to make sure we haven"t added extra keys that you weren"t expecting.
  33. [hadoop@hadoop1 ~]$ ssh-copy-id -i ~/.ssh/id_rsa.pub hadoop@localhost
  34. 28
  35. hadoop@localhost"s password:
  36. Now try logging into the machine, with "ssh "hadoop@localhost"", and check in:
  37. .ssh/authorized_keys
  38. to make sure we haven"t added extra keys that you weren"t expecting.
  39. 4 验证
  40. [hadoop@hadoop1 ~]$ ssh datanode1
  41. Last login: Thu Feb 2 09:01:16 2012 from 192.168.57.71
  42. [hadoop@hadoop2 ~]$ exit
  43. logout
  44. [hadoop@hadoop1 ~]$ ssh datanode2
  45. Last login: Thu Feb 2 09:01:18 2012 from 192.168.57.71
  46. [hadoop@hadoop3 ~]$ exit
  47. logout
  48. [hadoop@hadoop1 ~]$ ssh datanode3
  49. Last login: Thu Feb 2 09:01:20 2012 from 192.168.57.71
  50. [hadoop@hadoop4 ~]$ exit
  51. logout
  52. [hadoop@hadoop1 ~]$ ssh localhost
  53. Last login: Thu Feb 2 09:01:24 2012 from 192.168.57.71
  54. [hadoop@hadoop1 ~]$ exit
  55. logout

三 java环境配置

  1. 1 下载合适的jdk
  2. //此文件为64Linux 系统使用的 RPM包
  3. wget http://download.oracle.com/otn-pub/java/jdk/7/jdk-7-linux-x64.rpm
  4. 2 安装jdk
  5. rpm -ivh jdk-7-linux-x64.rpm
  6. 3 验证java
  7. [root@hadoop1 ~]# java -version
  8. java version "1.7.0"
  9. Java(TM) SE Runtime Environment (build 1.7.0-b147)
  10. Java HotSpot(TM) 64-Bit Server VM (build 21.0-b17, mixed mode)
  11. [root@hadoop1 ~]# ls /usr/java/
  12. default jdk1.7.0 latest
  13. 4 配置java环境变量
  14. #vim /etc/profile //在profile文件中加入如下信息:
  15. #add for hadoop
  16. export JAVA_HOME=/usr/java/jdk1.7.0
  17. export CLASSPATH=.:$JAVA_HOME/jre/lib/rt.jar:$JAVA_HOME/lib/dt.jar:$JAVA_HOME/
  18. export PATH=$PATH:$JAVA_HOME/bin
  19. //使环境变量生效
  20. source /etc/profile
  21. 5 拷贝 /etc/profile 到 datanode
  22. [root@hadoop1 src]# scp /etc/profile root@datanode1:/etc/
  23. The authenticity of host "datanode1 (192.168.57.86)" can"t be established.
  24. RSA key fingerprint is b5:00:d1:df:73:4c:94:f1:ea:1f:b5:cd:ed:3a:cc:e1.
  25. Are you sure you want to continue connecting (yes/no)? yes
  26. Warning: Permanently added "datanode1,192.168.57.86" (RSA) to the list of known hosts.
  27. root@datanode1"s password:
  28. profile 100% 1624 1.6KB/s 00:00
  29. [root@hadoop1 src]# scp /etc/profile root@datanode2:/etc/
  30. The authenticity of host "datanode2 (192.168.57.87)" can"t be established.
  31. RSA key fingerprint is 57:cf:96:15:78:a3:94:93:30:16:8e:66:47:cd:f9:cd.
  32. Are you sure you want to continue connecting (yes/no)? yes
  33. Warning: Permanently added "datanode2,192.168.57.87" (RSA) to the list of known hosts.
  34. root@datanode2"s password:
  35. profile 100% 1624 1.6KB/s 00:00
  36. [root@hadoop1 src]# scp /etc/profile root@datanode3:/etc/
  37. The authenticity of host "datanode3 (192.168.57.88)" can"t be established.
  38. RSA key fingerprint is 31:73:e8:3c:20:0c:1e:b2:59:5c:d1:01:4b:26:41:70.
  39. Are you sure you want to continue connecting (yes/no)? yes
  40. Warning: Permanently added "datanode3,192.168.57.88" (RSA) to the list of known hosts.
  41. root@datanode3"s password:
  42. profile 100% 1624 1.6KB/s 00:00
  43. 6 拷贝 jdk 安装包,并在每个datanode 节点安装 jdk 包
  44. [root@hadoop1 ~]# scp -r /home/hadoop/src/ hadoop@datanode1:/home/hadoop/
  45. hadoop@datanode1"s password:
  46. hadoop-0.20.203.0rc1.tar.gz 100% 58MB 57.8MB/s 00:01
  47. jdk-7-linux-x64.rpm 100% 78MB 77.9MB/s 00:01
  48. [root@hadoop1 ~]# scp -r /home/hadoop/src/ hadoop@datanode2:/home/hadoop/
  49. hadoop@datanode2"s password:
  50. hadoop-0.20.203.0rc1.tar.gz 100% 58MB 57.8MB/s 00:01
  51. jdk-7-linux-x64.rpm 100% 78MB 77.9MB/s 00:01
  52. [root@hadoop1 ~]# scp -r /home/hadoop/src/ hadoop@datanode3:/home/hadoop/
  53. hadoop@datanode3"s password:
  54. hadoop-0.20.203.0rc1.tar.gz 100% 58MB 57.8MB/s 00:01
  55. jdk-7-linux-x64.rpm 100% 78MB 77.9MB/s 00:01

四 hadoop 配置
//注意使用hadoop 用户 操作

  1. 1 配置目录
  2. [hadoop@hadoop1 ~]$ pwd
  3. /home/hadoop
  4. [hadoop@hadoop1 ~]$ ll
  5. total 59220
  6. lrwxrwxrwx 1 hadoop hadoop 17 Feb 1 16:59 hadoop -> hadoop-0.20.203.0
  7. drwxr-xr-x 12 hadoop hadoop 4096 Feb 1 17:31 hadoop-0.20.203.0
  8. -rw-r--r-- 1 hadoop hadoop 60569605 Feb 1 14:24 hadoop-0.20.203.0rc1.tar.gz
  9. 2 配置hadoop-env.sh,指定java位置
  10. vim hadoop/conf/hadoop-env.sh
  11. export JAVA_HOME=/usr/java/jdk1.7.0
  12. 3 配置core-site.xml //定位文件系统的 namenode
  13. [hadoop@hadoop1 ~]$ cat hadoop/conf/core-site.xml
  14. fs.default.namename>
  15. hdfs://namenode:9000value>
  16. property>
  17. configuration>
  18. 4 配置mapred-site.xml //定位jobtracker 所在的主节点
  19. [hadoop@hadoop1 ~]$ cat hadoop/conf/mapred-site.xml
  20. mapred.job.trackername>
  21. namenode:9001value>
  22. property>
  23. configuration>
  24. 5 配置hdfs-site.xml //配置HDFS副本数量
  25. [hadoop@hadoop1 ~]$ cat hadoop/conf/hdfs-site.xml
  26. dfs.replicationname>
  27. 3value>
  28. property>
  29. configuration>
  30. 6 配置 master 与 slave 配置文档
  31. [hadoop@hadoop1 ~]$ cat hadoop/conf/masters
  32. namenode
  33. [hadoop@hadoop1 ~]$ cat hadoop/conf/slaves
  34. datanode1
  35. datanode2
  36. 7 拷贝hadoop 目录到所有节点(datanode)
  37. [hadoop@hadoop1 ~]$ scp -r hadoop hadoop@datanode1:/home/hadoop/
  38. [hadoop@hadoop1 ~]$ scp -r hadoop hadoop@datanode2:/home/hadoop/
  39. [hadoop@hadoop1 ~]$ scp -r hadoop hadoop@datanode3:/home/hadoop
  40. 8 格式化 HDFS
  41. [hadoop@hadoop1 hadoop]$ bin/hadoop namenode -format
  42. 12/02/02 11:31:15 INFO namenode.NameNode: STARTUP_MSG:
  43. /************************************************************
  44. STARTUP_MSG: Starting NameNode
  45. STARTUP_MSG: host = hadoop1.test.com/127.0.0.1
  46. STARTUP_MSG: args = [-format]
  47. STARTUP_MSG: version = 0.20.203.0
  48. STARTUP_MSG: build = http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-security-203 -r 1099333; compiled by "oom" on Wed May 4 07:57:50 PDT 2011
  49. ************************************************************/
  50. Re-format filesystem in /tmp/hadoop-hadoop/dfs/name ? (Y or N) Y //这里输入Y
  51. 12/02/02 11:31:17 INFO util.GSet: VM type = 64-bit
  52. 12/02/02 11:31:17 INFO util.GSet: 2% max memory = 19.33375 MB
  53. 12/02/02 11:31:17 INFO util.GSet: capacity = 2^21 = 2097152 entries
  54. 12/02/02 11:31:17 INFO util.GSet: recommended=2097152, actual=2097152
  55. 12/02/02 11:31:17 INFO namenode.FSNamesystem: fsOwner=hadoop
  56. 12/02/02 11:31:18 INFO namenode.FSNamesystem: supergroupsupergroup=supergroup
  57. 12/02/02 11:31:18 INFO namenode.FSNamesystem: isPermissionEnabled=true
  58. 12/02/02 11:31:18 INFO namenode.FSNamesystem: dfs.block.invalidate.limit=100
  59. 12/02/02 11:31:18 INFO namenode.FSNamesystem: isAccessTokenEnabled=false accessKeyUpdateInterval=0 min(s), accessTokenLifetime=0 min(s)
  60. 12/02/02 11:31:18 INFO namenode.NameNode: Caching file names occuring more than 10 times
  61. 12/02/02 11:31:18 INFO common.Storage: Image file of size 112 saved in 0 seconds.
  62. 12/02/02 11:31:18 INFO common.Storage: Storage directory /tmp/hadoop-hadoop/dfs/name has been successfully formatted.
  63. 12/02/02 11:31:18 INFO namenode.NameNode: SHUTDOWN_MSG:
  64. /************************************************************
  65. SHUTDOWN_MSG: Shutting down NameNode at hadoop1.test.com/127.0.0.1
  66. ************************************************************/
  67. [hadoop@hadoop1 hadoop]$
  68. 9 启动hadoop 守护进程
  69. [hadoop@hadoop1 hadoop]$ bin/start-all.sh
  70. starting namenode, logging to /home/hadoop/hadoop/bin/../logs/hadoop-hadoop-namenode-hadoop1.test.com.out
  71. datanode1: starting datanode, logging to /home/hadoop/hadoop/bin/../logs/hadoop-hadoop-datanode-hadoop2.test.com.out
  72. datanode2: starting datanode, logging to /home/hadoop/hadoop/bin/../logs/hadoop-hadoop-datanode-hadoop3.test.com.out
  73. datanode3: starting datanode, logging to /home/hadoop/hadoop/bin/../logs/hadoop-hadoop-datanode-hadoop4.test.com.out
  74. starting jobtracker, logging to /home/hadoop/hadoop/bin/../logs/hadoop-hadoop-jobtracker-hadoop1.test.com.out
  75. datanode1: starting tasktracker, logging to /home/hadoop/hadoop/bin/../logs/hadoop-hadoop-tasktracker-hadoop2.test.com.out
  76. datanode2: starting tasktracker, logging to /home/hadoop/hadoop/bin/../logs/hadoop-hadoop-tasktracker-hadoop3.test.com.out
  77. datanode3: starting tasktracker, logging to /home/hadoop/hadoop/bin/../logs/hadoop-hadoop-tasktracker-hadoop4.test.com.out
  78. 10 验证
  79. //namenode
  80. [hadoop@hadoop1 logs]$ jps
  81. 2883 JobTracker
  82. 3002 Jps
  83. 2769 NameNode
  84. //datanode
  85. [hadoop@hadoop2 ~]$ jps
  86. 2743 TaskTracker
  87. 2670 DataNode
  88. 2857 Jps
  89. [hadoop@hadoop3 ~]$ jps
  90. 2742 TaskTracker
  91. 2856 Jps
  92. 2669 DataNode
  93. [hadoop@hadoop4 ~]$ jps
  94. 2742 TaskTracker
  95. 2852 Jps
  96. 2659 DataNode
  97. Hadoop 监控web页面
  98. http://192.168.57.75:50070/dfshealth.jsp



五 简单验证HDFS

  1. hadoop 的文件命令格式如下:
  2. hadoop fs -cmd
  3. //建立目录
  4. [hadoop@hadoop1 hadoop]$ bin/hadoop fs -mkdir /test-hadoop
  5. //査看目录
  6. [hadoop@hadoop1 hadoop]$ bin/hadoop fs -ls /
  7. Found 2 items
  8. drwxr-xr-x - hadoop supergroup 0 2012-02-02 13:32 /test-hadoop
  9. drwxr-xr-x - hadoop supergroup 0 2012-02-02 11:32 /tmp
  10. //査看目录包括子目录
  11. [hadoop@hadoop1 hadoop]$ bin/hadoop fs -lsr /
  12. drwxr-xr-x - hadoop supergroup 0 2012-02-02 13:32 /test-hadoop
  13. drwxr-xr-x - hadoop supergroup 0 2012-02-02 11:32 /tmp
  14. drwxr-xr-x - hadoop supergroup 0 2012-02-02 11:32 /tmp/hadoop-hadoop
  15. drwxr-xr-x - hadoop supergroup 0 2012-02-02 11:32 /tmp/hadoop-hadoop/mapred
  16. drwx------ - hadoop supergroup 0 2012-02-02 11:32 /tmp/hadoop-hadoop/mapred/system
  17. -rw------- 2 hadoop supergroup 4 2012-02-02 11:32 /tmp/hadoop-hadoop/mapred/system/jobtracker.info
  18. //添加文件
  19. [hadoop@hadoop1 hadoop]$ bin/hadoop fs -put /home/hadoop/hadoop-0.20.203.0rc1.tar.gz /test-hadoop
  20. [hadoop@hadoop1 hadoop]$ bin/hadoop fs -lsr /
  21. drwxr-xr-x - hadoop supergroup 0 2012-02-02 13:34 /test-hadoop
  22. -rw-r--r-- 2 hadoop supergroup 60569605 2012-02-02 13:34 /test-hadoop/hadoop-0.20.203.0rc1.tar.gz
  23. drwxr-xr-x - hadoop supergroup 0 2012-02-02 11:32 /tmp
  24. drwxr-xr-x - hadoop supergroup 0 2012-02-02 11:32 /tmp/hadoop-hadoop
  25. drwxr-xr-x - hadoop supergroup 0 2012-02-02 11:32 /tmp/hadoop-hadoop/mapred
  26. drwx------ - hadoop supergroup 0 2012-02-02 11:32 /tmp/hadoop-hadoop/mapred/system
  27. -rw------- 2 hadoop supergroup 4 2012-02-02 11:32 /tmp/hadoop-hadoop/mapred/system/jobtracker.info
  28. //获取文件
  29. [hadoop@hadoop1 hadoop]$ bin/hadoop fs -get /test-hadoop/hadoop-0.20.203.0rc1.tar.gz /tmp/
  30. [hadoop@hadoop1 hadoop]$ ls /tmp/*.tar.gz
  31. /tmp/1.tar.gz /tmp/hadoop-0.20.203.0rc1.tar.gz
  32. //删除文件
  33. [hadoop@hadoop1 hadoop]$ bin/hadoop fs -rm /test-hadoop/hadoop-0.20.203.0rc1.tar.gz
  34. Deleted hdfs://namenode:9000/test-hadoop/hadoop-0.20.203.0rc1.tar.gz
  35. [hadoop@hadoop1 hadoop]$ bin/hadoop fs -lsr /
  36. drwxr-xr-x - hadoop supergroup 0 2012-02-02 13:57 /test-hadoop
  37. drwxr-xr-x - hadoop supergroup 0 2012-02-02 11:32 /tmp
  38. drwxr-xr-x - hadoop supergroup 0 2012-02-02 11:32 /tmp/hadoop-hadoop
  39. drwxr-xr-x - hadoop supergroup 0 2012-02-02 11:32 /tmp/hadoop-hadoop/mapred
  40. drwx------ - hadoop supergroup 0 2012-02-02 11:32 /tmp/hadoop-hadoop/mapred/system
  41. -rw------- 2 hadoop supergroup 4 2012-02-02 11:32 /tmp/hadoop-hadoop/mapred/system/jobtracker.info
  42. drwxr-xr-x - hadoop supergroup 0 2012-02-02 13:36 /user
  43. -rw-r--r-- 2 hadoop supergroup 321 2012-02-02 13:36 /user/hadoop
  44. //删除目录
  45. [hadoop@hadoop1 hadoop]$ bin/hadoop fs -rmr /test-hadoop
  46. Deleted hdfs://namenode:9000/test-hadoop
  47. [hadoop@hadoop1 hadoop]$ bin/hadoop fs -lsr /
  48. drwxr-xr-x - hadoop supergroup 0 2012-02-02 11:32 /tmp
  49. drwxr-xr-x - hadoop supergroup 0 2012-02-02 11:32 /tmp/hadoop-hadoop
  50. drwxr-xr-x - hadoop supergroup 0 2012-02-02 11:32 /tmp/hadoop-hadoop/mapred
  51. drwx------ - hadoop supergroup 0 2012-02-02 11:32 /tmp/hadoop-hadoop/mapred/system
  52. -rw------- 2 hadoop supergroup 4 2012-02-02 11:32 /tmp/hadoop-hadoop/mapred/system/jobtracker.info
  53. drwxr-xr-x - hadoop supergroup 0 2012-02-02 13:36 /user
  54. -rw-r--r-- 2 hadoop supergroup 321 2012-02-02 13:36 /user/hadoop
  55. //hadoop fs 帮助(部分)
  56. [hadoop@hadoop1 hadoop]$ bin/hadoop fs -help
  57. hadoop fs is the command to execute fs commands. The full syntax is:
  58. hadoop fs [-fs ] [-conf ]
  59. [-D ] [-ls ] [-lsr ] [-du ]
  60. [-dus ] [-mv ] [-cp ] [-rm [-skipTrash] ]
  61. [-rmr [-skipTrash] ] [-put ... ] [-copyFromLocal ... ]
  62. [-moveFromLocal ... ] [-get [-ignoreCrc] [-crc]
  63. [-getmerge [addnl]] [-cat ]
  64. [-copyToLocal [-ignoreCrc] [-crc] ] [-moveToLocal ]
  65. [-mkdir ] [-report] [-setrep [-R] [-w] ]
  66. [-touchz ] [-test -[ezd] ] [-stat [format] ]
  67. [-tail [-f] ] [-text ]
  68. [-chmod [-R] PATH...]
  69. [-chown [-R] [OWNER][:[GROUP]] PATH...]
  70. [-chgrp [-R] GROUP PATH...]
  71. [-count[-q] ]
  72. [-help [cmd]]


更多Hadoop 相关知识

结束
Hadoop 环境搭建步骤繁琐,需要具备一定的Linux 系统知识,需要注意的是,通过以上步骤搭建的Hadoop 环境只能让你大体了解的hadoop ,如果想将HDFS 用于线上服务,还需对hadoop 配置文档做进一步配置 ,后续文档将继续以博文的形式发布,敬请期待。

 

本文出自 “dongnan” 博客,请务必保留此出处http://dngood.blog.51cto.com/446195/775368

文章版权归作者所有,未经允许请勿转载,若此文章存在违规行为,您可以联系管理员删除。

转载请注明本文地址:https://www.ucloud.cn/yun/3699.html

相关文章

  • Apache Hive-2.3.0 快速搭建与使用

    摘要:简介是一个基于的开源数据仓库工具,用于存储和处理海量结构化数据。它把海量数据存储于文件系统,而不是数据库,但提供了一套类数据库的数据存储和处理机制,并采用类语言对这些数据进行自动化管理和处理。 Hive 简介 Hive 是一个基于 hadoop 的开源数据仓库工具,用于存储和处理海量结构化数据。它把海量数据存储于 hadoop 文件系统,而不是数据库,但提供了一套类数据库的数据存储和处...

    ralap 评论0 收藏0
  • 跟上大数据的步伐:快速搭建Spark集群

    摘要:本文详细介绍了基于进行集群资源调度的数人云,如何部署集群。数人云集群正是通过进行集群资源调度,因此,数人云部署集群,有着天然的优势。 Spark 是 UC Berkeley AMP lab 开源的类 Hadoop MapReduce 的通用的并行计算框架,Spark 基于 map reduce 算法实现的分布式计算,拥有 Hadoop MapReduce 所具有的优点,并且 能更好地适...

    elina 评论0 收藏0
  • 盘点:世界前十 超人气基于Hadoop的大数据初创企业

    摘要:据估计,目前的生态系统市场价值大约为,,该研究公司预计,这一数字到年将迅速增加到亿美元。超人气初创公司前两名他们做什么主要为企业提供相关的大数据解决方案,并且将的袁术数据进行转换,为企业提供更智能的商业指导。 这已经不再是什么秘密了,全球的数据正在以几何数字增长,借助这股数据浪潮在全球范围内迅速成长起来一大批Hadoop的初创型公司。作为Apache的一个开源分支Hadoop几乎已经成为了大...

    marek 评论0 收藏0
  • 大数据集群环境搭建——服务器篇

    摘要:接下来,本小白带大家一步步搭建一个大数据的分布式环境。本文为大数据集群环境搭建的第一篇,服务器搭建篇。配置免登录大数据集群中各个服务器间通信都需要使用免登录,具体操作请大家自行搜索,网上的资源非常丰富。 如今我们已经进入大数据时代,在校大学生及各个公司很多都开始学习大数据。那么要学习大数据,第一步就是要搭建分布式集群环境。虽说dreamtecher也是个刚学习大数据没多久的小白,但是也...

    linkin 评论0 收藏0
  • 大数据集群环境搭建——服务器篇

    摘要:接下来,本小白带大家一步步搭建一个大数据的分布式环境。本文为大数据集群环境搭建的第一篇,服务器搭建篇。配置免登录大数据集群中各个服务器间通信都需要使用免登录,具体操作请大家自行搜索,网上的资源非常丰富。 如今我们已经进入大数据时代,在校大学生及各个公司很多都开始学习大数据。那么要学习大数据,第一步就是要搭建分布式集群环境。虽说dreamtecher也是个刚学习大数据没多久的小白,但是也...

    blastz 评论0 收藏0

发表评论

0条评论

最新活动
阅读需要支付1元查看
<