大忽悠(大数据)/hadoop相关配置

花木瑞记录运维约 746 字...

恕我直言,用什么hadoop hbase对于现在绝大多数项目而言就是闲得蛋疼,尤其是连搭环境都要搭半天,最后还只在本地机器开虚拟机然后只开一个的。 这些东西是为了规模而生的,并且也不是唯一解法,做出来玩具项目也只能说是摸一摸而已,而现在这种...连摸都不算,可能是hello world吧。 好讨厌java...也许是因为学校太傻比。————明明已经是实践课程了,却仍以那狗屁技术要点为中心,那些东西用什么写出来根本就无所谓,而在不了解具体实践与场景的情况下,也不可能把原本的技术目的与解决的问题搞清楚..所以连摸都不算。

记一下配置,也方便复制粘贴...Ubuntu22上装hadoop,hbase 然而还没装好... 这玩意有什么一键配置的操作吗,就像docker那种...艹,我应该先去找这玩意的。

Ubuntu 修改时区

timedatectl set-timezone Asia/Shanghai

安装jdk

apt-get update
apt-get install openjdk-11-jdk

创建hadoop用户

adduser hadoop
密码同为
hadoop
su hadoop

赋权

ssh-keygen -t rsa  
cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys
chmod 640 ~/.ssh/authorized_keys  
ssh localhost

cd /home/hadoop

下载

下载hadoop

wget https://dlcdn.apache.org/hadoop/common/hadoop-3.3.4/hadoop-3.3.4.tar.gz
tar xzf hadoop-3.3.4.tar.gz
mv hadoop-3.3.4 hadoop

下载hbase

wget https://dlcdn.apache.org/hbase/2.5.2/hbase-2.5.2-bin.tar.gz
tar xzf hbase-2.5.2-bin.tar.gz
mv hbase-2.5.2 hbase

环境变量

vim  ~/.bashrc
source ~/.bashrc
export JAVA_HOME=/usr/lib/jvm/java-11-openjdk-amd64
export HADOOP_HOME=/home/hadoop/hadoop
export HADOOP_INSTALL=$HADOOP_HOME
export HADOOP_MAPRED_HOME=$HADOOP_HOME
export HADOOP_COMMON_HOME=$HADOOP_HOME
export HADOOP_HDFS_HOME=$HADOOP_HOME
export HADOOP_YARN_HOME=$HADOOP_HOME
export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native
export HADOOP_OPTS="-Djava.library.path=$HADOOP_HOME/lib/native"
export HBASE_HOME=/home/hadoop/hbase
export PATH=$PATH:$HADOOP_HOME/sbin:$HADOOP_HOME/bin:$HBASE_HOME/bin
vim $HADOOP_HOME/etc/hadoop/hadoop-env.sh
export JAVA_HOME=/usr/lib/jvm/java-11-openjdk-amd64
vim  /home/hadoop/hbase/conf/hbase-env.sh
export JAVA_HOME=/usr/lib/jvm/java-11-openjdk-amd64
export HBASE_MANAGES_ZK=true

配置

配置hadoop

vim $HADOOP_HOME/etc/hadoop/core-site.xml
<configuration>
    <property>
        <name>fs.defaultFS</name>
        <value>hdfs://146.190.44.79:9000</value>
    </property>
</configuration>
vim $HADOOP_HOME/etc/hadoop/hdfs-site.xml
<configuration>
    <property>
        <name>dfs.replication</name>
        <value>1</value>
    </property>
    <property>
        <name>dfs.name.dir</name>
        <value>file:///home/hadoop/hadoopdata/hdfs/namenode</value>
    </property>
    <property>
        <name>dfs.data.dir</name>
        <value>file:///home/hadoop/hadoopdata/hdfs/datanode</value>
    </property>
</configuration>
vim $HADOOP_HOME/etc/hadoop/mapred-site.xml
<configuration>
    <property>
        <name>mapreduce.framework.name</name>
        <value>yarn</value>
    </property>
</configuration>
vim $HADOOP_HOME/etc/hadoop/yarn-site.xml
<configuration>
    <property>
        <name>yarn.nodemanager.aux-services</name>
        <value>mapreduce_shuffle</value>
    </property>
</configuration>

配置hbase

vim  /home/hadoop/hbase/conf/hbase-site.xml
<configuration>
  <property>
    <name>hbase.rootdir</name>
    <value>hdfs://146.190.44.79:9000/hbase</value>
  </property>
  <property>
    <name>hbase.master</name>
    <value>146.190.44.79:16000</value>
  </property>
  <property>
    <name>hbase.zookeeper.property.dataDir</name>
    <value>/tmp/zookeeper</value>
  </property>
  <property>
    <name>hbase.cluster.distributed</name>
    <value>true</value>
  </property>
  <property>
    <name>hbase.zookeeper.quorum</name>
    <value>146.190.36.12</value>
  </property>
</configuration>

启动

hdfs namenode -format  
start-all.sh
start-hbase.sh
rm /home/hadoop/hbase/lib/client-facing-thirdparty/log4j-slf4j-impl-2.17.2.jar

端口

146.190.44.79:8088
146.190.44.79:9870
146.190.44.79:16010

hbase建表

hbase-daemon.shopen in new window start master ## 启动Hbase hbase shell

create 'email_user','user'
create 'user_id','id'
create 'gid_disk','gid' 
create 'user_file','file'
create 'file','file'
create 'follow','name'
create 'followed','userid'
create 'share','content'
create 'shareed','shareid'
评论
  • 按正序
  • 按倒序
  • 按热度
Powered by Waline v2.14.4