一、windows客户端准备
双击安装hadoop依赖
放在非中文路径下,双击winutils.exe进行安装

如果安装错误,先安装微软运行库

新建Java Maven 工程
maven坐标
- <dependencies>
- <dependency>
- <groupId>org.apache.hadoop</groupId>
- <artifactId>hadoop-client</artifactId>
- <version>3.1.3</version>
- </dependency>
- <dependency>
- <groupId>junit</groupId>
- <artifactId>junit</artifactId>
- <version>4.12</version>
- </dependency>
- <dependency>
- <groupId>org.slf4j</groupId>
- <artifactId>slf4j-log4j12</artifactId>
- <version>1.7.30</version>
- </dependency>
- </dependencies>
复制代码
在sre/main/resources下新建log4j.properties配置文件
- log4j.rootLogger=INFO, stdout
- log4j.appender.stdout=org.apache.log4j.ConsoleAppender
- log4j.appender.stdout.layout=org.apache.log4j.PatternLayout
- log4j.appender.stdout.layout.ConversionPattern=%d %p [%c] - %m%n
- log4j.appender.logfile=org.apache.log4j.FileAppender
- log4j.appender.logfile.File=target/spring.log
- log4j.appender.logfile.layout=org.apache.log4j.PatternLayout
- log4j.appender.logfile.layout.ConversionPattern=%d %p [%c] - %m%n
复制代码
新建类 HdfsClient
二、API操作
创建文件夹
- package com.demo.hdfs;
- import org.apache.hadoop.conf.Configuration;
- import org.apache.hadoop.fs.FileSystem;
- import org.apache.hadoop.fs.Path;
- import org.junit.After;
- import org.junit.Before;
- import org.junit.Test;
- import java.io.IOException;
- import java.net.URI;
- import java.net.URISyntaxException;
- /**
- * 客户端代码操作
- * 1. 获取一个客户端对象
- * 2. 执行相关操作
- * 3. 关闭资源
- */
- public class HdfsClient {
- private FileSystem fs;
- @Before
- public void init() throws URISyntaxException, InterruptedException, IOException {
- //连接集群nn地址
- URI uri = new URI("hdfs://hadoop102:8020");
- //创建一个配置文件
- Configuration configuration = new Configuration();
- //用户
- String user = "root";
- //获取客户端对象
- fs = FileSystem.get(uri, configuration, user);
- }
- @After
- public void close() throws IOException {
- //关闭资源
- fs.close();
- }
- @Test
- public void testMkdir() throws IOException {
- //创建一个文件夹
- fs.mkdirs(new Path("/apipath1"));
- }
- }
复制代码
上传
- //上传
- @Test
- public void testPut() throws IOException {
- //参数1 :是否删除元数据
- //参数2: 是否允许覆盖
- //参数3: 目的路径
- fs.copyFromLocalFile(false, true, new Path("E:\\study\\BigData\\Hadoop\\脚本\\myhadoop.sh"), new Path("hdfs://hadoop102/code"));
- }
复制代码
下载
- //下载
- @Test
- public void testGet() throws IOException {
- //参数1 :是否删除原数据
- //参数2: 目的路径
- //参数3: 原文件路径
- //参数4:是否进行本地校验 ,false 开启,ture 关闭
- fs.copyToLocalFile(false, new Path("hdfs://hadoop102/code/myhadoop.sh"), new Path("E:\"), true);
- }
复制代码
.xxx.crc文件是一种校验方法
删除
- //删除
- @Test
- public void testRm() throws IOException {
- //参数1 :目的路径
- //参数2: 是否递归删除
-
- //1. 删除文件
- //fs.delete(new Path("hdfs://hadoop102/code/myhadoop.sh"), true);
- //2. 删除空目录
- //fs.delete( new Path("hdfs://hadoop102/code/code1"), true);
- //3. 删除非空目录,如果不是递归删除,则报错
- fs.delete(new Path("hdfs://hadoop102/code/code2"), true);
- }
复制代码
文件更名和移动
- //文件更名和移动
- @Test
- public void testMv() throws IOException {
- //参数1 :原文件路径
- //参数2: 目的文件路径
- //1.文件名称更改
- //fs.rename(new Path("hdfs://hadoop102/code/myhadoop.sh"), new Path("hdfs://hadoop102/code/myhadoop_rename.sh"));
- //2.文件移动并修改
- //fs.rename(new Path("hdfs://hadoop102/code/myhadoop_rename.sh"), new Path("hdfs://hadoop102/myhadoop_remove_and_rename.sh"));
- //3. 目录更名
- fs.rename(new Path("hdfs://hadoop102/code"), new Path("hdfs://hadoop102/code_rename"));
- }
复制代码
文件详细信息查看
查看文件名称、权限、长度、块信息
- //文件详细信息查看
- @Test
- public void testFileDetail() throws IOException {
- //参数1 :路径
- //参数2: 目的文件路径
- //获取所有文件信息
- RemoteIterator<LocatedFileStatus> listFiles = fs.listFiles(new Path("hdfs://hadoop102/"), true);
- //遍历文件信息
- while (listFiles.hasNext()) {
- LocatedFileStatus fileStatus = listFiles.next();
- System.out.println("-----------------------------------------");
- System.out.println(fileStatus.toString());
- System.out.println("-----------------------------------------");
- }
- }
复制代码
toString方法
- public String toString() {
- StringBuilder sb = new StringBuilder();
- sb.append(this.getClass().getSimpleName());
- sb.append("{");
- sb.append("path=" + this.path);
- sb.append("; isDirectory=" + this.isdir);
- if (!this.isDirectory()) {
- sb.append("; length=" + this.length);
- sb.append("; replication=" + this.block_replication);
- sb.append("; blocksize=" + this.blocksize);
- }
- sb.append("; modification_time=" + this.modification_time);
- sb.append("; access_time=" + this.access_time);
- sb.append("; owner=" + this.owner);
- sb.append("; group=" + this.group);
- sb.append("; permission=" + this.permission);
- sb.append("; isSymlink=" + this.isSymlink());
- if (this.isSymlink()) {
- try {
- sb.append("; symlink=" + this.getSymlink());
- } catch (IOException var3) {
- throw new RuntimeException("Unexpected exception", var3);
- }
- }
- sb.append("; hasAcl=" + this.hasAcl());
- sb.append("; isEncrypted=" + this.isEncrypted());
- sb.append("; isErasureCoded=" + this.isErasureCoded());
- sb.append("}");
- return sb.toString();
- }
复制代码
文件和文件夹判断
- //判断是文件夹还是文件
- @Test
- public void testFile() throws IOException {
- //获取所有文件信息
- FileStatus[] fileStatuses = fs.listStatus(new Path("hdfs://hadoop102/"));
- //遍历文件信息
- for (FileStatus fileStatus : fileStatuses) {
- if (fileStatus.isFile()) {
- System.out.println("this is file--" + fileStatus.toString());
- System.out.println("-----------------------------------------");
- } else {
- System.out.println("this is path--" + fileStatus.toString());
- System.out.println("-----------------------------------------");
- }
- }
- }
复制代码
来源:https://blog.caogenba.net/qq_42000631/article/details/122483873
免责声明:如果侵犯了您的权益,请联系站长,我们会及时删除侵权内容,谢谢合作! |