HDFS连接JAVA,HDFS常用API
2021/9/18 22:35:14
本文主要是介绍HDFS连接JAVA,HDFS常用API,对大家解决编程问题具有一定的参考价值,需要的程序猿们随着小编来一起学习吧!
先在pom.xml中导入依赖包
<dependencies> <!-- https://mvnrepository.com/artifact/org.apache.hadoop/hadoop-hdfs --> <dependency> <groupId>org.apache.hadoop</groupId> <artifactId>hadoop-hdfs</artifactId> <version>2.7.6</version> </dependency> <!-- https://mvnrepository.com/artifact/org.apache.hadoop/hadoop-common --> <dependency> <groupId>org.apache.hadoop</groupId> <artifactId>hadoop-common</artifactId> <version>2.7.6</version> </dependency> <!-- https://mvnrepository.com/artifact/org.apache.hadoop/hadoop-client --> <dependency> <groupId>org.apache.hadoop</groupId> <artifactId>hadoop-client</artifactId> <version>2.7.6</version> </dependency> <!-- https://mvnrepository.com/artifact/junit/junit --> <dependency> <groupId>junit</groupId> <artifactId>junit</artifactId> <version>4.3</version> </dependency> </dependencies>
hdfs连接Java
1.先获取配置文件(hdfs-site.xml)
Configuration cg = new Configuration(); 导入的是import org.apache.hadoop.conf.Configuration; cg.set("dfs.replication","1"); 1是备份数量
2.获取连接地址(core-site.xml)
URI uri = new URI("hdfs://master:9000");
3.创建(获取)hdfs文件管理系统的对象,同过对象操作hdfs
FileSystem fs = FileSystem.get(uri, cg);
常用HDFSAPi:
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.*; import org.junit.Before; import org.junit.Test; import java.io.*; import java.net.URI; public class HdfsApi { FileSystem fs; @Before public void main() throws Exception { Configuration cg = new Configuration(); cg.set("dfs.replication", "1"); URI uri = new URI("hdfs://master:9000"); fs = FileSystem.get(uri, cg); } @Test public void mk() throws IOException { boolean mk = fs.mkdirs(new Path("/test")); } @Test public void del() throws IOException { //false表示不迭代删除也可以不加,true可以进行多目录迭代删除 boolean del = fs.delete(new Path("/test"),false); System.out.println(del); } @Test public void listStatus() throws IOException { //对比图在下 //查看目录下文件列表 FileStatus[] fileStatuses = fs.listStatus(new Path("/data/data")); System.out.println(fileStatuses); //[Lorg.apache.hadoop.fs.FileStatus;@106cc338 System.out.println("-------------------------------"); for (FileStatus fileStatus : fileStatuses) { System.out.println(fileStatus.getLen()); //文件大小 以B字节为单位 System.out.println(fileStatus.getReplication()); //副本个数 System.out.println(fileStatus.getPermission()); //读写状态 System.out.println(fileStatus.getBlockSize()); //固定的一个block大小128MB System.out.println(fileStatus.getAccessTime()); //创建文件时的时间戳 System.out.println(fileStatus.getPath()); //文件路径 System.out.println("-----------------"); } /*输出结果 log4j:WARN No appenders could be found for logger (org.apache.hadoop.util.Shell). log4j:WARN Please initialize the log4j system properly. log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info. [Lorg.apache.hadoop.fs.FileStatus;@106cc338 ------------------------------- 180 1 rw-r--r-- 134217728 1631969331482 hdfs://master:9000/data/data/cource.txt ----------------- 138540 1 rw-r--r-- 134217728 1631963067585 hdfs://master:9000/data/data/score.txt ----------------- 41998 1 rw-r--r-- 134217728 1631963067890 hdfs://master:9000/data/data/students.txt ----------------- */ } @Test public void listBlockLocation() throws IOException { //对比图在下 BlockLocation[] fbl = fs.getFileBlockLocations( new Path("/data/data/students.txt"),0,1000000000); for (BlockLocation bl : fbl) { String[] hosts = bl.getHosts(); for (String host : hosts) { System.out.println(host); } //node1 表示文件存在node1,因为文件小于一个block,所以这里只存在一个节点 System.out.println(bl.getLength()); //41998 size大小 String[] names = bl.getNames(); for (String name : names) { System.out.println(name); } //192.168.163.120:50010 node1的地址 System.out.println(bl.getOffset()); //0 偏移量 String[] topologyPaths = bl.getTopologyPaths(); for (String topologyPath : topologyPaths) { System.out.println(topologyPath); } // /default-rack/192.168.163.120:50010 } } @Test public void open() throws IOException { FSDataInputStream open = fs.open(new Path("/data/data/students.txt")); BufferedReader br = new BufferedReader(new InputStreamReader(open)); //因为文件中有中文,所以将字节流转为字符流来读取 String len; while ((len=br.readLine())!=null){ System.out.println(len); } br.close(); } @Test public void create() throws IOException { FSDataOutputStream fos = fs.create(new Path("/data/data/test.txt")); BufferedWriter bw = new BufferedWriter(new OutputStreamWriter(fos)); bw.write("你好"); bw.write("世界"); bw.newLine(); bw.write("我和我的祖国"); bw.flush(); bw.close(); } }
这篇关于HDFS连接JAVA,HDFS常用API的文章就介绍到这儿,希望我们推荐的文章对大家有所帮助,也希望大家多多支持为之网!
- 2024-11-28知识管理革命:文档软件的新玩法了解一下!
- 2024-11-28低代码应用课程:新手入门全攻略
- 2024-11-28哪些办公软件适合团队协作,且能够清晰记录每个阶段的工作进展?
- 2024-11-28全栈低代码开发课程:零基础入门到初级实战
- 2024-11-28拖动排序课程:轻松掌握课程拖动排序功能
- 2024-11-28如何高效管理数字化转型项目
- 2024-11-28SMART法则好用吗?有哪些项目管理工具辅助实现?
- 2024-11-28深度剖析:6 款办公软件如何构建设计团队项目可视化管理新生态?
- 2024-11-28HTTP缓存课程:新手入门指南
- 2024-11-28实战丨证券 HTAP 混合业务场景的难点问题应对