对于非结构化数据,HDFS的访问控制有多精细?

时间:2021-08-29 13:49:39

I am looking for any piece of technical paper explaining how access control is conducted on unstructured data ingested by HDFS.

我正在寻找任何一篇技术论文,解释如何对HDFS摄取的非结构化数据进行访问控制。

  1. Can the granularity level be smaller than POSIX-ish file permissions?

    粒度级别是否可以小于POSIX-ish文件权限?

  2. Similarly, how would products like RecordService (from Cloudera), which provide an abstraction layer for security on storage components, work on unstructured data?

    同样,RecordService(来自Cloudera)的产品如何处理非结构化数据,它们为存储组件的安全性提供了一个抽象层?


For instance, if I have a very big emails archive file (more than a terabyte), would I be able to specify a more fine-grained ACL than one on the entire file itself? I am thinking about email headers, etc.

例如,如果我有一个非常大的电子邮件存档文件(超过1TB),我是否能够指定比整个文件本身更精细的ACL?我正在考虑电子邮件标题等。

1 个解决方案

#1


0  

  1. The granularity supported is to the row and column levels. See details.
  2. 支持的粒度是行级和列级。查看具体信息。
  3. Presently, for RecordService to work, your data must be organized as Hive Metastore tables. In the future, RecordService may infer structure/schema from the files themselves (but, not the case today).
  4. 目前,要使RecordService正常工作,您的数据必须组织为Hive Metastore表。将来,RecordService可能会从文件本身推断出结构/架构(但今天不是这样)。

#1


0  

  1. The granularity supported is to the row and column levels. See details.
  2. 支持的粒度是行级和列级。查看具体信息。
  3. Presently, for RecordService to work, your data must be organized as Hive Metastore tables. In the future, RecordService may infer structure/schema from the files themselves (but, not the case today).
  4. 目前,要使RecordService正常工作,您的数据必须组织为Hive Metastore表。将来,RecordService可能会从文件本身推断出结构/架构(但今天不是这样)。