I am looking for any piece of technical paper explaining how access control is conducted on unstructured data ingested by HDFS.
我正在寻找任何一篇技术论文,解释如何对HDFS摄取的非结构化数据进行访问控制。
-
Can the granularity level be smaller than POSIX-ish file permissions?
粒度级别是否可以小于POSIX-ish文件权限?
-
Similarly, how would products like RecordService (from Cloudera), which provide an abstraction layer for security on storage components, work on unstructured data?
同样,RecordService(来自Cloudera)的产品如何处理非结构化数据,它们为存储组件的安全性提供了一个抽象层?
For instance, if I have a very big emails archive file (more than a terabyte), would I be able to specify a more fine-grained ACL than one on the entire file itself? I am thinking about email headers, etc.
例如,如果我有一个非常大的电子邮件存档文件(超过1TB),我是否能够指定比整个文件本身更精细的ACL?我正在考虑电子邮件标题等。
1 个解决方案
#1
0
- The granularity supported is to the row and column levels. See details.
- 支持的粒度是行级和列级。查看具体信息。
- Presently, for RecordService to work, your data must be organized as Hive Metastore tables. In the future, RecordService may infer structure/schema from the files themselves (but, not the case today).
- 目前,要使RecordService正常工作,您的数据必须组织为Hive Metastore表。将来,RecordService可能会从文件本身推断出结构/架构(但今天不是这样)。
#1
0
- The granularity supported is to the row and column levels. See details.
- 支持的粒度是行级和列级。查看具体信息。
- Presently, for RecordService to work, your data must be organized as Hive Metastore tables. In the future, RecordService may infer structure/schema from the files themselves (but, not the case today).
- 目前,要使RecordService正常工作,您的数据必须组织为Hive Metastore表。将来,RecordService可能会从文件本身推断出结构/架构(但今天不是这样)。