Package org.apache.spark.input
Class PortableDataStream
Object
org.apache.spark.input.PortableDataStream
- All Implemented Interfaces:
- Serializable
A class that allows DataStreams to be serialized and moved around by not creating them
 until they need to be read
- See Also:
- Note:
- TaskAttemptContext is not serializable resulting in the confBytes construct, CombineFileSplit is not serializable resulting in the splitBytes construct
- 
Constructor SummaryConstructorsConstructorDescriptionPortableDataStream(org.apache.hadoop.mapreduce.lib.input.CombineFileSplit isplit, org.apache.hadoop.mapreduce.TaskAttemptContext context, Integer index) 
- 
Method Summary
- 
Constructor Details- 
PortableDataStreampublic PortableDataStream(org.apache.hadoop.mapreduce.lib.input.CombineFileSplit isplit, org.apache.hadoop.mapreduce.TaskAttemptContext context, Integer index) 
 
- 
- 
Method Details- 
getConfigurationpublic org.apache.hadoop.conf.Configuration getConfiguration()
- 
getPath
- 
openCreate a new DataInputStream from the split and context. The user of this method is responsible for closing the stream after usage.- Returns:
- (undocumented)
 
- 
toArraypublic byte[] toArray()Read the file as a byte array- Returns:
- (undocumented)
 
 
-