public abstract class AbstractHdfsBolt extends BaseRichBolt
Modifier and Type | Field and Description |
---|---|
protected OutputCollector |
collector |
protected String |
configKey |
protected FileNameFormat |
fileNameFormat |
protected Integer |
fileRetryCount |
protected org.apache.hadoop.fs.FileSystem |
fs |
protected String |
fsUrl |
protected org.apache.hadoop.conf.Configuration |
hdfsConfig |
protected Integer |
maxOpenFiles |
protected long |
offset |
protected Partitioner |
partitioner |
protected List<RotationAction> |
rotationActions |
protected Map<String,Integer> |
rotationCounterMap |
protected FileRotationPolicy |
rotationPolicy |
protected Timer |
rotationTimer |
protected SyncPolicy |
syncPolicy |
protected Integer |
tickTupleInterval |
protected Object |
writeLock |
protected Map<String,Writer> |
writers |
Constructor and Description |
---|
AbstractHdfsBolt() |
Modifier and Type | Method and Description |
---|---|
void |
cleanup()
Called when an IBolt is going to be shutdown.
|
void |
declareOutputFields(OutputFieldsDeclarer outputFieldsDeclarer)
Declare the output schema for all the streams of this topology.
|
protected abstract void |
doPrepare(Map conf,
TopologyContext topologyContext,
OutputCollector collector) |
void |
execute(Tuple tuple)
Process a single tuple of input.
|
protected org.apache.hadoop.fs.Path |
getBasePathForNextFile(Tuple tuple) |
Map<String,Object> |
getComponentConfiguration()
Declare configuration specific to this component.
|
protected abstract String |
getWriterKey(Tuple tuple) |
protected abstract Writer |
makeNewWriter(org.apache.hadoop.fs.Path path,
Tuple tuple) |
void |
prepare(Map conf,
TopologyContext topologyContext,
OutputCollector collector)
Marked as final to prevent override.
|
protected void |
rotateOutputFile(Writer writer) |
protected List<RotationAction> rotationActions
protected OutputCollector collector
protected transient org.apache.hadoop.fs.FileSystem fs
protected SyncPolicy syncPolicy
protected FileRotationPolicy rotationPolicy
protected FileNameFormat fileNameFormat
protected String fsUrl
protected String configKey
protected transient Object writeLock
protected transient Timer rotationTimer
protected long offset
protected Integer fileRetryCount
protected Integer tickTupleInterval
protected Integer maxOpenFiles
protected Partitioner partitioner
protected transient org.apache.hadoop.conf.Configuration hdfsConfig
protected void rotateOutputFile(Writer writer) throws IOException
IOException
public final void prepare(Map conf, TopologyContext topologyContext, OutputCollector collector)
Marked as final to prevent override. Subclasses should implement the doPrepare() method.
conf
- topologyContext
- collector
- public final void execute(Tuple tuple)
IBolt
Process a single tuple of input. The Tuple object contains metadata on it about which component/stream/task it came from. The values of the Tuple can be accessed using Tuple#getValue. The IBolt does not have to process the Tuple immediately. It is perfectly fine to hang onto a tuple and process it later (for instance, to do an aggregation or join).
Tuples should be emitted using the OutputCollector provided through the prepare method. It is required that all input tuples are acked or failed at some point using the OutputCollector. Otherwise, Storm will be unable to determine when tuples coming off the spouts have been completed.
For the common case of acking an input tuple at the end of the execute method, see IBasicBolt which automates this.
tuple
- The input tuple to be processed.public Map<String,Object> getComponentConfiguration()
IComponent
Declare configuration specific to this component. Only a subset of the “topology.*” configs can be overridden. The component configuration can be further overridden when constructing the topology using TopologyBuilder
getComponentConfiguration
in interface IComponent
getComponentConfiguration
in class BaseComponent
public void declareOutputFields(OutputFieldsDeclarer outputFieldsDeclarer)
IComponent
Declare the output schema for all the streams of this topology.
outputFieldsDeclarer
- this is used to declare output stream ids, output fields, and whether or not each output stream is a direct streampublic void cleanup()
IBolt
Called when an IBolt is going to be shutdown. There is no guarentee that cleanup will be called, because the supervisor kill -9’s worker processes on the cluster.
The one context where cleanup is guaranteed to be called is when a topology is killed when running Storm in local mode.
cleanup
in interface IBolt
cleanup
in class BaseRichBolt
protected org.apache.hadoop.fs.Path getBasePathForNextFile(Tuple tuple)
protected abstract void doPrepare(Map conf, TopologyContext topologyContext, OutputCollector collector) throws IOException
IOException
protected abstract Writer makeNewWriter(org.apache.hadoop.fs.Path path, Tuple tuple) throws IOException
IOException
Copyright © 2022 The Apache Software Foundation. All Rights Reserved.