本文整理汇总了Java中backtype.storm.metric.api.CountMetric类的典型用法代码示例。如果您正苦于以下问题:Java CountMetric类的具体用法?Java CountMetric怎么用?Java CountMetric使用的例子?那么恭喜您, 这里精选的类代码示例或许可以为您提供帮助。
CountMetric类属于backtype.storm.metric.api包,在下文中一共展示了CountMetric类的11个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于我们的系统推荐出更棒的Java代码示例。
示例1: open
import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
@Override
public void open(Map conf, TopologyContext context,
SpoutOutputCollector _collector) {
collector = _collector;
_spoutMetric = new CountMetric();
context.registerMetric(CatMetricUtil.getSpoutMetricName(topic, group),
_spoutMetric, Constants.EMIT_FREQUENCY_IN_SECONDS);
ConsumerConfig config = new ConsumerConfig();
try {
consumer = new Consumer(topic, group, config);
} catch (LionException e) {
throw new RuntimeException(e);
}
consumer.start();
stream = consumer.getStream();
fetchThread = new MessageFetcher(stream);
new Thread(fetchThread).start();
}
开发者ID:kkllwww007,项目名称:jstrom,代码行数:21,代码来源:BlackholeBlockingQueueSpout.java
示例2: open
import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
@Override
public void open(Map conf, TopologyContext context,
SpoutOutputCollector _collector) {
collector = _collector;
_spoutMetric = new CountMetric();
context.registerMetric(CatMetricUtil.getSpoutMetricName(topic, group),
_spoutMetric, Constants.EMIT_FREQUENCY_IN_SECONDS);
ConsumerConfig config = new ConsumerConfig();
try {
consumer = new Consumer(topic, group, config);
} catch (LionException e) {
throw new RuntimeException(e);
}
consumer.start();
stream = consumer.getStream();
}
开发者ID:kkllwww007,项目名称:jstrom,代码行数:18,代码来源:BlackholeSpout.java
示例3: registerMetrics
import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
@SuppressWarnings("rawtypes")
public void registerMetrics(Map conf, IMetricsContext context, String mapStateMetricName) {
int bucketSize = (Integer) (conf.get(Config.TOPOLOGY_BUILTIN_METRICS_BUCKET_SIZE_SECS));
String metricBaseName = "cassandra/" + mapStateMetricName;
_mreads = context.registerMetric(metricBaseName + "/readCount", new CountMetric(), bucketSize);
_mwrites = context.registerMetric(metricBaseName + "/writeCount", new CountMetric(), bucketSize);
_mexceptions = context.registerMetric(metricBaseName + "/exceptionCount", new CountMetric(), bucketSize);
}
开发者ID:hpcc-systems,项目名称:storm-cassandra-cql,代码行数:9,代码来源:CassandraCqlMapState.java
示例4: registerMetrics
import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
/**
* Register Metrics in Storm
* @param conf A set of properties
* @param context Metrics context
*/
public void registerMetrics(Map conf, IMetricsContext context) {
int bucketSize = (Integer) (conf.get(Config.TOPOLOGY_BUILTIN_METRICS_BUCKET_SIZE_SECS));
_mreads = context.registerMetric("mongodb/readCount", new CountMetric(), bucketSize);
_mwrites = context.registerMetric("mongodb/writeCount", new CountMetric(), bucketSize);
_mexceptions = context.registerMetric("mongodb/exceptionCount", new CountMetric(), bucketSize);
}
开发者ID:andressanchez,项目名称:storm-mongodb,代码行数:12,代码来源:MongoDBMapState.java
示例5: open
import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
@Override
public void open(Map conf, TopologyContext context, SpoutOutputCollector collector) {
_throttler = new WindowedTimeThrottler((Number) conf.get(Config.TOPOLOGY_TRIDENT_BATCH_EMIT_INTERVAL_MILLIS),
Integer.MAX_VALUE);
for (String spoutId : _managedSpoutIds) {
_states.add(TransactionalState.newCoordinatorState(conf, spoutId));
}
_currTransaction = getStoredCurrTransaction();
_collector = collector;
Number active = (Number) conf.get(Config.TOPOLOGY_MAX_SPOUT_PENDING);
if (active == null) {
_maxTransactionActive = 1;
} else {
_maxTransactionActive = active.intValue();
}
_attemptIds = getStoredCurrAttempts(_currTransaction, _maxTransactionActive);
avgWaittingTime = context.registerMetric("commit-wait-ms", new ReducedMetric(new MeanReducer()),
Utils.getInt(conf.get(Config.TOPOLOGY_BUILTIN_METRICS_BUCKET_SIZE_SECS)));
commitWaittingCount = context.registerMetric("commit-wait-cnt", new CountMetric(),
Utils.getInt(conf.get(Config.TOPOLOGY_BUILTIN_METRICS_BUCKET_SIZE_SECS)));
activeTxCount = context.registerMetric("active-tx-cnt", new ReducedMetric(new MeanReducer()),
Utils.getInt(conf.get(Config.TOPOLOGY_BUILTIN_METRICS_BUCKET_SIZE_SECS)));
for (int i = 0; i < _spouts.size(); i++) {
String txId = _managedSpoutIds.get(i);
_coordinators.add(_spouts.get(i).getCoordinator(txId, conf, context));
}
}
开发者ID:troyding,项目名称:storm-resa,代码行数:29,代码来源:MasterBatchCoordinator.java
示例6: prepare
import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
@Override
public void prepare(Map config, TopologyContext context) {
// Optional set logging
if (config.get(CONF_LOGGING) != null) {
m_logging = (Boolean) config.get(CONF_LOGGING);
} else {
m_logging = false;
}
// Tuple counter metric
if (config.get(CONF_METRIC_LOGGING_INTERVALL) != null) {
m_countMetric = new CountMetric();
context.registerMetric("tuple_count", m_countMetric,
((Long) config.get(CONF_METRIC_LOGGING_INTERVALL)).intValue());
}
LOG.info("Loading SVM model...");
Dataset dataset = Configuration.getDataSetSemEval2013();
m_model = SerializationUtils.deserialize(dataset.getDatasetPath()
+ File.separator + SVM.SVM_MODEL_FILE_SER);
if (m_model == null) {
LOG.error("Could not load SVM model! File: " + dataset.getDatasetPath()
+ File.separator + SVM.SVM_MODEL_FILE_SER);
throw new RuntimeException();
}
}
开发者ID:millecker,项目名称:storm-apps,代码行数:28,代码来源:SVMBolt.java
示例7: PartitionManager
import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
public PartitionManager(DynamicPartitionConnections connections, String topologyInstanceId, ZkState state, Map stormConf, SpoutConfig spoutConfig, Partition id) {
_partition = id;
_connections = connections;
_spoutConfig = spoutConfig;
_topologyInstanceId = topologyInstanceId;
_consumer = connections.register(id.host, id.partition);
_state = state;
_stormConf = stormConf;
numberAcked = numberFailed = 0;
String jsonTopologyId = null;
Long jsonOffset = null;
String path = committedPath();
try {
Map<Object, Object> json = _state.readJSON(path);
LOG.info("Read partition information from: " + path + " --> " + json );
if (json != null) {
jsonTopologyId = (String) ((Map<Object, Object>) json.get("topology")).get("id");
jsonOffset = (Long) json.get("offset");
}
} catch (Throwable e) {
LOG.warn("Error reading and/or parsing at ZkNode: " + path, e);
}
Long currentOffset = KafkaUtils.getOffset(_consumer, spoutConfig.topic, id.partition, spoutConfig);
if (jsonTopologyId == null || jsonOffset == null) { // failed to parse JSON?
_committedTo = currentOffset;
LOG.info("No partition information found, using configuration to determine offset");
} else if (!topologyInstanceId.equals(jsonTopologyId) && spoutConfig.forceFromStart) {
_committedTo = KafkaUtils.getOffset(_consumer, spoutConfig.topic, id.partition, spoutConfig.startOffsetTime);
LOG.info("Topology change detected and reset from start forced, using configuration to determine offset");
} else {
_committedTo = jsonOffset;
LOG.info("Read last commit offset from zookeeper: " + _committedTo + "; old topology_id: " + jsonTopologyId + " - new topology_id: " + topologyInstanceId );
}
if (currentOffset - _committedTo > spoutConfig.maxOffsetBehind || _committedTo <= 0) {
LOG.info("Last commit offset from zookeeper: " + _committedTo);
_committedTo = currentOffset;
LOG.info("Commit offset " + _committedTo + " is more than " +
spoutConfig.maxOffsetBehind + " behind, resetting to startOffsetTime=" + spoutConfig.startOffsetTime);
}
LOG.info("Starting Kafka " + _consumer.host() + ":" + id.partition + " from offset " + _committedTo);
_emittedToOffset = _committedTo;
_fetchAPILatencyMax = new CombinedMetric(new MaxMetric());
_fetchAPILatencyMean = new ReducedMetric(new MeanReducer());
_fetchAPICallCount = new CountMetric();
_fetchAPIMessageCount = new CountMetric();
}
开发者ID:redBorder,项目名称:rb-bi,代码行数:53,代码来源:PartitionManager.java
示例8: execute
import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
@Override
public void execute(Tuple input) {
// triggered by the arrival of a tuple
// be it a tick or normal one
flushQueues();
if (isTickTuple(input)) {
return;
}
CountMetric metric = metricGauge.scope("activethreads");
metric.getValueAndReset();
metric.incrBy(this.activeThreads.get());
metric = metricGauge.scope("in queues");
metric.getValueAndReset();
metric.incrBy(this.fetchQueues.inQueues.get());
metric = metricGauge.scope("queues");
metric.getValueAndReset();
metric.incrBy(this.fetchQueues.queues.size());
LOG.info("[Fetcher #{}] Threads : {}\tqueues : {}\tin_queues : {}",
taskIndex, this.activeThreads.get(),
this.fetchQueues.queues.size(), this.fetchQueues.inQueues.get());
if (!input.contains("url")) {
LOG.info("[Fetcher #{}] Missing field url in tuple {}", taskIndex,
input);
// ignore silently
_collector.ack(input);
return;
}
String url = input.getStringByField("url");
// has one but what about the content?
if (StringUtils.isBlank(url)) {
LOG.info("[Fetcher #{}] Missing value for field url in tuple {}",
taskIndex, input);
// ignore silently
_collector.ack(input);
return;
}
fetchQueues.addFetchItem(input);
}
开发者ID:zaizi,项目名称:alfresco-apache-storm-demo,代码行数:49,代码来源:FetcherBolt.java
示例9: initMetrics
import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
private void initMetrics(TopologyContext context)
{
_countMetric = new CountMetric();
context.registerMetric("record_count", _countMetric, 1);
}
开发者ID:dsarlis,项目名称:datix,代码行数:7,代码来源:SFlowBolt.java
示例10: PartitionManager
import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
public PartitionManager(DynamicPartitionConnections connections, String topologyInstanceId, ZkState state, Map stormConf, SpoutConfig spoutConfig, Partition id) {
_partition = id;
_connections = connections;
_spoutConfig = spoutConfig;
_topologyInstanceId = topologyInstanceId;
_consumer = connections.register(id.host, id.partition);
_state = state;
_stormConf = stormConf;
String jsonTopologyId = null;
Long jsonOffset = null;
String path = committedPath();
try {
Map<Object, Object> json = _state.readJSON(path);
LOG.info("Read partition information from: " + path + " --> " + json );
if (json != null) {
jsonTopologyId = (String) ((Map<Object, Object>) json.get("topology")).get("id");
jsonOffset = (Long) json.get("offset");
}
} catch (Throwable e) {
LOG.warn("Error reading and/or parsing at ZkNode: " + path, e);
}
if (jsonTopologyId == null || jsonOffset == null) { // failed to parse JSON?
_committedTo = KafkaUtils.getOffset(_consumer, spoutConfig.topic, id.partition, spoutConfig);
LOG.info("No partition information found, using configuration to determine offset");
} else if (!topologyInstanceId.equals(jsonTopologyId) && spoutConfig.forceFromStart) {
_committedTo = KafkaUtils.getOffset(_consumer, spoutConfig.topic, id.partition, spoutConfig.startOffsetTime);
LOG.info("Topology change detected and reset from start forced, using configuration to determine offset");
} else {
_committedTo = jsonOffset;
LOG.info("Read last commit offset from zookeeper: " + _committedTo + "; old topology_id: " + jsonTopologyId + " - new topology_id: " + topologyInstanceId );
}
LOG.info("Starting " + _partition + " from offset " + _committedTo);
_emittedToOffset = _committedTo;
_fetchAPILatencyMax = new CombinedMetric(new MaxMetric());
_fetchAPILatencyMean = new ReducedMetric(new MeanReducer());
_fetchAPICallCount = new CountMetric();
_fetchAPIMessageCount = new CountMetric();
}
开发者ID:metamx,项目名称:incubator-storm,代码行数:43,代码来源:PartitionManager.java
示例11: prepare
import backtype.storm.metric.api.CountMetric; //导入依赖的package包/类
@Override
public void prepare(@SuppressWarnings("rawtypes") Map stormNativeConfig,
TopologyContext context, OutputCollector collector) {
this.stormNativeConfig = stormNativeConfig;
this.collector = collector;
originId = operationClass.getName() + "." + context.getThisTaskIndex();
// connect to the zoopkeeper configuration
try {
zookeeperStormConfiguration = ZookeeperStormConfigurationFactory
.getInstance().getStormConfiguration(stormNativeConfig);
} catch (StormConfigurationException e) {
logger.error("Can not connect to zookeeper for get Storm configuration. Reason: "
+ e.getMessage());
// create empty config to avoid errors
zookeeperStormConfiguration = new EmptyStormConfiguration();
}
String msg = "SensorStormBolt instance created for";
if (fieldGrouperId == null) {
msg = msg + " a single_instance operation class \""
+ operationClass.getName() + "\"";
} else {
msg = msg + " a fieldGrouping operation class \""
+ operationClass.getName() + "\" grouped on tuple field \""
+ fieldGrouperId + "\"";
}
if (batcherClass != null) {
msg = msg + ", with a batcher class \"" + batcherClass.getName()
+ "\"";
} else {
msg = msg + ", with no batcher.";
}
logger.info(msg);
syncBuffer = new FlushingSyncBuffer(syncBufferSize);
operationManagers = new HashMap<String, OperationManager>();
fieldGrouperValues = new HashMap<Particle, String>();
bufferRejectMetric = new CountMetric();
context.registerMetric("syncbuffer_rejects", bufferRejectMetric,
TIME_BUCKET_SIZE_IN_SECS);
}
开发者ID:sensorstorm,项目名称:SensorStorm,代码行数:42,代码来源:SensorStormBolt.java
注:本文中的backtype.storm.metric.api.CountMetric类示例整理自Github/MSDocs等源码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。 |
请发表评论