不透明分區事務


title: storm不透明分區事務
date: 2017-08-27
categoties:
-storm
tags:

  • storm
  • java

不透明分區事務

一般情況下storm是在構造函數中對基本數據類型和可序列化的對象進行賦值和實例化,在prepare()方法中對不可序列化的對象進行實例化

接口說明

IOpaquePartitionedTransactionalSpout<T>:不透明分區事務Spout
--IOpaquePartitionedTransactionalSpout.Coordinator
--isReady():為true開啟一個新的事務,沒提交一個事務停頓一段時間,可以在此進行sleep
--IOpaquePartitionedTransactionalSpout.Emitter<X>
-- emitPartitionBatch(TransactionAttempt tx,BatchOutputCollector collector,int partition,X lastPartitionMeta)
-- numPartitions()

不透明分區事務特點

以下X表示元數據
它不區分發新消息還是重發舊消息,全部用emitPartitionBatch搞定。雖然emitPartitionBatch返回的X應該是下一批次供自己使用的(emitPartitionBatch的第4個參數),但是只有一個批次成功以后X才會更新到ZooKeeper中,如果失敗重發,emitPartitionBatch讀取的X還是舊的。所以這時候自定義的X不需要記錄當前批次的開始位置和下一批次的開始位置兩個值,只需要記錄下一批次開始位置一個值即可,例如:
public class BatchMeta{
public long nextOffset; //下一批次的偏移量
}

實例

制造一批數據對數據按照時間進行內的進行統計

定義不透明分區事務的spout,在構造函數中制造數據源

import org.apache.storm.coordination.BatchOutputCollector;
import org.apache.storm.task.TopologyContext;
import org.apache.storm.topology.OutputFieldsDeclarer;
import org.apache.storm.transactional.TransactionAttempt;
import org.apache.storm.transactional.partitioned.IOpaquePartitionedTransactionalSpout;
import org.apache.storm.tuple.Fields;
import org.apache.storm.tuple.Values;
import org.apache.storm.utils.Utils;
import java.util.HashMap;
import java.util.Map;
import java.util.Random;

public class MyOpaquePtTxSpout implements IOpaquePartitionedTransactionalSpout<MyMdata> {


    private static final long serialVersionUID = -1889920861418121463L;

    public static Map<Integer,Map<Long,String>> Partion_Data_Base = new HashMap<Integer, Map<Long, String>>();
    public static int PART_COUNT = 5;

    public MyOpaquePtTxSpout() {

        Random _rand = new Random();
        String[] hosts={"www.haoyidao.com"};
        String[] session_id =  { "ABYH6Y4V4SCVXTG6DPB4VH9U123", "XXYH6YCGFJYERTT834R52FDXV9U34", "BBYH61456FGHHJ7JL89RG5VV9UYU7",
                "CYYH6Y2345GHI899OFG4V9U567", "VVVYH6Y4V4SFXZ56JIPDPB4V678" };

        String[] time = { "2014-01-07 08:40:50", "2014-01-07 08:40:51", "2014-01-07 08:40:52", "2014-01-07 08:40:53",
                "2014-01-07 09:40:49", "2014-01-07 10:40:49", "2014-01-07 11:40:49", "2014-01-07 12:40:49" };

        for (int i=0;i<PART_COUNT;i++){
            Map<Long,String> map = new HashMap<Long, String>();
            for (long j=0;j<30;j++){
                map.put(j,hosts[0]+"\t"+session_id[_rand.nextInt(5)]+"\t"+_rand.nextInt(8));
            }
            Partion_Data_Base.put(i,map);
        }

        System.err.println("MtPtTxSpout start.....");

    }

    public Emitter<MyMdata> getEmitter(Map map, TopologyContext topologyContext) {
        return new MyEmitter();
    }

    public Coordinator getCoordinator(Map map, TopologyContext topologyContext) {
        return new MyCoordinator();
    }

    public void declareOutputFields(OutputFieldsDeclarer outputFieldsDeclarer) {

        outputFieldsDeclarer.declare(new Fields("tx","today","log"));
    }

    public Map<String, Object> getComponentConfiguration() {
        return null;
    }

    public class MyCoordinator implements IOpaquePartitionedTransactionalSpout.Coordinator{

        public boolean isReady() {
            Utils.sleep(1000);
            return true;
        }

        public void close() {

        }
    }

    public static final long BITCHNUM = 10;

    public class MyEmitter implements IOpaquePartitionedTransactionalSpout.Emitter<MyMdata>{

        public MyMdata emitPartitionBatch(TransactionAttempt transactionAttempt, BatchOutputCollector batchOutputCollector, int i, MyMdata myMdata) {

            long beginPoint = 0;
            if (myMdata == null){
                beginPoint = 0;
            }else {
                beginPoint = myMdata.getBeginPoint()+myMdata.getNum();
            }
            MyMdata myMdata1 = new MyMdata();
            myMdata1.setBeginPoint(beginPoint);
            myMdata1.setNum(BITCHNUM);
            System.err.println("啟動一個事務:"+myMdata1.toString());

            //進行發射數據
            Map<Long,String> batchMap = Partion_Data_Base.get(i);
            for (Long j = myMdata1.getBeginPoint();j<myMdata1.getBeginPoint()+myMdata1.getNum();j++){
                //表示一個分區消息發送完了
                if (batchMap.size()<j){
                    break;
                }
                batchOutputCollector.emit(new Values(transactionAttempt,"2017-08-09",batchMap.get(j)));
            }

            return myMdata1;
        }

        public int numPartitions() {
            return 5;
        }

        public void close() {

        }
    }
}

定義bolt

import org.apache.storm.coordination.BatchOutputCollector;
import org.apache.storm.coordination.IBatchBolt;
import org.apache.storm.task.TopologyContext;
import org.apache.storm.topology.OutputFieldsDeclarer;
import org.apache.storm.topology.base.BaseTransactionalBolt;
import org.apache.storm.transactional.TransactionAttempt;
import org.apache.storm.tuple.Fields;
import org.apache.storm.tuple.Tuple;
import org.apache.storm.tuple.Values;

import java.text.DateFormat;
import java.text.SimpleDateFormat;
import java.util.HashMap;
import java.util.Map;

public class Mybolt implements IBatchBolt<TransactionAttempt> {

    private static  Map<String,Integer> countMap = new HashMap<String, Integer>();
    private BatchOutputCollector batchOutputCollector;
    private TransactionAttempt transactionAttempt;
    DateFormat dateFormat = new SimpleDateFormat("yyyy-MM-dd");
    String today;
    Integer count;



    public void prepare(Map map, TopologyContext topologyContext, BatchOutputCollector batchOutputCollector, TransactionAttempt transactionAttempt) {
        this.batchOutputCollector = batchOutputCollector;
        this.transactionAttempt = transactionAttempt;

    }

    public void execute(Tuple tuple) {
        TransactionAttempt transactionAttempt = (TransactionAttempt) tuple.getValue(0);
        String log = (String) tuple.getValue(1);
        String[] strings = log.split("\t");
        today = "2014-01-07";
         count = countMap.get(today);
        if (count == null){
            count = 0;
        }
        count++;
        countMap.put(today,count);

    }

    public void finishBatch() {
        System.err.println(this.transactionAttempt.toString()+"---"+today+"----"+count);
        this.batchOutputCollector.emit(new Values(this.transactionAttempt,today,count));

    }

    public void declareOutputFields(OutputFieldsDeclarer outputFieldsDeclarer) {

        outputFieldsDeclarer.declare(new Fields("tx","today","count"));

    }

    public Map<String, Object> getComponentConfiguration() {
        return null;
    }
}

定義提交報告的commit

import org.apache.storm.coordination.BatchOutputCollector;
import org.apache.storm.task.TopologyContext;
import org.apache.storm.topology.OutputFieldsDeclarer;
import org.apache.storm.topology.base.BaseTransactionalBolt;
import org.apache.storm.transactional.ICommitter;
import org.apache.storm.transactional.TransactionAttempt;
import org.apache.storm.tuple.Tuple;
import org.apache.storm.tuple.Values;

import java.math.BigInteger;
import java.util.HashMap;
import java.util.Map;

public class MyCommitbolt extends BaseTransactionalBolt implements ICommitter{

    private static final long serialVersionUID = 1550322821850864035L;

    public static Map<String,DbValue> dbValueMap = new HashMap<String, DbValue>();
    Map<String,Integer> countMap = new HashMap<String, Integer>();
    TransactionAttempt transactionAttempt;
    BatchOutputCollector collector;
    String today = null;


    public void prepare(Map map, TopologyContext topologyContext, BatchOutputCollector batchOutputCollector, TransactionAttempt transactionAttempt) {

        this.transactionAttempt = transactionAttempt;
        this.collector = batchOutputCollector;
    }

    public void execute(Tuple tuple) {
        today = tuple.getString(1);
        Integer count = tuple.getInteger(2);
        transactionAttempt = (TransactionAttempt) tuple.getValue(0);
        if (today!=null&&count!=null){
            Integer batchCount = countMap.get(today);
            if (batchCount==null){
                batchCount = 0;
            }
            batchCount = batchCount+count;
            countMap.put(today,batchCount);
        }

    }

    public void finishBatch() {
        if (countMap.size()>0){
            DbValue dbValue = dbValueMap.get(today);

            DbValue newValue;

            if (dbValue==null||!dbValue.txid.equals(transactionAttempt.getTransactionId())){
                newValue = new DbValue();
                newValue.txid = transactionAttempt.getTransactionId();
                newValue.dataStr = today;
                if (dbValue==null){
                    newValue.count = countMap.get(today);
                    newValue.pre_cout = 0;
                }else {
                    newValue.count = dbValue.count+countMap.get(today);
                    newValue.pre_cout = dbValue.count;
                }

                dbValueMap.put(today,newValue);
            }

            System.out.println("total==========:"+dbValueMap.get(today).count);

        }


    }

    public void declareOutputFields(OutputFieldsDeclarer outputFieldsDeclarer) {

    }

    public static class DbValue{
        BigInteger txid;
        int count = 0;
        String dataStr;
        int pre_cout;
    }
}

定義topo結構

import org.apache.storm.coordination.BatchOutputCollector;
import org.apache.storm.task.TopologyContext;
import org.apache.storm.topology.OutputFieldsDeclarer;
import org.apache.storm.topology.base.BaseTransactionalBolt;
import org.apache.storm.transactional.ICommitter;
import org.apache.storm.transactional.TransactionAttempt;
import org.apache.storm.tuple.Tuple;
import org.apache.storm.tuple.Values;

import java.math.BigInteger;
import java.util.HashMap;
import java.util.Map;

public class MyCommitbolt extends BaseTransactionalBolt implements ICommitter{

    private static final long serialVersionUID = 1550322821850864035L;

    public static Map<String,DbValue> dbValueMap = new HashMap<String, DbValue>();
    Map<String,Integer> countMap = new HashMap<String, Integer>();
    TransactionAttempt transactionAttempt;
    BatchOutputCollector collector;
    String today = null;


    public void prepare(Map map, TopologyContext topologyContext, BatchOutputCollector batchOutputCollector, TransactionAttempt transactionAttempt) {

        this.transactionAttempt = transactionAttempt;
        this.collector = batchOutputCollector;
    }

    public void execute(Tuple tuple) {
        today = tuple.getString(1);
        Integer count = tuple.getInteger(2);
        transactionAttempt = (TransactionAttempt) tuple.getValue(0);
        if (today!=null&&count!=null){
            Integer batchCount = countMap.get(today);
            if (batchCount==null){
                batchCount = 0;
            }
            batchCount = batchCount+count;
            countMap.put(today,batchCount);
        }

    }

    public void finishBatch() {
        if (countMap.size()>0){
            DbValue dbValue = dbValueMap.get(today);

            DbValue newValue;

            if (dbValue==null||!dbValue.txid.equals(transactionAttempt.getTransactionId())){
                newValue = new DbValue();
                newValue.txid = transactionAttempt.getTransactionId();
                newValue.dataStr = today;
                if (dbValue==null){
                    newValue.count = countMap.get(today);
                    newValue.pre_cout = 0;
                }else {
                    newValue.count = dbValue.count+countMap.get(today);
                    newValue.pre_cout = dbValue.count;
                }

                dbValueMap.put(today,newValue);
            }

            System.out.println("total==========:"+dbValueMap.get(today).count);

        }


    }

    public void declareOutputFields(OutputFieldsDeclarer outputFieldsDeclarer) {

    }

    public static class DbValue{
        BigInteger txid;
        int count = 0;
        String dataStr;
        int pre_cout;

    }

}
最后編輯于
?著作權歸作者所有,轉載或內容合作請聯系作者
平臺聲明:文章內容(如有圖片或視頻亦包括在內)由作者上傳并發布,文章內容僅代表作者本人觀點,簡書系信息發布平臺,僅提供信息存儲服務。
  • 序言:七十年代末,一起剝皮案震驚了整個濱河市,隨后出現的幾起案子,更是在濱河造成了極大的恐慌,老刑警劉巖,帶你破解...
    沈念sama閱讀 229,885評論 6 541
  • 序言:濱河連續發生了三起死亡事件,死亡現場離奇詭異,居然都是意外死亡,警方通過查閱死者的電腦和手機,發現死者居然都...
    沈念sama閱讀 99,312評論 3 429
  • 文/潘曉璐 我一進店門,熙熙樓的掌柜王于貴愁眉苦臉地迎上來,“玉大人,你說我怎么就攤上這事。” “怎么了?”我有些...
    開封第一講書人閱讀 177,993評論 0 383
  • 文/不壞的土叔 我叫張陵,是天一觀的道長。 經常有香客問我,道長,這世上最難降的妖魔是什么? 我笑而不...
    開封第一講書人閱讀 63,667評論 1 317
  • 正文 為了忘掉前任,我火速辦了婚禮,結果婚禮上,老公的妹妹穿的比我還像新娘。我一直安慰自己,他們只是感情好,可當我...
    茶點故事閱讀 72,410評論 6 411
  • 文/花漫 我一把揭開白布。 她就那樣靜靜地躺著,像睡著了一般。 火紅的嫁衣襯著肌膚如雪。 梳的紋絲不亂的頭發上,一...
    開封第一講書人閱讀 55,778評論 1 328
  • 那天,我揣著相機與錄音,去河邊找鬼。 笑死,一個胖子當著我的面吹牛,可吹牛的內容都是我干的。 我是一名探鬼主播,決...
    沈念sama閱讀 43,775評論 3 446
  • 文/蒼蘭香墨 我猛地睜開眼,長吁一口氣:“原來是場噩夢啊……” “哼!你這毒婦竟也來了?” 一聲冷哼從身側響起,我...
    開封第一講書人閱讀 42,955評論 0 289
  • 序言:老撾萬榮一對情侶失蹤,失蹤者是張志新(化名)和其女友劉穎,沒想到半個月后,有當地人在樹林里發現了一具尸體,經...
    沈念sama閱讀 49,521評論 1 335
  • 正文 獨居荒郊野嶺守林人離奇死亡,尸身上長有42處帶血的膿包…… 初始之章·張勛 以下內容為張勛視角 年9月15日...
    茶點故事閱讀 41,266評論 3 358
  • 正文 我和宋清朗相戀三年,在試婚紗的時候發現自己被綠了。 大學時的朋友給我發了我未婚夫和他白月光在一起吃飯的照片。...
    茶點故事閱讀 43,468評論 1 374
  • 序言:一個原本活蹦亂跳的男人離奇死亡,死狀恐怖,靈堂內的尸體忽然破棺而出,到底是詐尸還是另有隱情,我是刑警寧澤,帶...
    沈念sama閱讀 38,998評論 5 363
  • 正文 年R本政府宣布,位于F島的核電站,受9級特大地震影響,放射性物質發生泄漏。R本人自食惡果不足惜,卻給世界環境...
    茶點故事閱讀 44,696評論 3 348
  • 文/蒙蒙 一、第九天 我趴在偏房一處隱蔽的房頂上張望。 院中可真熱鬧,春花似錦、人聲如沸。這莊子的主人今日做“春日...
    開封第一講書人閱讀 35,095評論 0 28
  • 文/蒼蘭香墨 我抬頭看了看天上的太陽。三九已至,卻和暖如春,著一層夾襖步出監牢的瞬間,已是汗流浹背。 一陣腳步聲響...
    開封第一講書人閱讀 36,385評論 1 294
  • 我被黑心中介騙來泰國打工, 沒想到剛下飛機就差點兒被人妖公主榨干…… 1. 我叫王不留,地道東北人。 一個月前我還...
    沈念sama閱讀 52,193評論 3 398
  • 正文 我出身青樓,卻偏偏與公主長得像,于是被迫代替她去往敵國和親。 傳聞我的和親對象是個殘疾皇子,可洞房花燭夜當晚...
    茶點故事閱讀 48,431評論 2 378

推薦閱讀更多精彩內容