Changes between Version 17 and Version 18 of HyperTable
- Timestamp:
- Apr 17, 2008, 10:11:41 AM (17 years ago)
Legend:
- Unmodified
- Added
- Removed
- Modified
-
HyperTable
v17 v18 49 49 * The AOL query logs were inserted into an 8-node Hypertable cluster. The average size of each row key was ~7 bytes and each value was ~15 bytes. The insert rate (with 4 simultaneous insert processes) was approximately 410K inserts/s. The table was scanned at a rate of approximately 671K cells/s. 50 50 == 【How we use it】 == 51 * 以陽明的病人屬性分類資料為範本來建立搜尋引擎 52 * 參考資料: 53 * [http://trac.nchc.org.tw/grid/wiki/BioInfo_Project/2008-03-19 待解問題1] 54 * [http://trac.nchc.org.tw/grid/wiki/BioInfo_Project/2008-04-10 轉檔程式] 55 * 匯入方式:[[BR]] 56 將陽明的病人屬性分類資料(XML格式),轉換成Hypertable可批次讀取的TSV格式,再使用HQL指令載入到Hypertable當中 57 * 匯出方式:[[BR]] 58 使用HQL指令匯出TSV檔案 59 * 搜尋方式:[[BR]] 60 目前Hypertable只提供以Rowkey完全比對及Rowkey或Timestamp的區間比對,以陽明的範例為例:[[BR]] 61 Source是指檢查之前診斷判定的病因;[[BR]] 62 Primary Site是指檢查之後診斷判定的病因;[[BR]] 63 如果用Source為Rowkey來搜尋"Breast",搜尋結果會查出所有Source為"Breast"的資料[[BR]] 64 如果用Primary Site為Rowkey來搜尋"Breast",搜尋結果會查出所有Primary Site為"Breast"的資料[[BR]] 65 * NOTE: 交互比對的部份,Hypertable目前似乎沒提供,這個有可能要另外的程式再來處理過 51 66 == 【Similar Project】 == 52 67 * [http://hadoop.apache.org/hbase/ HBase]: Bigtable-like structured storage for Hadoop HDFS