|
@@ -1,10 +1,12 @@
|
|
|
package com.persagy.iot.utils
|
|
|
|
|
|
import org.apache.hadoop.conf.Configuration
|
|
|
-import org.apache.hadoop.hbase.{HBaseConfiguration, HColumnDescriptor, HTableDescriptor, TableName}
|
|
|
-import org.apache.hadoop.hbase.client.{Admin, Connection, ConnectionFactory, Put}
|
|
|
+import org.apache.hadoop.hbase.{Cell, CellUtil, HBaseConfiguration, HColumnDescriptor, HTableDescriptor, TableName}
|
|
|
+import org.apache.hadoop.hbase.client.{Admin, Connection, ConnectionFactory, Get, Put, Result, ResultScanner, Scan, Table}
|
|
|
import org.apache.hadoop.hbase.util.Bytes
|
|
|
|
|
|
+import java.util
|
|
|
+import scala.collection.JavaConverters.asScalaBufferConverter
|
|
|
import scala.util.Try
|
|
|
|
|
|
object HbaseUtil {
|
|
@@ -39,10 +41,27 @@ object HbaseUtil {
|
|
|
*/
|
|
|
def insertData(conn: Connection, rowKey: String, name: String, valueMap: Map[Object, Map[Object, Object]]): Unit ={
|
|
|
|
|
|
+ val get = new Get("4rowKey".getBytes)
|
|
|
+ val scan = new Scan()
|
|
|
+ val table: Table = conn.getTable(TableName.valueOf("test_create"))
|
|
|
+ var scanner: ResultScanner = null
|
|
|
+ val value: util.Iterator[Result] = scanner.iterator()
|
|
|
+ while(value.hasNext){
|
|
|
+ val result: Result = value.next()
|
|
|
+ val rowKey: String = Bytes.toString(result.getRow)
|
|
|
+ for (cell:Cell <- result.listCells().asScala){
|
|
|
+ val string: String = Bytes.toString(CellUtil.cloneQualifier(cell))
|
|
|
+ string match {
|
|
|
+ case "date" => println(rowKey + " : " + string)
|
|
|
+ case _ => println("no")
|
|
|
+ }
|
|
|
+
|
|
|
+ }
|
|
|
+ }
|
|
|
//获取表连接
|
|
|
val tableName = TableName.valueOf(name)
|
|
|
val table = conn.getTable(tableName)
|
|
|
- valueMap.
|
|
|
+
|
|
|
val put = new Put(rowKey.getBytes());
|
|
|
put.addColumn(Bytes.toBytes("data"), Bytes.toBytes("name"), Bytes.toBytes(valueMap.head._2.head._2.toString))
|
|
|
|