微软Azure云平台Hbase 的使用

简介: In this article What is HBase? Prerequisites Provision HBase clusters using Azure Management portal Mange HBase tables using HBase shell Use Hiv...

In this article

What is HBase?

HBase is a low-latency NoSQL database that allows online transactional processing of big data. HBase is offered as a managed cluster integrated into the Azure environment. The clusters are configured to store data directly in Azure Blob storage, which provides low latency and increased elasticity in performance/cost choices. This enables customers to build interactive websites that work with large datasets, to build services that store sensor and telemetry data from millions of end points, and to analyze this data with Hadoop jobs. For more information on HBase and the scenarios it can be used for, see HDInsight HBase overview.

Prerequisites

Before you begin this tutorial, you must have the following:

Provision an HBase cluster on the Azure portal

This section describes how to provision an HBase cluster using the Azure Management portal.

To provision an HDInsight cluster in the Azure Management portal

  1. Sign in to the Azure Management Portal.
  2. Click NEW on the lower left, and then click DATA SERVICES, HDINSIGHT, HBASE.
  3. Enter CLUSTER NAME, CLUSTER SIZE, CLUSTER USER PASSWORD, and STORAGE ACCOUNT.

    Choosing and HBase cluster type and entering cluster login credentials.

  4. Click on the check icon on the lower left to create the HBase cluster.

Create an HBase sample table from the HBase shell

This section describes how to enable and use the Remote Desktop Protocol (RDP) to access the HBase shell and then use it to create an HBase sample table, add rows, and then list the rows in the table.

It assumes you have completed the procedure outlined in the first section, and so have already successfully created an HBase cluster.

To enable the RDP connection to the HBase cluster

  1. From the Management portal, click HDINSIGHT from the left to view the list of the existing clusters.
  2. Click the HBase cluster where you want to open HBase Shell.
  3. Click CONFIGURATION from the top.
  4. Click ENABLE REMOTE from the bottom.
  5. Enter the RDP user name and password. The user name must be different from the cluster user name you used when provisioning the cluster. TheEXPIRES ON data can be up to seven days from today.
  6. Click the check on the lower right to enable remote desktop.
  7. After the RPD is enabled, click CONNECT from the bottom of the CONFIGURATION tab, and follow the instructions.

To open the HBase Shell

  1. Within your RDP session, click on the Hadoop Command Line shortcut located on the desktop.

  2. Change the folder to the HBase home directory:

    cd %HBASE_HOME%\bin
  3. Open the HBase shell:

    hbase shell

To create a sample table, add data and retrieve the data

  1. Create a sample table:

    create 'sampletable', 'cf1'
  2. Add a row to the sample table:

    put 'sampletable', 'row1', 'cf1:col1', 'value1'
  3. List the rows in the sample table:

    scan 'sampletable'

Check cluster status in the HBase WebUI

HBase also ships with a WebUI that helps monitoring your cluster, for example by providing request statistics or information about regions. On the HBase cluster you can find the WebUI under the address of the zookeepernode.

http://zookeepernode:60010/master-status

In a HighAvailability (HA) cluster, you will find a link to the current active HBase master node hosting the WebUI.

Bulk load a sample table

  1. Create samplefile1.txt containing the following data, and upload to Azure Blob Storage to /tmp/samplefile1.txt:

    row1    c1  c2
    row2    c1  c2
    row3    c1  c2
    row4    c1  c2
    row5    c1  c2
    row6    c1  c2
    row7    c1  c2
    row8    c1  c2
    row9    c1  c2
    row10    c1 c2
  2. Change the folder to the HBase home directory:

    cd %HBASE_HOME%\bin
  3. Execute ImportTsv:

    hbase org.apache.hadoop.hbase.mapreduce.ImportTsv -Dimporttsv.columns="HBASE_ROW_KEY,a:b,a:c" -Dimporttsv.bulk.output=/tmpOutput sampletable2 /tmp/samplefile1.txt
  4. Load the output from prior command into HBase:

    hbase org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles /tmpOutput sampletable2

Use Hive to query an HBase table

Now you have an HBase cluster provisioned and have created an HBase table, you can query it using Hive. This section creates a Hive table that maps to the HBase table and uses it to queries the data in your HBase table.

To open cluster dashboard

  1. Sign in to the Azure Management Portal.
  2. Click HDINSIGHT from the left pane. You shall see a list of clusters created including the one you just created in the last section.
  3. Click the cluster name where you want to run the Hive job.
  4. Click QUERY CONSOLE from the bottom of the page to open cluster dashboard. It opens a Web page on a different browser tab.
  5. Enter the Hadoop User account username and password. The default username is admin, the password is what you entered during the provision process. A new browser tab is opened.
  6. Click Hive Editor from the top. The Hive Editor looks like :

    HDInsight cluster dashboard.

To run Hive queries

  1. Enter the HiveQL script below into Hive Editor and click SUBMIT to create an Hive Table mapping to the HBase table. Make sure that you have created the sampletable table referenced here in HBase using the HBase Shell before executing this statement.

    CREATE EXTERNAL TABLE hbasesampletable(rowkey STRING, col1 STRING, col2 STRING) STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH SERDEPROPERTIES ('hbase.columns.mapping' = ':key,cf1:col1,cf1:col2') TBLPROPERTIES ('hbase.table.name' = 'sampletable');

    Wait until the Status is updated to Completed.

  2. Enter the HiveQL script below into Hive Editor, and then click SUBMIT button. The Hive query queries the data in the HBase table:

    SELECT count(*) FROM hbasesampletable;
  3. To retrieve the results of the Hive query, click on the View Details link in the Job Session window when the job finishes executing. The Job Output shall be 1 because you only put one record into the HBase table.

To browse the output file

  1. From Query Console, click File Browser from the top.
  2. Click the Azure Storage account used as the default file system for the HBase cluster.
  3. Click the HBase cluster name. The default Azure storage account container uses the cluster name.
  4. Click user.
  5. Click admin. This is the Hadoop user name.
  6. Click the job name with the Last Modified time matching the time when the SELECT Hive query ran.
  7. Click stdout. Save the file and open the file with Notepad. The output shall be 1.

    HDInsight HBase Hive Editor File Browser

Use HBase REST Client Library for .NET C# APIs to create an HBase table and retrieve data from the table

The Microsoft HBase REST Client Library for .NET project must be downloaded from GitHub and the project built to use the HBase .NET SDK. The following procedure includes the instructions for this task.

  1. Create a new C# Visual Studio Windows Desktop Console application.
  2. Open NuGet Package Manager Console by click the TOOLS menu, NuGet Package Manager, Package Manager Console.
  3. Run the following NuGet command in the console:

    Install-Package Microsoft.HBase.Client

  4. Add the following using statements on the top of the file:

    using Microsoft.HBase.Client; using org.apache.hadoop.hbase.rest.protobuf.generated;
  5. Replace the Main function with the following:

    static void Main(string[] args) { string clusterURL = "https://<yourHBaseClusterName>.azurehdinsight.net"; string hadoopUsername= "<yourHadoopUsername>"; string hadoopUserPassword = "<yourHadoopUserPassword>"; string hbaseTableName = "sampleHbaseTable"; // Create a new instance of an HBase client. ClusterCredentials creds = new ClusterCredentials(new Uri(clusterURL), hadoopUsername, hadoopUserPassword); HBaseClient hbaseClient = new HBaseClient(creds); // Retrieve the cluster version var version = hbaseClient.GetVersion(); Console.WriteLine("The HBase cluster version is " + version); // Create a new HBase table. TableSchema testTableSchema = new TableSchema(); testTableSchema.name = hbaseTableName; testTableSchema.columns.Add(new ColumnSchema() { name = "d" }); testTableSchema.columns.Add(new ColumnSchema() { name = "f" }); hbaseClient.CreateTable(testTableSchema); // Insert data into the HBase table. string testKey = "content"; string testValue = "the force is strong in this column"; CellSet cellSet = new CellSet(); CellSet.Row cellSetRow = new CellSet.Row { key = Encoding.UTF8.GetBytes(testKey) }; cellSet.rows.Add(cellSetRow); Cell value = new Cell { column = Encoding.UTF8.GetBytes("d:starwars"), data = Encoding.UTF8.GetBytes(testValue) }; cellSetRow.values.Add(value); hbaseClient.StoreCells(hbaseTableName, cellSet); // Retrieve a cell by its key. cellSet = hbaseClient.GetCells(hbaseTableName, testKey); Console.WriteLine("The data with the key '" + testKey + "' is: " + Encoding.UTF8.GetString(cellSet.rows[0].values[0].data)); // with the previous insert, it should yield: "the force is strong in this column" //Scan over rows in a table. Assume the table has integer keys and you want data between keys 25 and 35. Scanner scanSettings = new Scanner() { batch = 10, startRow = BitConverter.GetBytes(25), endRow = BitConverter.GetBytes(35) }; ScannerInformation scannerInfo = hbaseClient.CreateScanner(hbaseTableName, scanSettings); CellSet next = null; Console.WriteLine("Scan results"); while ((next = hbaseClient.ScannerGetNext(scannerInfo)) != null) { foreach (CellSet.Row row in next.rows) { Console.WriteLine(row.key + " : " + Encoding.UTF8.GetString(row.values[0].data)); } } Console.WriteLine("Press ENTER to continue ..."); Console.ReadLine(); }
  6. Set the first three variables in the Main function.

  7. Press F5 to run the application.

What's Next?

In this tutorial, you have learned how to provision an HBase cluster, how to create tables, and and view the data in those tables from the HBase shell. You also learned how use Hive to query the data in HBase tables and how to use the HBase C# APIs to create an HBase table and retrieve data from the table.

To learn more, see:

相关实践学习
lindorm多模间数据无缝流转
展现了Lindorm多模融合能力——用kafka API写入,无缝流转在各引擎内进行数据存储和计算的实验。
云数据库HBase版使用教程
&nbsp; 相关的阿里云产品:云数据库 HBase 版 面向大数据领域的一站式NoSQL服务,100%兼容开源HBase并深度扩展,支持海量数据下的实时存储、高并发吞吐、轻SQL分析、全文检索、时序时空查询等能力,是风控、推荐、广告、物联网、车联网、Feeds流、数据大屏等场景首选数据库,是为淘宝、支付宝、菜鸟等众多阿里核心业务提供关键支撑的数据库。 了解产品详情:&nbsp;https://cn.aliyun.com/product/hbase &nbsp; ------------------------------------------------------------------------- 阿里云数据库体验:数据库上云实战 开发者云会免费提供一台带自建MySQL的源数据库&nbsp;ECS 实例和一台目标数据库&nbsp;RDS实例。跟着指引,您可以一步步实现将ECS自建数据库迁移到目标数据库RDS。 点击下方链接,领取免费ECS&amp;RDS资源,30分钟完成数据库上云实战!https://developer.aliyun.com/adc/scenario/51eefbd1894e42f6bb9acacadd3f9121?spm=a2c6h.13788135.J_3257954370.9.4ba85f24utseFl
相关文章
|
分布式计算 Hadoop 分布式数据库
《CDP企业数据云平台从入门到实践》——HBase 迁移到 CDP(1)
《CDP企业数据云平台从入门到实践》——HBase 迁移到 CDP(1)
200 0
|
Shell 分布式数据库 API
《CDP企业数据云平台从入门到实践》——HBase 迁移到 CDP(2)
《CDP企业数据云平台从入门到实践》——HBase 迁移到 CDP(2)
162 0
|
安全 Shell 测试技术
《CDP企业数据云平台从入门到实践》——HBase 迁移到 CDP(3)
《CDP企业数据云平台从入门到实践》——HBase 迁移到 CDP(3)
189 0
|
机器学习/深度学习 人工智能 编解码
微软行星云计算Planetary Computer——微软云平台的主要可视化APP(全球10米的最新土地覆盖,森林毁坏监测,生态系统监测,森林碳风险评估,AI加持的的土地覆盖制图)
微软行星云计算Planetary Computer——微软云平台的主要可视化APP(全球10米的最新土地覆盖,森林毁坏监测,生态系统监测,森林碳风险评估,AI加持的的土地覆盖制图)
376 0
微软行星云计算Planetary Computer——微软云平台的主要可视化APP(全球10米的最新土地覆盖,森林毁坏监测,生态系统监测,森林碳风险评估,AI加持的的土地覆盖制图)
|
SQL 分布式计算 Hadoop
Azure 云平台用 SQOOP 将 SQL server 2012 数据表导入 HIVE / HBASE
My name is Farooq and I am with HDinsight support team here at Microsoft. In this blog I will try to give some brief overview of Sqoop in HDinsight an...
1079 0