<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hbase on Blg</title><link>https://blg.robot-house.us/tags/hbase/</link><description>Recent content in Hbase on Blg</description><generator>Hugo -- gohugo.io</generator><language>en-us</language><lastBuildDate>Fri, 26 Feb 2021 07:46:00 +0000</lastBuildDate><atom:link href="https://blg.robot-house.us/tags/hbase/index.xml" rel="self" type="application/rss+xml"/><item><title>An OpenTSDB Retrospective</title><link>https://blg.robot-house.us/posts/opentsdb-retrospective/</link><pubDate>Fri, 26 Feb 2021 07:46:00 +0000</pubDate><guid>https://blg.robot-house.us/posts/opentsdb-retrospective/</guid><description>Over a year ago, I wrote about the success I&amp;rsquo;d had at $JOB with OpenTSDB. As with all success, this particular success meant we&amp;rsquo;d be punished&amp;hellip;with more metrics! As our need for an effective TSDB grew, OpenTSDB started to creak. We&amp;rsquo;re now in the position where we are PoCing alternative metrics storage platforms to address our growing needs.
The major problems:
The community feels stale and possibly dying. There hasn&amp;rsquo;t been a release since 2018.</description></item><item><title>Upgrading HBase Away From Ambari</title><link>https://blg.robot-house.us/posts/upgrading-hbase-away-from-ambari/</link><pubDate>Wed, 11 Mar 2020 01:39:54 +0000</pubDate><guid>https://blg.robot-house.us/posts/upgrading-hbase-away-from-ambari/</guid><description>Hortonworks and Cloudera have merged: https://www.cloudera.com/about/news-and-blogs/press-releases/2019-01-03-cloudera-and-hortonworks-complete-planned-merger.html. At $JOB, we knew this would mean our Hortonworks install may change or be deprecated. At the same time, the HBase 2.0.2 install we had been using was starting to have some trouble.
We had a few problems:
Master procedure WAL files were not being properly cleared up. This meant that a master failover would take &amp;gt; 5 minutes to complete while the master read each old procedure WAL for the last ~10 months.</description></item><item><title>Repairing Overlapping Regions in HBase >= 2.0</title><link>https://blg.robot-house.us/posts/repairing-region-overlaps-hbase/</link><pubDate>Fri, 02 Aug 2019 12:14:00 +0000</pubDate><guid>https://blg.robot-house.us/posts/repairing-region-overlaps-hbase/</guid><description>There is potential, when working with faulty hardware, that your HBase cluster can crash in weird ways, causing the hbase:meta region to report two different regions with the same start or end key. In HBase &amp;lt; 2.0, the hbck tool could fix these overlapping regions, but since that tool is disabled in HBase &amp;gt;= 2.0, how do we fix the problem?
Bringing it up means I&amp;rsquo;ve experienced it, right? Anyway&amp;hellip;</description></item><item><title>Merging Regions in HBase</title><link>https://blg.robot-house.us/posts/merging-regions-in-hbase/</link><pubDate>Mon, 10 Jun 2019 18:46:00 +0000</pubDate><guid>https://blg.robot-house.us/posts/merging-regions-in-hbase/</guid><description>So, like with any job using physical hardware, getting new machines for your growing Hadoop cluster takes time. And sometimes, the amount of ingest starts putting pressure on components before new hosts can be added. Today, our HBase hosts are edging dangerously close to &amp;gt;=200 regions per host. The HBase Book suggests 200+ regions per host is bad.
Okay, so we&amp;rsquo;re edging closer and closer to worse and worse performance as we add data (and regions naturally split) without adding hosts.</description></item><item><title>Learning About OpenTSDB</title><link>https://blg.robot-house.us/posts/learning-about-opentsdb/</link><pubDate>Tue, 19 Feb 2019 18:46:00 +0000</pubDate><guid>https://blg.robot-house.us/posts/learning-about-opentsdb/</guid><description>At $JOB, we needed a highly scalable metric storage system and we had a nice little Hadoop cluster I&amp;rsquo;d set up that could handle some IOPs. So we decided to test OpenTSDB and Druid.
Druid was certainly appealing:
They claimed OLAP powers Ambari had a built-in install There was a Grafana plugin The problem with Druid was simple: we couldn&amp;rsquo;t successfully query data once we got it in. Both the Grafana plugin and CLI tools could not produce readable data after weeks of effort.</description></item><item><title>Graphite and HBase</title><link>https://blg.robot-house.us/posts/graphite-and-hbase/</link><pubDate>Fri, 16 May 2014 10:23:00 +0000</pubDate><guid>https://blg.robot-house.us/posts/graphite-and-hbase/</guid><description>So this is exciting&amp;hellip;
We run Graphite at work for (relatively) real-time metric collection of our environment. That&amp;rsquo;s cool. It&amp;rsquo;s a powerful tool that allows great insight into what is actually happening in our environment.
However&amp;hellip;
Graphite relies (by default) on a flat-file, rrd-like database called whisper. That&amp;rsquo;s fine when you&amp;rsquo;re not getting that many metrics (or when you can spring for SSDs to write your data), but when you&amp;rsquo;re getting over 2.</description></item></channel></rss>