Bokuto ni yoru kendo kihon waza keiko hou

Bokuto ni yoru kendo kihon waza keiko hou

基本一 一本打ち Kihon ichi, ippon uchi
面、小手、胴、突き men, kote, do, tsuki

基本二 2段技 Kihon ni, nidan waza
小手面 kote men

基本三 払い技 Kihon san, harai waza
払い面 harai men

基本四 引き技 Kihon yon, hiki waza
面引き胴 men hiki do

基本五 抜き技 Kihon go, nuki waza
面抜き胴 men nuki do

基本六 すりあげ技 Kihon roku, suri age waza
小手すりあげ面 kote suri age men

基本七 出鼻技 Kihon nana, debana waza
面出鼻小手 men debana kote

基本八 返し技 Kihon hachi, kaeshi waza
面返し胴 men kaeshi do

基本九 打ち落とし Kihon kyu, uchi otoshi waza
胴打ち落とし面 do uchi otoshi men

comparison between Apache Kafka and RabbitMQ

Apache Kafka RabbitMQ
Creation year 2011 2007
License Apache (Open source) Mozilla Public License (Open source)
Programming language Scala Erlang
AMQP compliant No Yes
Officially supported clients in JAVA JAVA, .NET/C#, Erlang

Table 1 – General information comparison between Apache Kafka and RabbitMQ

Continue reading “comparison between Apache Kafka and RabbitMQ”

Overall structure of RabbitMQ

RabbitMQ is an open source message broker middleware created in 2007 and that is now managed by GoPivotal.

Most of the operations are performed in memory. RabbitMQ is not “disk-oriented”:
messages are received by brokers via an exchange (i.e. a logical entry point that will decide based on some criteria in which queue(s) the broker should place a message) and then pushed to the registered consumers. The broker pushes randomly queued messages toward the consumers. They thus receive unordered messages, and do not need to remember anything about the queue state (as messages are unordered and pushed by the brokers. They do not and cannot fetch specific messages on their own). Messages are paged out to disc only if there is no more memory available, or if they are explicitly told to be stored.

Continue reading “Overall structure of RabbitMQ”

Apache JMeter Introduction on MAC

Commend Version

You can install by HomeBrew

  1. Install HomeBrew

To install HomeBrew:

/usr/bin/ruby -e "$(curl -fsSL"

It should take only a couple of minutes. Before installing JMeter, let’s now update HomeBrew package definitions:

brew update

Continue reading “Apache JMeter Introduction on MAC”

(Hadoop) analysis of Hive Meta Store Entity Using A Hook Function

Currently, I was working to implement a project to save metadata of Hive Program. Basically, I keep A SQL database to save and update the metadata of every execution of HQL sentence with a internal hook.

Basically, There are two important Set of Entities Class:
Set inputs and Set outputs

In this article, I only introduce the data inside of those Entities above, the exact structure of Hive program will be introduced in another article.

In the source code of org.apache.hadoop.hive.ql.plan.HiveOperation, you can found tens of different hive operation. For our goal, a metadata store system, I only care about those operation related to metadata.



input: null, or location if set location while create table.
output: new table, current database
log: operation is CREATETABLE,inputs :[],outputs:[db@tml_2, database:db]


input: deleted table
output: deleted table
log: operation is DROPTABLE,inputs :[db@tml_1],outputs:[db@tml_1]


input: old table
output: old table, new table
log: operation is ALTERTABLE_RENAME,inputs :[db@tml_2],outputs:[db@tml_2, db@tml_3]


input: null
output: new table
log: operation is ALTERTABLE_RENAMECOL,inputs :[],outputs:[db@tml_3]


input: null
output: new table
log: operation is ALTERTABLE_RENAMECOL,inputs :[],outputs:[db@tml_3]


input: table, old partition
output: old partition, new partition
log: operation is ALTERTABLE_RENAMEPART,inputs :[db@tml_part, ks_xs@tml_part@dt=2008-08-08/hour=14],outputs:[db@tml_part@dt=2008-08-08/hour=14, db@tml_part@dt=2008-08-08/hour=15]

    input: partition
    output: location, partition
    log: operation is ALTERPARTITION_LOCATION,inputs :[db@tml_part, db@tml_part@dt=2008-08-08/hour=15],outputs:[viewfs://hadoop-lt-cluster/home/dp/data/userprofile/db.db/tml_part/dt, db@tml_part@dt=2008-08-08/hour=15]


In the org.apache.hadoop.hive.ql.hooks.ENtity, You can found all the Type of Entity.

   * The type of the entity.
  public static enum Type {

What’s strange is that there is no COLUMN in them. So when we try to catch the operation of add/rename/replace columns, we have to get the data from their parent table.

Besides, we can get meta data easily with specific type.

Some experience about upload/download files on HDFS by JAVA Spring framework

  1. Using Stream method to transfer data

At first, I tried to use a easy way to transfer files, read them into memory as binary arrays, and then upload to HDFS. However, I met problem with java.lang.OutOfMemoryError: Java heap space. The reason is that I use some redundant function like File.getBytes(), which occupied too much memory.

Then I fixed the problem by using double stream. From Client(browser) to Web Server, and from Web Server to HDFS Server.

  • From Browser To Web Server

We can use the normal way, MultipartFile file to upload file.

The file contents are either stored in memory or temporarily on disk. In either case, the user is responsible for copying file contents to a session-level or persistent store as and if desired. The temporary storage will be cleared at the end of request processing.

Developer can config the threshold to decide if the temporary file is saved in memory or in disk by modify proterty file like following.


So that we can support high level concurrent visit.

  • From Web Server to HDFS Server

After initialized FileSystem by Config Class, The process is easily.

FSDataOutputStream out = fs.create(path, true);
IOUtils.copyBytes(file.getInputStream(), out, 1024, true);
  • From HDFS to Web Server

While downloading file, Hadoop API offers powerful API, only if we input the file path, we can create a InputStream.

FSDataInputStream in =;
  • From Web Server to Browser

Normally, in Spring framework, developer can use Resource Instance to transer file. And kindly, Resource offers a function InputStreamResource(InputStream), so that we can use it to download files from HDFS to browser directly.

  1. File Name Code Problem

This problem is exactly for Chinese developer. Chinese characters will be grabled in Http header. The solution is that change charset to “ISO8859-1”

String fileName = new String(fileName.getBytes(), "ISO-8859-1");

(MongoDB) How to use –fork On Windows for multiple servers

Although for most of developers, deploying database On Windows is not advised. For beginners like me, always need do some practise on Windows.

Currently I am studying replica sets, which asks me to run multiple servers at the same time with following commands.

Continue reading “(MongoDB) How to use –fork On Windows for multiple servers”