Manado, Indonesia. 95252
(+62) 823-9602-9583

Author: Bayu Dwiyan Satria

Software Engineer | DevOps Engineer

kubernetes

Introduction to Basic Concept of Kubernetes

Have you heard Kubernetes? You must be interested in that topic. That’s why you open this article. This article is about the basic concept of Kubernetes and how to use it.

hadoop

Setup And Configure Single-Node Hadoop Installation

Single node cluster means only one DataNode running and setting up all the NameNode, DataNode, ResourceManager and NodeManager on a single machine. It can easily and efficiently the sequential workflow in a smaller environment as compared to large environments which contains terabytes of data distributed across hundreds of machines.

laptop

How to Reduce Technology Distractions for a More Productive Workplace

Human resources make up 60% of the entire operating costs in an organization. Therefore, managing the workforce holds the for management to maintain and increase their profits. While IT was expected to come as a harbinger of change, that can elevate workplace productivity to uncharted heights, new evidence suggests that this is clearly not the case. Instead, the productivity of employees is experiencing a gradual decline due to the complexity of current IT strategies and policies.
#IT #Blog #Technology

VPN Users

Free VPN, is it Safe

VPN is a connection between networks that is private, carried out through public networks and allows the exchange of resources privately.

web-crawling

Best Java Web Crawler Implementation

Implementing a Java web crawler is a fun and challenging task often given in university programming classes. You may also actually need a Java web crawler in your own applications from time to time. You can also learn a lot about Java networking and multi-threading while implementing a Java web crawler. This tutorial will go through the challenges and design decisions you face when implementing a Java web crawler.

spark-vs-hadoop

What is differences about Apache Hadoop vs Apache Spark

What is Big Data? What size of Data is considered to be big and will be termed as Big Data? We have many relative assumptions for the term Big Data. It is possible that, the amount of data say 50 terabytes can be considered as Big Data for Startup’s but it may not be Big Data for the companies like Google and Facebook. It is because they have infrastructure to store and process this vast amount of data. Apache Hadoop and Apache Spark are both Big Data analytics frameworks they provide some of the most popular tools used to carry out common Big Data-related tasks.

hadoop

Apache Hadoop Introduction

The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using simple programming models. It is designed to scale up from single servers to thousands of machines, each offering local computation and storage.

Rather than rely on hardware to deliver high-availability, the library itself is designed to detect and handle failures at the application layer. So delivering a highly-available service on top of a cluster of computers, each of which may be prone to failures.

Featured

Apache Spark Parallel Program Flows

Apache Spark Flows – Apache Spark consists of several purpose-built components as we have discuss at the introduction of apache spark. Let’s see what a typical Spark program looks like. Imagine that a 300 MB log file is stored in a three-node HDFS cluster. Hadoop File System (HDFS) automatically splits the file into 128 MB parts and places each part on a separate node of the cluster.

apache_spark_component

Apache Spark Component Parallel Processing

Apache Spark consists of several purpose-built components as we have discuss at the introduction of apache spark. Apache spark provides high-level APIs in Java, Scala, Python and R, and an optimized engine that supports general execution graphs. It also supports a rich set of higher-level tools. These are Apache Spark Component : Spark Core, Spark SQL, Spark Streaming, Spark GraphX, Spark MLlib.

These components make Spark a feature-packed unifying platform: it can be used for many tasks that previously had to be accomplished with several different frameworks. A brief description of each Apache Spark component follows.

Apache_hadoop

Setup And Configure Cluster Node Hadoop Installation

This describes how to setup and configure a cluster-node Hadoop installation so that you can quickly perform simple operations using Hadoop MapReduce and the Hadoop Distributed File System (HDFS).