Prepare Kafka for IoT with DCOS

Prepare Kafka for IoT with DCOS

In the blog post Install dev environment for Kafka, Nodejs and node-red I described to install a kafka – webstorm environment on a virtual machine. For a productive use, we need high availability kafka installation. The idea is to use an orchestration framework like DCOS. It is based on the documentation of mesosphere https://docs.mesosphere.com/1.8/usage/tutorials/iot_pipeline/.

Continue reading

Install dev environment for Kafka, Nodejs and node-red

To prepare a nodejs docker image with a javascript application for DCOS I need a small development environment. I plan to install on a linux based system kafka, nodejs and node-red. The idea is, to have a small environment to develop a web interface for websocket connections.

Continue reading

Cloud Control 13.2 and SOA Suite 12.2.1.1

Cloud Control 13.2 and SOA Suite 12.2.1.1

The intention is to monitor a SOA/BPM Cluster installation with 2 managed servers on two hosts with Cloud Control 13.2. There is a different user for weblogic and cloud control agent.

First step is to install the agents on 2 servers. Each server has one managed server. The easiest installation for agents is to install it with AgentPull Scripts.

Here is a short script, for further information about agent install see oracle documentation.

Continue reading

how oracle exadata write-back flash cache (wbfc) turns your life

how oracle exadata write-back flash cache (wbfc) turns your life

Write-Back flash cache is a feature of oracle Exadata since version 11.2.3.2 and is  available since 2012. By default it’s turned off. Do I need wbfc?

specifications

When you look at performance benchmark of oracle Exadata disk performance it decrease from 10.800 IOPS (X2-2 Quarter, HP Disks) to 7.800 IOPS (X6-2 Quarter, HC Disks). The size increased from 600GB to 8TB. If you want to write random IO to X6 without write-back flash cache it can be a nightmare in opposite to a X2-2 machine.

Continue reading

ACFS Replication between two Exadata’s

In this blog I describe the configuration of ACFS between two Exadata. ACFS is a cluster file system based on Oracle ASM (Automatic Storage Management). It is available at Oracle Database Cluster with ASM and also on Exadata installations. I found some documentation, but only few details are available. In my blog I added information about monitoring and it is optimized for use with Exadata. In my blog the installation is based on X5 Exadata hardware and Oracle 12C (12.1.0.2) grid and database software.
Continue reading

Data Transformation in Big Data Appliance or Oracle Exadata – a comparison

In a data ware house environment new data is often read out of csv files. This csv files are connected as an external table to the database. In this scenario you need a high available file system, where the text files are stored. The idea is now to use hdfs – hadoop file system – as a high available, high-performance filesystem. Hdfs is than an high performance, high available filesystem and in additional a database where first transformation can be executed without stress the main database.

In the blog I show the way to integrate this data in the data ware house and perform some easy transformations like generate keys. I use  BigDataSQL external tables to connect. On the one hand I create everything in the data base on the other hand I create the keys in an Hive/hadoop environment. The keys are based on Md5 hashes as described in the last blog.
Continue reading