LANDOOP Blog

drive02 drive02
10 Aug 2018

Lenses 2.1

· by Stefan Bocutiu · Read in about 5 min · #Release

August just became more hot; with great pleasure we —team Landoop— announce the immediate release of Lenses 2.1. Following an ambitious road-map, this version focuses on fortifying our SQL engine’s capabilities, the all-new global streaming topology graph and an improved user experience. Lenses SQL now supports ALL data formats The Lenses SQL streaming engine for Apache Kafka can now handle any type of serialization format, including the much requested Google’s Protobuf.


10 Jul 2018

Lenses on HDInsight

· by Marios Andreopoulos · Read in about 12 min · #lenses #azure

So you gave our Lenses Box a spin and were sold immediately, now you are ready to run Lenses against your own Kafka cluster. We actually get this a lot! As it happens your cluster is on Azure, maybe even on Azure HDInsight; Microsoft after all, a few days ago, announced the general availability of Kafka 1.0 for HDInsight. As always, we’ve got you covered.

In this article we’ll go through a simple demonstration on how to setup Lenses on Azure and connect it to a HDInsight Kafka cluster, maybe even throw in a Schema Registry instance for good measure. Note that I am an Azure beginner myself, so many clouds to learn, so little time. Luckily Lenses is designed to work effortlessly with any Kafka installation. We love, support and learn from all vendors.


1 Jun 2018

Using the Lenses SQL Jdbc driver with Apache Spark

· by Stephen Samuel · Read in about 8 min · #kafka #jdbc

In our previous blog we introduced the exciting new open source JDBC driver for Apache Kafka via Lenses. In this article we’ll delve deeper and show how we can use the driver in conjunction with Apache Spark. For those who are new to Spark, Apache Spark is an in-memory distributed processing engine which supports both a programatic and SQL API. Spark will split a dataset into partitions and distribute these partitions across a cluster.


5 May 2018

Data Governance and GDPR with Apache Kafka

· by Antonios Chalkiopoulos · Read in about 7 min · #GDPR

Last December we announced our commitment to provide the necessary capabilities for data streaming systems, that will enable data-driven businesses to achieve compliance with GDPR prior to the regulation’s effective date (May 25, 2018), and this post explains how Lenses delivers by providing Data Governance capabilities and GDPR compliance by design. The immutable nature of modern high-performance distributed systems, provides a lot of competitive advantages to various industries that are interested in fast loading streams of events and apply low latency queries and scalable processing for data in motion.


1 May 2018

Using the Lenses JDBC driver with Squirrel SQL

· by Stephen Samuel · Read in about 5 min · #kafka #jdbc

One of the most well supported protocols that the JDK has introduced is the JDBC (Java Database Connectivity) interface for accessing relational databases. Over two decades since it debuted, the number of supported databases has grown to include databases that are not relational, and in some cases, not even databases. Now, through the recently released Lenses JDBC driver, Apache Kafka can be added to the list of supported technologies.


22 Apr 2018

Lenses 2.0 for Apache Kafka

· by Antonios Chalkiopoulos · Read in about 6 min · #Release

Today, we are very pleased to announce the release of Lenses v.2.0

Lenses is the streaming management platform for Apache Kafka. This release focuses on improvements based on the feedback we’ve received and introduces a ton of exciting new features.

Here’s a quick overview:

  • Lenses SQL Engine supports JDBC and introduces a Kafka JDBC driver
  • Lenses platform now ships with Lenses CLI
  • New Clients for Python & Go
  • Notification Gateway and native integration with Alert Manager
  • New Features for Data Governance and Multi-tenancy management
  • Enhanced Topology with Metrics for Processors, Connectors & Topics
  • Cloud Native capabilities


15 Mar 2018

Cassandra to Kafka - Tuning the Kafka Connector (part 2)

· by Mike Barlotta · Read in about 9 min · #Cassandra #KafkaConnect

Mike Barlotta, Agile Data Engineer at WalmartLabs introduces how Kafka Connect and Stream Reactor can be leveraged to bring data from Cassandra into Apache Kafka.



In the first part of this series (see Getting started with the Kafka Connect Cassandra Source) we looked at how to get Kafka Connect setup with the Cassandra Source connector from Landoop. We also took a look at some design considerations for the Cassandra tables. In this post we will examine some of the options we have for tuning the Cassandra Source connector.


14 Mar 2018

Getting started with the Kafka Connect Cassandra Source

· by Mike Barlotta · Read in about 10 min · #Cassandra #KafkaConnect

Mike Barlotta, Agile Data Engineer at WalmartLabs introduces how Kafka Connect and Stream Reactor can be leveraged to bring data from Cassandra into Apache Kafka.



This post will look at how to setup and tune the Cassandra Source connector that is available from Landoop. The Cassandra Source connector is used to read data from a Cassandra table, writing the contents into a Kafka topic using only a configuration file. This enables data that has been saved to be easily turned into an event stream.


1 Mar 2018

IoT for Smart Homes and trillions of messages from Kafka to Elasticsearch

· by Angelos Petheriotis · Read in about 9 min · #IoT #ElasticSearch #KafkaConnect #Kubernetes

Angelos Petheriotis, Senior Data Engineer at Centrica (Hive Home/British Gas) shares parts of their data journey, building IoT realtime data pipelines with Stream-Reactor, Kafka and Kubernetes.



Driving billions of messages per day through multiple processing pipelines requires a significant amount of processing and persisting jobs. We designed our pipelines having in mind a real time, durable and stable continuous data pipeline. In order to achieve this goal we made our services and our infrastructure as decoupled as possible.


30 Jan 2018

MQTT. Kafka. InfluxDB. SQL. IoT Harmony.

· by Stefan Bocutiu · Read in about 16 min · #Lenses #LensesHowTos #LensesSQL #Kafka #LSQL #IoT #InfluxDB #KafkaStreams #MQTT #COAP

IoT with Kafka via Lenses The rapidly growing number of interconnected devices confirms the Internet of Things (IoT) is a fast maturing technology. The digital economy has its own currency and that is data. Similar to the standard currencies, data is valuable if you can use it. The IoT is a driver for being data rich. However, having the data is not quite enough; you need to be able to analyze the data and take the appropriate action.


17 Jan 2018

Stream Reactor released for Kafka Connect 0.11 and 1.0

· by Andrew Stevenson · Read in about 3 min · #Release #StreamReactor #Connectors

Stream Reactor, the largest open-source collection of Apache Kafka connectors, has released today many new features, bug fixes and new connectors for Apache Pulsar!

  • Version 0.4.0 for Kafka Connect 0.11
  • Version 1.0.0 for Kafka Connect 1.0


11 Jan 2018

Using Lenses to easily scale SQL processors in Kubernetes - Part 4

· by Andrew Stevenson · Read in about 3 min · #LensesHowTos #Video #Kubernetes #SQL #Kafka

In this previous post we showed how to scale out Lenses SQL processors with Kafka Connect. Connect is one on three execution modes for LSQL processors via Lenses, we also have in process, mainly for developers and Kubernetes, the subject of this post.


10 Jan 2018

Lenses 1.1 Release

· by Christina Daskalaki · Read in about 5 min · #Release

We are super excited to announce the new Lenses release v1.1! Lenses is a streaming platform for Apache Kafka which supports the core elements of Kafka, vital enterprise features and a rich web interface to simplify your Kafka development and operations. Lenses also ships with a free single broker development environment which provides a pre-setup Kafka environment with connectors and examples for your local development. Since November’s release, Lenses has been widely adopted and we would like to thank you all for your valuable feedback which we have taken into account as part of this release.


19 Dec 2017

JavaScript lib for Kafka development

· by Sebastian Dragomir · Read in about 6 min · #Release

More and more companies are embracing the data streaming revolution in order to get a better insight of their business and thus drive growth upwards. With Apache Kafka getting a lot of momentum in the space, it is normal to want to connect a Javascript UI application to it. This Javascript library build on redux brings the best practises of real time data streaming to the browser and provides a new way for building solutions to get and send data on Apache Kafka.


7 Dec 2017

Kafka stream processing via Lenses SQL, scale with Kafka - part 3

· by Andrew Stevenson · Read in about 2 min · #LensesHowTos #Video

As mentioned in previous post, Lenses SQL leverages Kafka Streams to process data and currently provides 3 execution modes to run Lenses SQL processors. In this video we demonstrate how to scale out using CONNECT mode as well as how to manage the Lenses SQL processors via the Lenses web interface or the CLI tool.


1 Dec 2017

Apache Kafka and GDPR Compliance

· by Antonios Chalkiopoulos · Read in about 5 min · #DataProtection #GDPR

General Data Protection Regulation Read the May 2018 updated article GDPR - Data Governance with Apache Kafka and Lenses 2.0 GDPR is an important piece of legislation designed to strengthen and unify data protection laws for all individuals within the European Union. The regulations becomes effective and enforceable on the 25th May 2018. Our commitment is to provide the necessary capabilities in data streaming systems, to allow your data-driven business to achieve compliance with GDPR prior to the regulation’s effective date.


21 Nov 2017

Kafka stream processing via SQL - part 2

· by Antonios Chalkiopoulos · Read in about 7 min · #LensesHowTos

In this post we are going to see how you can leverage Lenses and SQL (Lenses SQL - our own SQL layer solution for Apache Kafka called LSQL) to create, execute and monitor Kafka Streams application defined with SQL. If you’ve worked with data before, a lot of time is allocated to extracting and massaging the data from various sources, and enhancing them into the required format. Lenses SQL Engine for Apache Kafka makes your ETL challenges a quick and integrated experience.


12 Nov 2017

How to explore data in Kafka topics with Lenses - part 1

· by Christina Daskalaki · Read in about 5 min · #LensesHowTos

In this post we are going to see how Lenses can help you explore data in Kafka. Lenses comes with a powerful user interface for Kafka to explore historical or in motion data, for which you can run Lenses SQL Engine queries. This helps to quickly access data for debugging, analyzing or reporting but at the same time is not requiring being a developer to do so. In addition, Lenses comes with a set of REST and Web Socket endpoints that makes integration with your Kafka data simple.


15 Oct 2017

Apache Kafka Streaming, count on Lenses SQL

· by Stefan Bocutiu · Read in about 5 min · #LensesHowTos #LensesSQL

In this brief entry we will discuss how count aggregation can be coded faster with Lenses SQL. Count aggregations is a very common scenario in stream processing. Some common use cases include aggregated time reports of transaction counts for a payment provider, views of products for an e-commerce site, how many customers are viewing a hotel and many more. In this article we will see how Lenses allows you to run these aggregations leveraging Lenses SQL engine.


11 Oct 2017

Describe and execute Kafka stream topologies with Lenses SQL

· by Stefan Bocutiu · Read in about 15 min · #LensesHowTos #LensesSQL

Streaming Topologies out of the box! Lenses SQL in action Lenses SQL for Analyze, Process, Connect Lenses SQL supports the 3 major layers of your data streaming pipeline: Analyze: Run ad-hoc queries over Kafka topic in real time or history. Browsing your Kafka topics has never been easier and more powerful. Process: on top of Kafka Streams API in order to run comprehensive and production-quality streaming analytics. Connect: We build all our connectors by bringing SQL capability at the ingestion process.


19 Sep 2017

Athens Big Data meetup - Landoop (Sep 17)

· by Antonios Chalkiopoulos · Read in about 1 min · #Meetup

Excited to have attended the Athens Big Data meetup in mid September to present our open source contributions to the streaming technologies and in particular around Apache Kafka®.


29 Aug 2017

Lenses Announced for Apache Kafka™ !

· by Christina Daskalaki · Read in about 3 min · #Announcement

We are happy to announce our unified platform Lenses that enables us to zoom into Apache Kafka®


17 Jul 2017

Fast Data CSD 3.2 release

· by Marios Andreopoulos · Read in about 3 min · #Release

Fast Data 3.2 is officially out! This release cycle took longer than usual but it brings many changes that will provide you with a more streamlined Kafka experience and let us build, test and enhance future releases quicker and with more confidence. The release has been available for our clients since last month.

Fast Data is our solution for installing and managing a modern Kafka stack through Cloudera Manager. Check here for an overview and request a trial today!

If you already use our CSD, read our documentation for instructions on how to upgrade without downtime. We are always available to help and can arrange for an engineer to walk you through.


19 Jun 2017

Kafka security via data encryption

· by Antonios Chalkiopoulos · Read in about 3 min · #DataProtection

Secure cluster, but what happens with your data?

When it comes to security, Apache Kafka as every distributed system provides the mechanisms to transfer data securely across the components being involved. Depending on your set up this might involve different services such as Kerberos, relying on multiple TLS certificates and advanced ACL setup in brokers and Zookeeper. In many cases, with encryption features enabled, performance is also taking a penalty hit.


12 May 2017

Kafka Topics UI (rest proxy v2)

· by Christina Daskalaki · Read in about 1 min · #Release


The new version of Kafka Topics UI is now available!

The new version implements the Rest proxy v2 API so make sure you upgrade το the right version of rest proxy.


6 May 2017

How British Gas is streaming 4 Billion messages with our connectors to enable Connected Homes!

· by Christina Daskalaki · Read in about 1 min · #UseCases

This presentation is by Angelos Petheriotis, senior engineer at HiveHome British Gas, at Apache Kafka meetup. Angelos presented the team’s journey to 50K msg/s from IoT Devices featuring our DM Stream-reactor connectors, how they use Kafka Connect Query Language and Landoop Web Tools. Enjoy! Streaming 4 billion Messages per day. Lessons Learned. from Angelos Petheriotis


3 May 2017

Fast Avro Write

· by Stefan Bocutiu · Read in about 5 min · #Techies

This article presents how Avro lib writes to files and how we can achieve significant performance improvements by parallelizing the write. A (JVM) library has been implemented and is available on Github fast-avro-write The reason we proceeded with this implementation was a project that required writing multiple Μillions of Avro messages from Kafka onto a star DW (data warehouse) in HIVE (HDFS). You might have heard about (or even dealt with) the challenges of working with HDFS.


4 Apr 2017

Kafka Connect Pipelines, sink to Elasticsearch

· by Christina Daskalaki · Read in about 9 min · #Tutorial

Introduction

In this mini tutorial we will explore how to create a Kafka Connect Pipeline using the Kafka Development Environment (fast-data-dev) in order to move real time telemetry data into Elasticsearch and finally visualize the positions in a Kibana Tile Map by writing zero code…!


20 Feb 2017

Kafka connect for FTP data

· by Antonios Chalkiopoulos · Read in about 8 min · #KafkaConnect

An FTP server, together with a pair of credentials is a common pattern, on how data providers expose data as a service. In this article we are going to implement custom file transformers to efficiently load files over FTP and using Kafka Connect convert them to meaningful events in Avro format. Depending on data subscriptions we might get access to FTP locations with files updated daily , weekly or monthly. File structures might be positional, csv, json , xml or even binary.


28 Jan 2017

From MQTT to Kafka with Connect and Stream Reactor

· by Marios Andreopoulos · Read in about 8 min · #KafkaConnect

MQTT stands for MQ Telemetry Transport. It is a lightweight messaging protocol, designed for embedded hardware, low-power or limited-network applications and microcontrollers with limited RAM and/or CPU. It is a protocol that drives the IoT expansion. On the other hand, large numbers of small devices that produce frequent readings, lead to big data and the need for analysis in both time and space domain (spatial-temporal analysis). Kafka can be the highway that connects your IoT with your backend analytics and persistence.


18 Jan 2017

Apache Kafka London Meetup - by Landoop

· by Antonios Chalkiopoulos · Read in about 1 min · #Meetup

How to simplify your ETL process using Kafka Connect for (E) and (L). Introducing KCQL - the Kafka Connect Query Language for fast-data pipelines. Using KCQL to set up Kafka Connectors for popular in-memory and analytical systems (live demos) such as HazelCast, Redis and InfluxDB. Use fast-data-dev docker for your kafka development environment. Enhancing your existing Cloudera (Hadoop) clusters with fast-data capabilities. Demos: http://schema-registry-ui.landoop.com http://kafka-topics-ui.landoop.com http://kafka-connect-ui.landoop.com https://fast-data-dev.demo.landoop.com/ Code https://github.com/landoop/ Connectors


1 Dec 2016

Time-Series with Kafka, Kafka Connect & InfluxDB

· by Christina Daskalaki · Read in about 6 min · #Tutorials

Time-series datastores are of particular interest these days and influxDB is a popular open source distributed time-series database. In this tutorial we will integrate Kafka with InfluxDB using Kafka Connect and implement a Scala avro message producer to test the setup. The steps we are going to follow are: Setup a docker development environment Run an InfluxDB Sink Kafka Connector Create a Kafka Avro producer in Scala (use the schema registry) Generate some messages in Kafka Finally, we will verify the data in influxDB and visualise them in Chronograph.


21 Aug 2016

Coyote Testing Tool

· by Marios Andreopoulos · Read in about 7 min · #Release

A few days ago we open source’d Coyote, a tool we created in order to automate testing of our Landoop Boxes, which features a large range of environments for Big Data and Fast Data (see Kafka). Coyote does one simple thing: it takes a .yml file with a list of commands to setup, run and check their exit code and/or output. It has some other functionality too, but its essence is this.


8 Aug 2016

Confluent Platform 3.0.0 CSD

· by Marios Andreopoulos · Read in about 2 min · #Release

This is now in General Availability. See it here and request a trial today!

Today we release our first beta CSD for Confluent Platform 3.0.0. It is robust enough to use in production; but we want to add at least some small touches before the final release which we expect to be fully compatible with the beta —drop-in replacement and upgrade.


7 Aug 2016

Kafka Topics UI

· by Antonios Chalkiopoulos · Read in about 3 min · #Release

Hey,
check out the new version for Kafka Topic UI here.

If you are looking for an all in one UI for Kafka check out Lenses for topic data, schemas, connectors, processors, consumers UIs and Lenses SQL engine support!

Kafka is now the de-facto platform for streaming architectures, and it’s eco-system is maturing, but is not just yet as Enterprise Ready as many people in Big | Fast Data would like it to be. Landoop is a London based start-up that wants to drive Kafka faster to the future, and thus..

We are announcing the kafka-topics-ui a User Interface that allows browsing data from Kafka Topics and a lot more


6 Aug 2016

Schema Registry UI for Kafka

· by Antonios Chalkiopoulos · Read in about 3 min · #Release

Hey, If you are looking for an all in one UI for Kafka check out Lenses for topic data, schemas, connectors, processors, consumers UIs and Lenses SQL engine support!

If you are looking for a safe way to interchange messages while using a fast streaming architecture such as Kafka, you need to look no further than Confluent’s schema-registry. This simple and state-less micro-service, uses the _schemas topic to hold schema versions, can run as a single-master multiple-slave architecture and supports multi data-center deployments.

We are happy to announce a UI, the schema-registry-ui a fully-featured tool for your underlying schema registry that allows visualization and exploration of registered schemas and a lot more…


14 Jul 2016

Our Argos and Accenture presentation on Big and Fast Data

· by Antonios Chalkiopoulos · Read in about 1 min · #Meetup

We want to thank @Argos - the third largest retailer in UK - for inviting Landoop and @Accenture for hosting our presentation in one of the most beautiful theaters in the world, the IMAX theater in SCIENCE MUSEUM, London.

View our presentation on Big to Fast Data. How #kafka and #kafka-connect can redefine you ETL and a bit about #stream-processing


8 Jul 2016

Confluent Platform 2.0.1 on Cloudera (CSD)

· by Marios Andreopoulos · Read in about 5 min · #Release

Important. Our Confluent CSD is deprecated and replaced by our most complete yet solution for a managed Kafka stack through Cloudera Manager, including monitoring, alerts and our exclusive UIs. See it here and request a trial today!

We are happy to announce the first version of our Confluent CSD.

Utilizing Landoop’s Confluent CSD you can create a Kafka Cluster with support services such as REST Proxy, Schema Registry and Kafka Connect in a few clicks.


7 Feb 2016

Ansible Nginx Let's Encrypt Automation

· by Marios Andreopoulos · Read in about 5 min · #Techies

Automatic SSL certificate issuance and renew with Ansible and Let’s Encrypt Here on Landoop we prototype fast and new (sub)domains are frequently added to complement our back and front-end services. Since the beginning our specifications included “ssl everywhere”. The journey into providing fully secure and encrypted services is a long one; hence we need an adventure in the SSL land. The tools, the needs. We use ansible to manage our servers.


Subscribe to get updates!


2 Minute Overview


Discover awesome features


Community


Join us at Landoop Community


Resources


Repos, Docs, Trainings, Tutorials


Free Download


ALL-IN-ONE free for developers!