Packages

  • package root
    Definition Classes
    root
  • package com
    Definition Classes
    root
  • package adform
    Definition Classes
    com
  • package streamloader

    The entry point of the stream loader library is the StreamLoader class, which requires a KafkaSource and a Sink.

    The entry point of the stream loader library is the StreamLoader class, which requires a KafkaSource and a Sink. Once started it will subscribe to the provided topics and will start polling and sinking records. The sink has to be able to persist records and to look up committed offsets (technically this is optional, but without it there would be no way to provide any delivery guarantees). A large class of sinks are batch based, implemented as RecordBatchingSink. This sink accumulate batches of records using some RecordBatcher and once ready, stores them to some underlying RecordBatchStorage. A common type of batch is file based, i.e. a batcher might write records to a temporary file and once the file is full the sink commits the file to some underlying storage, such as a database or a distributed file system like HDFS.

    A sketch of the class hierarchy illustrating the main classes and interfaces can be seen below.



    For concrete storage implementations see the clickhouse, hadoop, s3 and vertica packages. They also contain more file builder implementations than just the CsvFileBuilder included in the core library.

    Definition Classes
    adform
  • package clickhouse
  • package hadoop
  • package model
  • package s3
  • package sink
  • package source
  • package util
  • package vertica
  • BuildInfo
  • StreamLoader
p

com.adform

streamloader

package streamloader

The entry point of the stream loader library is the StreamLoader class, which requires a KafkaSource and a Sink. Once started it will subscribe to the provided topics and will start polling and sinking records. The sink has to be able to persist records and to look up committed offsets (technically this is optional, but without it there would be no way to provide any delivery guarantees). A large class of sinks are batch based, implemented as RecordBatchingSink. This sink accumulate batches of records using some RecordBatcher and once ready, stores them to some underlying RecordBatchStorage. A common type of batch is file based, i.e. a batcher might write records to a temporary file and once the file is full the sink commits the file to some underlying storage, such as a database or a distributed file system like HDFS.

A sketch of the class hierarchy illustrating the main classes and interfaces can be seen below.



For concrete storage implementations see the clickhouse, hadoop, s3 and vertica packages. They also contain more file builder implementations than just the CsvFileBuilder included in the core library.

Linear Supertypes
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. streamloader
  2. AnyRef
  3. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. Protected

Package Members

  1. package clickhouse
  2. package hadoop
  3. package model
  4. package s3
  5. package sink
  6. package source
  7. package util
  8. package vertica

Type Members

  1. class StreamLoader extends Logging with Metrics

    The main stream loading class, given a KafkaSource and some Sink continuously polls the source and sinks records to the sink.

    The main stream loading class, given a KafkaSource and some Sink continuously polls the source and sinks records to the sink.

    Runs in the active thread and blocks it. When running multiple instances in multiple threads you must provide separate sources and sinks.

Value Members

  1. case object BuildInfo extends Product with Serializable

    This object was generated by sbt-buildinfo.

Inherited from AnyRef

Inherited from Any

Ungrouped