Selfuel Docs
  • Welcome to Selfuel Platform
    • Features
    • Capabilities
    • Target Audience
    • $150 Free Trial
  • Registration and Login
  • Platform UI
  • Stream Processing with Cortex
    • Cortex Quickstart Guide
    • Cortex Elements
      • Streams
      • Attributes
      • Mappings
        • 🚧Source Mapping Types
        • 🚧Sink Mapping Types
      • Node and Application Healthchecks
      • Nodes
        • Node Preview
        • Node Connectivites
        • Node Units
      • Expression Builder
        • 🚧Built-in Functions
      • Windows
        • Cron Window
        • Delay Window
        • Unique Event Window
        • First Event Window
        • Sliding Event Count Window
        • Tumbling Event Count Window
        • Session Window
        • Tumbling Event Sort Window
        • Sliding Time Window
        • Tumbling Time Window
        • Sliding Time and Event Count Window
      • Store and Cache
        • RDBMS
        • MongoDB
        • Redis
        • Elasticsearch
    • Applications
      • Applications Page
      • Creating Applications using Canvas
      • Connector Nodes Cluster
        • Source Nodes
          • CDC Source
          • Email Source
          • HTTP Source
          • HTTP Call Response Source
          • HTTP Service Source
          • Kafka Source
          • RabbitMQ Source
          • gRPC Source
          • JMS Source
          • Kafka Multi DC Source
          • JMS Source
          • AWS S3 Source
          • Google Pub-sub Source
          • AWS SQS Source
          • MQTT Source
          • Google Cloud Storage Source
          • HTTP SSE Source
          • WebSubHub Source
        • Sink Nodes
          • Email Sink
          • HTTP Sink
          • HTTP Service Response Sink
          • HTTP Call Sink
          • Kafka Sink
          • RabbitMQ Sink
          • gRPC Sink
          • JMS Sink
          • Kafka Multi DC Sink
          • AWS S3 Sink
          • Google Pub-sub Sink
          • AWS SQS Sink
          • MQTT Sink
          • Google Cloud Storage Sink
          • HTTP SSE Sink
          • WebSubHub Sink
      • Processing Nodes Cluster
        • Query
        • Join
        • Pattern
        • Sequence
        • Processor
        • 🚧On-demand Query
      • Buffer Nodes Cluster
        • Stream
        • Table
        • Window
        • Aggregation
        • Trigger
    • Run Applications
      • Run Applications Using Runners
      • Update Running Applications
      • Application Versioning
  • Data Integration with Nexus
    • Nexus Quickstart Guide
    • Nexus Elements
      • Concept
        • Config
        • Schema Feature
        • Speed Control
      • Connectors
        • Source
          • Source Connector Features
          • Source Common Options
          • AmazonDynamoDB
          • AmazonSqs
          • Cassandra
          • Clickhouse
          • CosFile
          • DB2
          • Doris
          • Easysearch
          • Elasticsearch
          • FakeSource
          • FtpFile
          • Github
          • Gitlab
          • GoogleSheets
          • Greenplum
          • Hbase
          • HdfsFile
          • Hive
          • HiveJdbc
          • Http
          • Apache Iceberg
          • InfluxDB
          • IoTDB
          • JDBC
          • Jira
          • Kingbase
          • Klaviyo
          • Kudu
          • Lemlist
          • Maxcompute
          • Milvus
          • MongoDB CDC
          • MongoDB
          • My Hours
          • MySQL CDC
          • MySQL
          • Neo4j
          • Notion
          • ObsFile
          • OceanBase
          • OneSignal
          • OpenMldb
          • Oracle CDC
          • Oracle
          • OssFile
          • OssJindoFile
          • Paimon
          • Persistiq
          • Phoenix
          • PostgreSQL CDC
          • PostgreSQL
          • Apache Pulsar
          • Rabbitmq
          • Redis
          • Redshift
          • RocketMQ
          • S3File
          • SftpFile
          • Sls
          • Snowflake
          • Socket
          • SQL Server CDC
          • SQL Server
          • StarRocks
          • TDengine
          • Vertica
          • Web3j
          • Kafka
        • Sink
          • Sink Connector Features
          • Sink Common Options
          • Activemq
          • AmazonDynamoDB
          • AmazonSqs
          • Assert
          • Cassandra
          • Clickhouse
          • ClickhouseFile
          • CosFile
          • DB2
          • DataHub
          • DingTalk
          • Doris
          • Druid
          • INFINI Easysearch
          • Elasticsearch
          • Email
          • Enterprise WeChat
          • Feishu
          • FtpFile
          • GoogleFirestore
          • Greenplum
          • Hbase
          • HdfsFile
          • Hive
          • Http
          • Hudi
          • Apache Iceberg
          • InfluxDB
          • IoTDB
          • JDBC
          • Kafka
          • Kingbase
          • Kudu
          • Maxcompute
          • Milvus
          • MongoDB
          • MySQL
          • Neo4j
          • ObsFile
          • OceanBase
          • Oracle
          • OssFile
          • OssJindoFile
          • Paimon
          • Phoenix
          • PostgreSql
          • Pulsar
          • Rabbitmq
          • Redis
          • Redshift
          • RocketMQ
          • S3Redshift
          • S3File
          • SelectDB Cloud
          • Sentry
          • SftpFile
          • Slack
          • Snowflake
          • Socket
          • SQL Server
          • StarRocks
          • TDengine
          • Tablestore
          • Vertica
        • Formats
          • Avro format
          • Canal Format
          • CDC Compatible Debezium-json
          • Debezium Format
          • Kafka source compatible kafka-connect-json
          • MaxWell Format
          • Ogg Format
        • Error Quick Reference Manual
      • Transform
        • Transform Common Options
        • Copy
        • FieldMapper
        • FilterRowKind
        • Filter
        • JsonPath
        • LLM
        • Replace
        • Split
        • SQL Functions
        • SQL
    • Integrations
      • Integrations Page
      • Creating Integrations Using Json
    • Run Integrations
      • Run Integrations Using Runners
      • Integration Versioning
  • Batch Processing/Storage with Maxim
    • Maxim Quickstart Guide
    • Maxim Elements
    • Queries
    • Run Queries
  • Orchestration with Routines
    • Routines Quickstart Guide
    • Routines Elements
    • Routines
    • Run Routines
  • Runners
    • Runners Page
    • Create a Runner to Run Applications
  • Security
    • Vaults
      • Vaults Page
      • Create Vaults
        • Runner-level Vaults
        • Application-level Vaults
      • Edit and Delete Vaults
      • 🚧Utilizing Vaults in Applications and Runners
    • Certificates
      • Certificates Page
      • 🚧Utilizing Certificates in Applications
      • 🟨Setting Up Security Settings
  • Monitoring Performance
    • Dashboard
    • Application Details
    • Runner Details
  • Logging
    • Log Types
  • Cost Management
    • SaaS
      • Pay-as-you-go
        • Hard Budget Cap
        • Soft Budget Cap
      • Subscriptions
    • On-prem
  • Organization Settings
    • General
    • Access Controls
      • User Roles and Privileges
    • Current Costs
    • Billing Addresses
    • Payment Accounts
    • Subscriptions
    • Pricing
    • Invoicing
  • User Settings
  • Troubleshooting
  • FAQs
Powered by GitBook
On this page
  • Step 4 - Kafka Connection Settings
  • Kafka Bootstrap Servers
  • Topic List
  • Kafka Source Group ID
  • Step 5 - Threading and Partition Management
  • Threading Options
  • Partition Numbers
  • Step 6 - Event Handling
  • Sequence Tracking
  • Binary Message Reception
  • Topic Offset Mapping
  • Step 7 - Offset Commit Settings
  • Enable Offset Commiting
  • Auto Commit Interval
  • Enable Asynchronous Commiting
  • Step 8 - Additional Settings
  • Additional Configurations
  • Step 9 - Preview
  1. Stream Processing with Cortex
  2. Applications
  3. Connector Nodes Cluster
  4. Source Nodes

Kafka Source

PreviousHTTP Service SourceNextRabbitMQ Source

Last updated 1 year ago

Step 4 - Kafka Connection Settings

Kafka Bootstrap Servers

Indicates a list of Kafka servers the Kafka source should connect to. This list should be provided as comma-separated values

e.g. localhost:9092, localhost:9093

Default Value
Possible Data Type

STRING

Topic List

Indicates the list of topics the source should subscribe to. This list should be provided as comma-separated values.

e.g. topic_one, topic_two

Default Value
Possible Data Type

STRING

Kafka Source Group ID

Identifies the Kafka source group through an ID. The Group ID ensures that sources having the same topic and partition within the same group don’t receive duplicate events.

Default Value
Possible Data Type

STRING

Step 5 - Threading and Partition Management

Threading Options

Specifies if the Kafka source should operate on a single thread or multiple threads. Possible values are

  • Single Thread: o run the Kafka source on a single thread.

  • Topic-wise: To use a separate thread for each topic.

  • Partition-wise: To use a separate thread for each partition.

Default Value
Possible Data Type

Partition Numbers

Indicates a list of the partition numbers for the chosen topic. This list should be provided as comma-separated values.

e.g. 0, 2, 4

Default Value
Possible Data Type

0

STRING

Step 6 - Event Handling

Sequence Tracking

When set to ON, it takes into account the sequence of events received through the source. Each event must include a sequence number to indicate its order.

Default Value
Possible Data Type

OFF

STRING

Binary Message Reception

Must be set to ON to enable the Kafka source to receive binary events.

Default Value
Possible Data Type

OFF

STRING

Topic Offset Mapping

Defines the reading offsets for each topic and partition in key value pairs. When an offset is set, the Kafka source skips reading messages before that offset.

Unless an offset is defined for a specific topic; Cortex reads messages from the very beginning.

e.g. reading2=20, temperature=500 reads from the 21th message of the reading2 topic, and from the 501th message of the temperature topic.

Default Value
Possible Data Type

STRING

Step 7 - Offset Commit Settings

Enable Offset Commiting

Default Value
Possible Data Type

STRING

Auto Commit Interval

Default Value
Possible Data Type

STRING

Enable Asynchronous Commiting

Default Value
Possible Data Type

STRING

Step 8 - Additional Settings

Additional Configurations

Contains any other configurations for the consumer in key-value format. Some supported configurations can be exemplified as

e.g. Keystore Type: 'ssl.keystore.type' as key and 'JKS' as value.

e.g. Batch Size: 'batch.size' as key and '200' as value.

Default Value
Possible Data Type

STRING

Step 9 - Preview

In Preview Step, you're provided with a concise summary of all the changes you've made to the Kafka Source Node. This step is pivotal for reviewing and ensuring that your configurations are as intended before completing node setup.

  • Viewing Configurations: Preview Step presents a consolidated view of your node setup.

  • Saving and Exiting: Use the Complete button to save your changes and exit the node and return back to Canvas.

  • Revisions: Use the Back button to return to any Step of modify node setup.

The Preview Step offers a user-friendly summary to manage and finalize node settings in Cortex.