Last active
June 13, 2018 07:06
-
-
Save affo/56de309bdbcf92720d69fbbeed31bd92 to your computer and use it in GitHub Desktop.
config file for flink
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
################################################################################ | |
# Licensed to the Apache Software Foundation (ASF) under one | |
# or more contributor license agreements. See the NOTICE file | |
# distributed with this work for additional information | |
# regarding copyright ownership. The ASF licenses this file | |
# to you under the Apache License, Version 2.0 (the | |
# "License"); you may not use this file except in compliance | |
# with the License. You may obtain a copy of the License at | |
# | |
# http://www.apache.org/licenses/LICENSE-2.0 | |
# | |
# Unless required by applicable law or agreed to in writing, software | |
# distributed under the License is distributed on an "AS IS" BASIS, | |
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |
# See the License for the specific language governing permissions and | |
# limitations under the License. | |
################################################################################ | |
#============================================================================== | |
# Common | |
#============================================================================== | |
# The external address of the host on which the JobManager runs and can be | |
# reached by the TaskManagers and any clients which want to connect. This setting | |
# is only used in Standalone mode and may be overwritten on the JobManager side | |
# by specifying the --host <hostname> parameter of the bin/jobmanager.sh executable. | |
# In high availability mode, if you use the bin/start-cluster.sh script and setup | |
# the conf/masters file, this will be taken care of automatically. Yarn/Mesos | |
# automatically configure the host name based on the hostname of the node where the | |
# JobManager runs. | |
jobmanager.rpc.address: jobmanager | |
# The RPC port where the JobManager is reachable. | |
jobmanager.rpc.port: 6123 | |
# The heap size for the JobManager JVM | |
jobmanager.heap.mb: 4096 | |
# The heap size for the TaskManager JVM | |
taskmanager.heap.mb: 4096 | |
# The number of task slots that each TaskManager offers. Each slot runs one parallel pipeline. | |
taskmanager.numberOfTaskSlots: 4 | |
# Specify whether TaskManager memory should be allocated when starting up (true) or when | |
# memory is required in the memory manager (false) | |
# Important Note: For pure streaming setups, we highly recommend to set this value to `false` | |
# as the default state backends currently do not use the managed memory. | |
taskmanager.memory.preallocate: false | |
# The parallelism used for programs that did not specify and other parallelism. | |
parallelism.default: 1 | |
#============================================================================== | |
# Web Frontend | |
#============================================================================== | |
# The address under which the web-based runtime monitor listens. | |
# | |
#jobmanager.web.address: 0.0.0.0 | |
# The port under which the web-based runtime monitor listens. | |
# A value of -1 deactivates the web server. | |
jobmanager.web.port: 8081 | |
# Flag to specify whether job submission is enabled from the web-based | |
# runtime monitor. Uncomment to disable. | |
#jobmanager.web.submit.enable: false | |
#============================================================================== | |
# HistoryServer | |
#============================================================================== | |
# The HistoryServer is started and stopped via bin/historyserver.sh (start|stop) | |
# Directory to upload completed jobs to. Add this directory to the list of | |
# monitored directories of the HistoryServer as well (see below). | |
#jobmanager.archive.fs.dir: hdfs:///completed-jobs/ | |
# The address under which the web-based HistoryServer listens. | |
#historyserver.web.address: 0.0.0.0 | |
# The port under which the web-based HistoryServer listens. | |
#historyserver.web.port: 8082 | |
# Comma separated list of directories to monitor for completed jobs. | |
#historyserver.archive.fs.dir: hdfs:///completed-jobs/ | |
# Interval in milliseconds for refreshing the monitored directories. | |
#historyserver.archive.fs.refresh-interval: 10000 | |
#============================================================================== | |
# Streaming state checkpointing | |
#============================================================================== | |
# The backend that will be used to store operator state checkpoints if | |
# checkpointing is enabled. | |
# | |
# Supported backends: jobmanager, filesystem, rocksdb, <class-name-of-factory> | |
# | |
# state.backend: filesystem | |
# Directory for storing checkpoints in a Flink-supported filesystem | |
# Note: State backend must be accessible from the JobManager and all TaskManagers. | |
# Use "hdfs://" for HDFS setups, "file://" for UNIX/POSIX-compliant file systems, | |
# (or any local file system under Windows), or "S3://" for S3 file system. | |
# | |
# state.backend.fs.checkpointdir: hdfs://namenode-host:port/flink-checkpoints | |
#============================================================================== | |
# Advanced | |
#============================================================================== | |
# The number of buffers for the network stack. | |
# | |
taskmanager.network.numberOfBuffers: 32384 | |
# Directories for temporary files. | |
# | |
# Add a delimited list for multiple directories, using the system directory | |
# delimiter (colon ':' on unix) or a comma, e.g.: | |
# /data1/tmp:/data2/tmp:/data3/tmp | |
# | |
# Note: Each directory entry is read from and written to by a different I/O | |
# thread. You can include the same directory multiple times in order to create | |
# multiple I/O threads against that directory. This is for example relevant for | |
# high-throughput RAIDs. | |
# | |
# If not specified, the system-specific Java temporary directory (java.io.tmpdir | |
# property) is taken. | |
# | |
# taskmanager.tmp.dirs: /tmp | |
# Path to the Hadoop configuration directory. | |
# | |
# This configuration is used when writing into HDFS. Unless specified otherwise, | |
# HDFS file creation will use HDFS default settings with respect to block-size, | |
# replication factor, etc. | |
# | |
# You can also directly specify the paths to hdfs-default.xml and hdfs-site.xml | |
# via keys 'fs.hdfs.hdfsdefault' and 'fs.hdfs.hdfssite'. | |
# | |
# fs.hdfs.hadoopconf: /path/to/hadoop/conf/ | |
#============================================================================== | |
# High Availability | |
#============================================================================== | |
# The high-availability mode. Possible options are 'NONE' or 'zookeeper'. | |
# | |
# high-availability: zookeeper | |
# The path where metadata for master recovery is persisted. While ZooKeeper stored | |
# the small ground truth for checkpoint and leader election, this location stores | |
# the larger objects, like persisted dataflow graphs. | |
# | |
# Must be a durable file system that is accessible from all nodes | |
# (like HDFS, S3, Ceph, nfs, ...) | |
# | |
# high-availability.storageDir: hdfs:///flink/ha/ | |
# The list of ZooKeeper quorum peers that coordinate the high-availability | |
# setup. This must be a list of the form: | |
# "host1:clientPort,host2:clientPort,..." (default clientPort: 2181) | |
# | |
# high-availability.zookeeper.quorum: localhost:2181 | |
# ACL options are based on https://zookeeper.apache.org/doc/r3.1.2/zookeeperProgrammers.html#sc_BuiltinACLSchemes | |
# It can be either "creator" (ZOO_CREATE_ALL_ACL) or "open" (ZOO_OPEN_ACL_UNSAFE) | |
# The default value is "open" and it can be changed to "creator" if ZK security is enabled | |
# | |
# high-availability.zookeeper.client.acl: open | |
#============================================================================== | |
# Flink Cluster Security Configuration (optional configuration) | |
#============================================================================== | |
# Kerberos authentication for various components - Hadoop, ZooKeeper, and connectors - | |
# may be enabled in four steps: | |
# 1. configure the local krb5.conf file | |
# 2. provide Kerberos credentials (either a keytab or a ticket cache w/ kinit) | |
# 3. make the credentials available to various JAAS login contexts | |
# 4. configure the connector to use JAAS/SASL | |
# The below configure how Kerberos credentials are provided. A keytab will be used instead of | |
# a ticket cache if the keytab path and principal are set. | |
# security.kerberos.login.use-ticket-cache: true | |
# security.kerberos.login.keytab: /path/to/kerberos/keytab | |
# security.kerberos.login.principal: flink-user | |
# The configuration below defines which JAAS login contexts | |
# security.kerberos.login.contexts: Client,KafkaClient | |
#============================================================================== | |
# ZK Security Configuration (optional configuration) | |
#============================================================================== | |
# Below configurations are applicable if ZK ensemble is configured for security | |
# Override below configuration to provide custom ZK service name if configured | |
# zookeeper.sasl.service-name: zookeeper | |
# The configuration below must match one of the values set in "security.kerberos.login.contexts" | |
# zookeeper.sasl.login-context-name: Client | |
akka.ask.timeout: 120s | |
akka.client.timeout: 120s | |
akka.startup-timeout: 20s | |
akka.framesize: 20000000b |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment