|
|
# Version 9.2.2.20240415
|
|
|
#
|
|
|
############################################################################
|
|
|
# OVERVIEW
|
|
|
############################################################################
|
|
|
# DO NOT EDIT THIS FILE!
|
|
|
# Changes to default files will be lost on update and are difficult to
|
|
|
# manage and support.
|
|
|
#
|
|
|
# Please make any changes to system defaults by overriding them in
|
|
|
# apps or $SPLUNK_HOME/etc/system/local
|
|
|
# (See "Configuration file precedence" in the web documentation).
|
|
|
#
|
|
|
# To override a specific setting, copy the name of the stanza and
|
|
|
# setting to the file where you wish to override it.
|
|
|
#
|
|
|
# This file configures various limits to the Splunk's search commands.
|
|
|
# CAUTION: Do not alter the settings in limits.conf unless you know what
|
|
|
# you are doing.
|
|
|
#
|
|
|
# Improperly configured limits may result in splunkd crashes and/or
|
|
|
# memory overuse.
|
|
|
#
|
|
|
############################################################################
|
|
|
# GLOBAL SETTINGS
|
|
|
############################################################################
|
|
|
|
|
|
[default]
|
|
|
max_mem_usage_mb = 200
|
|
|
|
|
|
[searchresults]
|
|
|
maxresultrows = 50000
|
|
|
# Maximum number of times to try in the atomic write operation
|
|
|
# (1 = no retries)
|
|
|
tocsv_maxretry = 5
|
|
|
# Retry period is 1/2 second (500 milliseconds)
|
|
|
tocsv_retryperiod_ms = 500
|
|
|
|
|
|
compression_level = 1
|
|
|
|
|
|
[search_info]
|
|
|
# These setting control logging of error messages to info.csv
|
|
|
# All messages will be logged to search.log regardless of these settings.
|
|
|
# maximum number of error messages to log in info.csv
|
|
|
# Set to 0 to remove limit, may affect search performance
|
|
|
max_infocsv_messages = 20
|
|
|
# log level = DEBUG | INFO | WARN | ERROR
|
|
|
infocsv_log_level = INFO
|
|
|
# Log warnings if search returns no results because user has no
|
|
|
# permissions to search on queried indexes.
|
|
|
show_warn_on_filtered_indexes = false
|
|
|
# Log level of messages when search returns no results because user has
|
|
|
# no permissions to search on queried indexes.
|
|
|
filteredindexes_log_level = DEBUG
|
|
|
|
|
|
|
|
|
[subsearch]
|
|
|
# Maximum number of results to return from a subsearch.
|
|
|
maxout = 10000
|
|
|
# Maximum number of seconds to run a subsearch before finalizing.
|
|
|
maxtime = 60
|
|
|
# Time to cache a given subsearch's results.
|
|
|
ttl = 300
|
|
|
|
|
|
############################################################################
|
|
|
# SEARCH COMMAND
|
|
|
############################################################################
|
|
|
# This section contains the settings for the search command.
|
|
|
# The settings are organized in subsections by type of setting.
|
|
|
|
|
|
[search]
|
|
|
|
|
|
############################################################################
|
|
|
# Batch search
|
|
|
############################################################################
|
|
|
# This section contains settings for batch search.
|
|
|
|
|
|
# Allow batch mode which searches in non-time order for certain classes
|
|
|
# of searches.
|
|
|
allow_batch_mode = true
|
|
|
|
|
|
# When batch mode attempts to retry the search on a peer that failed wait
|
|
|
# at least this many seconds.
|
|
|
batch_retry_min_interval = 5
|
|
|
|
|
|
# When batch mode attempts to retry the search on a peer that failed wait
|
|
|
# at most this many seconds.
|
|
|
batch_retry_max_interval = 300
|
|
|
|
|
|
# After a retry attempt fails increase the time to wait before trying
|
|
|
# again by this scaling factor.
|
|
|
batch_retry_scaling = 1.5
|
|
|
|
|
|
# When in batch mode what is the max number of index values to read in
|
|
|
# at one time.
|
|
|
batch_search_max_index_values = 10000000
|
|
|
|
|
|
# Number of search pipelines created per batch search
|
|
|
batch_search_max_pipeline = 1
|
|
|
|
|
|
# Default size of the aggregator queue to which all the search pipelines
|
|
|
# dump the search results on the indexer.
|
|
|
batch_search_max_results_aggregator_queue_size = 100000000
|
|
|
|
|
|
# Default size of the serialized results queue where all the serialized
|
|
|
# results are kept before transmission.
|
|
|
batch_search_max_serialized_results_queue_size = 100000000
|
|
|
|
|
|
|
|
|
############################################################################
|
|
|
# Bundles
|
|
|
############################################################################
|
|
|
# This section contains settings for bundles and bundle replication.
|
|
|
|
|
|
# Avoid loading remote bundles in splunkd.
|
|
|
load_remote_bundles = false
|
|
|
|
|
|
# Bundle replication file ttl.
|
|
|
replication_file_ttl = 600
|
|
|
|
|
|
# The minimum bundle replication period.
|
|
|
replication_period_sec = 60
|
|
|
|
|
|
# Whether bundle replication is synchronous (and thus blocking searches).
|
|
|
sync_bundle_replication = auto
|
|
|
|
|
|
# Bundle status expiry time
|
|
|
bundle_status_expiry_time = 1hr
|
|
|
|
|
|
############################################################################
|
|
|
# Concurrency
|
|
|
############################################################################
|
|
|
# This section contains settings for search concurrency limits.
|
|
|
# If 'total_search_concurrency_limit = auto', the total limit of concurrent
|
|
|
# historical searches is
|
|
|
# max_hist_searches = max_searches_per_cpu x number_of_cpus + base_max_searches.
|
|
|
|
|
|
# The maximum number of concurrent historical searches in the search head.
|
|
|
total_search_concurrency_limit = auto
|
|
|
|
|
|
# The base number of concurrent historical searches.
|
|
|
base_max_searches = 6
|
|
|
|
|
|
# Max real-time searches = max_rt_search_multiplier x max historical searches.
|
|
|
max_rt_search_multiplier = 1
|
|
|
|
|
|
# The maximum number of concurrent historical searches per CPU.
|
|
|
max_searches_per_cpu = 1
|
|
|
|
|
|
# Whether maximum number of concurrent searches are enforced cluster-wide
|
|
|
# for admission of adhoc searches
|
|
|
shc_adhoc_quota_enforcement = off
|
|
|
|
|
|
############################################################################
|
|
|
# Distributed search
|
|
|
############################################################################
|
|
|
# This section contains settings for distributed search connection
|
|
|
# information.
|
|
|
|
|
|
# Limit on the skew permitted when adding a search peer.
|
|
|
# Peers with a skew larger than this will be rejected.
|
|
|
addpeer_skew_limit = 600
|
|
|
|
|
|
# Defaults to download all remote logs other than saved search logs and
|
|
|
# oneshot search logs.
|
|
|
fetch_remote_search_log = disabledSavedSearches
|
|
|
|
|
|
# Maximum size of the chunk queue.
|
|
|
max_chunk_queue_size = 10000000
|
|
|
|
|
|
# Search results combiner maximum in-memory buffer size (in events).
|
|
|
max_combiner_memevents = 50000
|
|
|
|
|
|
# Absolute value of largest time skew we will tolerate between the search
|
|
|
# head and the peer (in seconds).
|
|
|
max_tolerable_skew = 60
|
|
|
|
|
|
# Maximum number of worker threads in Round Robin policy.
|
|
|
max_workers_searchparser = 5
|
|
|
|
|
|
# The minimum number of results blobs to keep for consumption by the
|
|
|
# search head.
|
|
|
results_queue_min_size = 10
|
|
|
|
|
|
# Corresponds to the size of the results queue in the dispatch fetch level
|
|
|
result_queue_max_size = 100000000
|
|
|
|
|
|
# If all currently active peers have finished with the search wait this
|
|
|
# many seconds before giving up on peers we are attempting to reconnect
|
|
|
# to for a retry.
|
|
|
results_queue_read_timeout_sec = 900
|
|
|
|
|
|
# by default we do not enable throttling
|
|
|
remote_search_requests_throttling_type = disabled
|
|
|
|
|
|
# By default, send user capabilities to the search peers.
|
|
|
remote_search_requests_send_capabilities_list = true
|
|
|
|
|
|
# By default, allow remote search execution even if the capability list is missing.
|
|
|
remote_search_requests_reject_if_capabilities_list_absent = false
|
|
|
|
|
|
############################################################################
|
|
|
# Field stats
|
|
|
############################################################################
|
|
|
# This section contains settings for field statistics.
|
|
|
|
|
|
# How often to update the field summary statistics, as a ratio to the
|
|
|
# elapsed run time so far.
|
|
|
fieldstats_update_freq = 0
|
|
|
|
|
|
# Maximum period for updating field summary statistics in seconds.
|
|
|
fieldstats_update_maxperiod = 60
|
|
|
|
|
|
# The minimum frequency of a field displayed in the /summary endpoint.
|
|
|
min_freq = 0.01
|
|
|
|
|
|
|
|
|
############################################################################
|
|
|
# History
|
|
|
############################################################################
|
|
|
# This section contains settings for search history.
|
|
|
|
|
|
# Enable search history?
|
|
|
enable_history = true
|
|
|
|
|
|
# Max number of searches to store in history
|
|
|
# for each user/app, if search_history_storage_mode is csv;
|
|
|
# for each user, if search_history_storage_mode is kvstore.
|
|
|
max_history_length = 500
|
|
|
|
|
|
# Max time search history records to store in history (for each user).
|
|
|
max_history_storage_retention_time = 90d
|
|
|
|
|
|
# History storage
|
|
|
search_history_storage_mode = csv
|
|
|
|
|
|
############################################################################
|
|
|
# Memory tracker
|
|
|
############################################################################
|
|
|
# This section contains settings for the memory tracker.
|
|
|
|
|
|
# If memory tracker is disabled, search won't be terminated even if it
|
|
|
# exceeds the memory limit.
|
|
|
# By default memory tracking is disabled.
|
|
|
enable_memory_tracker = false
|
|
|
|
|
|
# Default value for percentage memory usage for the splunk search
|
|
|
# process is set to 25%.
|
|
|
search_process_memory_usage_percentage_threshold = 25
|
|
|
|
|
|
# Default value for memory usage for the Splunk search process is set to 4GB.
|
|
|
search_process_memory_usage_threshold = 4000
|
|
|
|
|
|
|
|
|
############################################################################
|
|
|
# Meta search
|
|
|
############################################################################
|
|
|
# This section contains settings for meta search.
|
|
|
|
|
|
# Allow inexact metasearch?
|
|
|
allow_inexact_metasearch = false
|
|
|
|
|
|
|
|
|
############################################################################
|
|
|
# Misc
|
|
|
############################################################################
|
|
|
# This section contains miscellaneous search settings.
|
|
|
|
|
|
|
|
|
# Determines if the saved searches handler uses a removable cache
|
|
|
use_removable_search_cache = true
|
|
|
|
|
|
# Specifies after how long a paused search should be auto canceled,
|
|
|
# in seconds.
|
|
|
# 0 means do not auto cancel the paused search.
|
|
|
auto_cancel_after_pause = 0
|
|
|
|
|
|
dispatch_dir_warning_size = 5000
|
|
|
|
|
|
# Enable concatenation of successively occurring evals into a single
|
|
|
# comma separated eval during generation of data model searches.
|
|
|
enable_datamodel_meval = true
|
|
|
|
|
|
# Determines whether or not scoped conditional expansion of knowledge
|
|
|
# objects occurs during search string expansion. This only applies on
|
|
|
# the search head.
|
|
|
# NOTE: Do not change unless instructed to do so by Splunk Support.
|
|
|
enable_conditional_expansion = true
|
|
|
|
|
|
# If true, always dispatch saved searches as the requesting user.
|
|
|
# The default for dispatchAs in the savedsearches.conf.spec.in file
|
|
|
# is 'owner'.
|
|
|
force_saved_search_dispatch_as_user = false
|
|
|
|
|
|
# Max length of custom job id when passing spawning new job.
|
|
|
max_id_length = 150
|
|
|
|
|
|
# Specifies the maximum length of a generated or custom search job ID before
|
|
|
# the Splunk software shortens the directory name. The search job ID itself
|
|
|
# remains the same.
|
|
|
max_id_length_before_hash = 230
|
|
|
|
|
|
# Specifies whether the Splunk software reruns all or elements of a currently
|
|
|
# running search process when there are indexer failures in an indexer
|
|
|
# clustering environment.
|
|
|
search_retry = false
|
|
|
|
|
|
|
|
|
# Sets how long, in seconds, 'search_retry' waits to get updated
|
|
|
# indexer information.
|
|
|
search_retry_waiting_time = 70
|
|
|
|
|
|
# Maximum number of attempts made to retry a historical search before failing
|
|
|
# Only applied when search_retry is set to true
|
|
|
search_retry_max_historical = 15
|
|
|
|
|
|
# Stack size of the search executing thread.
|
|
|
stack_size = 4194304
|
|
|
|
|
|
# Use precomputed summaries if possible?
|
|
|
summary_mode = all
|
|
|
|
|
|
# Track indextime range of searches (shown in job inspector).
|
|
|
track_indextime_range = true
|
|
|
|
|
|
# By default use bloom filter.
|
|
|
use_bloomfilter = true
|
|
|
|
|
|
# By default use metadata elimination.
|
|
|
use_metadata_elimination = true
|
|
|
|
|
|
# Serialization format and compression algorithm used for search results
|
|
|
results_serial_format = srs
|
|
|
results_compression_algorithm = zstd
|
|
|
|
|
|
# Record search telemetry in search_telemetry.json in the dispatch dir
|
|
|
# The setting in telemetry.conf controls whether the data is sent back,
|
|
|
# this setting controls whether we generate search telemetry data in the
|
|
|
# dispatch dir. Search telemetry data is also put into _introspection.
|
|
|
record_search_telemetry = true
|
|
|
|
|
|
|
|
|
# Number of files to use as a threshold at which to stop adding more
|
|
|
# files to var/run/splunk/search_telemetry for indexing search telemetry
|
|
|
# data. This setting applies only to telemetry on the search head.
|
|
|
search_telemetry_file_limit = 500
|
|
|
|
|
|
|
|
|
# Limit, in bytes, for each of the constituent components of the search
|
|
|
# telemetry json representation
|
|
|
search_telemetry_component_limit = 10000
|
|
|
|
|
|
# track of the number of events of each sourcetype that match a search
|
|
|
track_matching_sourcetypes = true
|
|
|
|
|
|
# timeout to launch a search job
|
|
|
search_launch_timeout_seconds = 180
|
|
|
|
|
|
# timeout to initialize startup configuration, in milliseconds
|
|
|
search_startup_config_timeout_ms = 3000
|
|
|
|
|
|
# maximum number of tracked search result sourcetypes to add to audit.log
|
|
|
max_audit_sourcetypes = 100
|
|
|
|
|
|
use_search_evaluator_v2 = true
|
|
|
|
|
|
# The maximum number of field metadata displayed in the /jobs/fieldmeta endpoint.
|
|
|
max_fieldmeta_cnt_ui = 1000
|
|
|
|
|
|
############################################################################
|
|
|
# Parsing
|
|
|
############################################################################
|
|
|
# This section contains settings related to parsing searches.
|
|
|
|
|
|
# Max recursion depth for macros.
|
|
|
# Considered a search exception if macro expansion does not stop after
|
|
|
# this many levels.
|
|
|
max_macro_depth = 100
|
|
|
|
|
|
# Max recursion depth for subsearch.
|
|
|
# Considered a search exception if subsearch does not stop after
|
|
|
# this many levels.
|
|
|
max_subsearch_depth = 8
|
|
|
|
|
|
# The minimum length of a prefix before a * to ask the index about.
|
|
|
min_prefix_len = 1
|
|
|
|
|
|
# When true, always search the lexicon for both field::val and val for
|
|
|
# field=val searches, unless INDEXED=true is set for the field in
|
|
|
# in fields.conf (in which case only field::val is searched)
|
|
|
always_include_indexedfield_lispy = true
|
|
|
|
|
|
# When set to true, we will scope every indexed field=val statement
|
|
|
# with sourcetype and convert it to the indexed form (field::val)
|
|
|
indexed_fields_expansion = true
|
|
|
|
|
|
############################################################################
|
|
|
# Preview
|
|
|
############################################################################
|
|
|
# This section contains settings for previews.
|
|
|
|
|
|
# The maximum time to spend generating previews, as a fraction of total
|
|
|
# search time.
|
|
|
preview_duty_cycle = 0.25
|
|
|
|
|
|
|
|
|
############################################################################
|
|
|
# Quota or queued searches
|
|
|
############################################################################
|
|
|
# This section contains settings for quota or queued searches.
|
|
|
|
|
|
# Default setting for allowing async jobs to be queued if quota violation.
|
|
|
default_allow_queue = true
|
|
|
|
|
|
# The maximum number of times to retry to dispatch a search when the
|
|
|
# quota has been reached.
|
|
|
dispatch_quota_retry = 4
|
|
|
|
|
|
# Milliseconds between retrying to dispatch a search if a quota has been
|
|
|
# reached. We retry the given number of times, with each successive wait
|
|
|
# 2x longer than the previous.
|
|
|
dispatch_quota_sleep_ms = 100
|
|
|
|
|
|
# Enforce cumulative role based quotas
|
|
|
enable_cumulative_quota = false
|
|
|
|
|
|
# how often to retry queued jobs (in seconds)
|
|
|
queued_job_check_freq = 1
|
|
|
|
|
|
|
|
|
############################################################################
|
|
|
# Reading chunk controls
|
|
|
############################################################################
|
|
|
# This section contains settings for reading chunk controls.
|
|
|
|
|
|
# max_results_perchunk, min_results_perchunk, and target_time_perchunk
|
|
|
# are multiplied by this for a long running search.
|
|
|
chunk_multiplier = 5
|
|
|
|
|
|
# Time in seconds until a search is considered "long running”.
|
|
|
long_search_threshold = 2
|
|
|
|
|
|
# Maximum raw size of results for each call to search (in dispatch).
|
|
|
# 0 = no limit, not affected by chunk_multiplier.
|
|
|
max_rawsize_perchunk = 100000000
|
|
|
|
|
|
# Maximum results per call to search (in dispatch).
|
|
|
# Must be <= maxresultrows.
|
|
|
max_results_perchunk = 2500
|
|
|
|
|
|
# Minimum results per call to search (in dispatch).
|
|
|
# Must be <= max_results_perchunk.
|
|
|
min_results_perchunk = 100
|
|
|
|
|
|
# Target duration of a particular call to fetch search results in ms.
|
|
|
target_time_perchunk = 2000
|
|
|
|
|
|
|
|
|
############################################################################
|
|
|
# Real-time
|
|
|
############################################################################
|
|
|
# This section contains settings for real-time searches.
|
|
|
|
|
|
# For real-time searches in the UI, maximum number of events stored
|
|
|
# (as a FIFO buffer).
|
|
|
realtime_buffer = 10000
|
|
|
|
|
|
|
|
|
############################################################################
|
|
|
# Remote storage
|
|
|
############################################################################
|
|
|
# This section contains settings for remote storage.
|
|
|
|
|
|
# Maximum number of remote buckets to localize as a look-ahead on searches.
|
|
|
bucket_localize_max_lookahead = 5
|
|
|
|
|
|
# Specifies which bucket prediction algorithm to use.
|
|
|
bucket_predictor = consec_not_needed
|
|
|
|
|
|
|
|
|
############################################################################
|
|
|
# Results storage
|
|
|
############################################################################
|
|
|
# This section contains settings for storing final search results.
|
|
|
|
|
|
# The maximum number of end results to store globally
|
|
|
# (when status_buckets=0).
|
|
|
max_count = 500000
|
|
|
|
|
|
# By default, no timeline information is retained. UI will supply the
|
|
|
# status_buckets as needed.
|
|
|
status_buckets = 0
|
|
|
|
|
|
# Truncate report output to max_count?
|
|
|
truncate_report = false
|
|
|
|
|
|
# Do we write multi-file results to results_dir?
|
|
|
write_multifile_results_out = true
|
|
|
|
|
|
# For event searches should we read final results from the timeliner
|
|
|
read_final_results_from_timeliner = true
|
|
|
|
|
|
# Field filters are turned off by default in the preview release.
|
|
|
field_filters = false
|
|
|
|
|
|
############################################################################
|
|
|
# Search process
|
|
|
############################################################################
|
|
|
# This section contains settings for search process configurations.
|
|
|
|
|
|
# Even if the search process has not been idle for the above time, check its
|
|
|
# internal caches for stale data after this many searches.
|
|
|
idle_process_cache_search_count = 8
|
|
|
|
|
|
# If a search process is idle for this many seconds, take the opportunity
|
|
|
# to scan its internal caches for stale data.
|
|
|
idle_process_cache_timeout = 0.5
|
|
|
|
|
|
# Periodically we'll check if we have too many idle search processes.
|
|
|
# This controls how often that happens (in seconds).
|
|
|
idle_process_reaper_period = auto
|
|
|
|
|
|
# Inside a search process, keep up to this many compiled regex artifacts
|
|
|
# before checking for stale ones. Normally the above idle_process_cache_*
|
|
|
# settings will check for stale entries before this limit is hit.
|
|
|
idle_process_regex_cache_hiwater = 2500
|
|
|
|
|
|
# When running a search, scan at most this many idle processes before
|
|
|
# launching a new one.
|
|
|
launcher_max_idle_checks = auto
|
|
|
|
|
|
# Number of server threads dedicated to managing communication with
|
|
|
# search processes.
|
|
|
# Negative number means automatically pick a sensible value
|
|
|
launcher_threads = -1
|
|
|
|
|
|
# Maximum number of preforked search processes that are idle
|
|
|
# and wait for next search execution
|
|
|
max_idle_process_count = auto
|
|
|
|
|
|
# Memory (RSS) limit of a search process that can be idle and reusable, in KB
|
|
|
# Number 0 is set to use the system default, 1024*1024 KB (1GB)
|
|
|
# Negative number means no limit
|
|
|
max_idle_process_memory = auto
|
|
|
|
|
|
# The number of search processes constructing a pool to run searches, which is
|
|
|
# dependent on system resources (CPU and memory) available to Splunk server
|
|
|
# Number 0 is set to automatically pick a sensible value
|
|
|
# Negative number means no limit
|
|
|
max_search_process_pool = 2048
|
|
|
|
|
|
# When reaping idle search processes, allow one to be reaped if it is
|
|
|
# not using the most recent configuration bundle, and its bundle has not
|
|
|
# been used in at least this many seconds.
|
|
|
max_old_bundle_idle_time = auto
|
|
|
|
|
|
# On UNIX we can run more that one search per process.
|
|
|
# Set this to a number greater than one to enable.
|
|
|
max_searches_per_process = 500
|
|
|
|
|
|
# When running more than one search per process, limit the number of new
|
|
|
# searches that can be started before allowing time to service the ones
|
|
|
# that are already running
|
|
|
max_searches_started_per_cycle = 30
|
|
|
|
|
|
# When running more than one search per process, don't allow a process to
|
|
|
# accumulate more than this number of seconds running searches. Note that a
|
|
|
# search can run longer than this without being terminated, it only prevents
|
|
|
# the process from being used for another search
|
|
|
max_time_per_process = auto
|
|
|
|
|
|
# When running more than one search per process, do not reuse a process
|
|
|
# if it is older than this number of seconds. This is different than
|
|
|
# max_time_per_process because it includes time the process spent idle.
|
|
|
process_max_age = 7200.0
|
|
|
|
|
|
# Don't reuse a process that last served a different user unless it has
|
|
|
# been idle this long (in seconds).
|
|
|
process_min_age_before_user_change = auto
|
|
|
|
|
|
search_process_mode = auto
|
|
|
|
|
|
# Whether to increase oom_score of search processes to make splunk more stable.
|
|
|
search_process_configure_oom_score_adj = true
|
|
|
|
|
|
# The value added to search process, ranging between 0 and 1000.
|
|
|
# Only applies when 'search_process_configure_oom_score_adj' is set to true.
|
|
|
search_process_set_oom_score_adj = 700
|
|
|
|
|
|
############################################################################
|
|
|
# search_messages.log
|
|
|
############################################################################
|
|
|
|
|
|
# Specifies whether splunkd promotes user-facing search messages
|
|
|
# from $SPLUNK_HOME/var/run/splunk/dispatch/<sid>/info.csv to
|
|
|
# $SPLUNK_HOME/var/log/splunk/search_messages.log.
|
|
|
log_search_messages = true
|
|
|
|
|
|
# When 'log_search_messages = true', this setting specifies the lowest
|
|
|
# severity of message that splunkd logs to search_messages.log.
|
|
|
# Splunkd ignores all messages with a lower severity.
|
|
|
# Possible values in ascending order: DEBUG, INFO, WARN, ERROR
|
|
|
search_messages_severity = WARN
|
|
|
|
|
|
############################################################################
|
|
|
# Search reuse
|
|
|
############################################################################
|
|
|
# This section contains settings for search reuse.
|
|
|
|
|
|
|
|
|
############################################################################
|
|
|
# Splunk Analytics for Hadoop
|
|
|
############################################################################
|
|
|
# This section contains settings for use with Splunk Analytics for Hadoop.
|
|
|
|
|
|
# The maximum time to spend doing reduce, as a fraction of total search time.
|
|
|
reduce_duty_cycle = 0.25
|
|
|
|
|
|
# The frequency with which try to reduce intermediate data when there is
|
|
|
# a non-streaming and non-stateful streaming command. (0 = never)
|
|
|
reduce_freq = 10
|
|
|
|
|
|
|
|
|
############################################################################
|
|
|
# Status
|
|
|
############################################################################
|
|
|
# This section contains settings for search status.
|
|
|
|
|
|
# The number of search job metadata to cache in RAM.
|
|
|
status_cache_size = 10000
|
|
|
|
|
|
|
|
|
############################################################################
|
|
|
# Timelines
|
|
|
############################################################################
|
|
|
# This section contains settings for timelines.
|
|
|
|
|
|
# Size of thread pool for remote event download framework.
|
|
|
remote_event_download_initialize_pool = 5
|
|
|
remote_event_download_finalize_pool = 5
|
|
|
remote_event_download_local_pool = 5
|
|
|
|
|
|
# Allow timeline to be map/reduced?
|
|
|
remote_timeline = true
|
|
|
|
|
|
# Whether to fetch all events accessible through the timeline from the
|
|
|
# remote peers before the job is considered done.
|
|
|
remote_timeline_fetchall = 1
|
|
|
|
|
|
# Minimum number of peers required to utilize remote timelining.
|
|
|
remote_timeline_min_peers = 1
|
|
|
|
|
|
# How often to touch remote artifacts to keep them from being reaped
|
|
|
# when search has not finished? (in seconds).
|
|
|
remote_timeline_touchperiod = 300
|
|
|
|
|
|
# Timeouts for fetching remote timeline events.
|
|
|
remote_timeline_connection_timeout = 5
|
|
|
remote_timeline_send_timeout = 10
|
|
|
remote_timeline_receive_timeout = 10
|
|
|
|
|
|
# In ms
|
|
|
search_keepalive_frequency = 30000
|
|
|
|
|
|
# Maximum number of uninterrupted keepalives before the connection is closed.
|
|
|
search_keepalive_max = 100
|
|
|
|
|
|
# Enable timeline preview
|
|
|
timeline_events_preview = false
|
|
|
|
|
|
############################################################################
|
|
|
# TTL
|
|
|
############################################################################
|
|
|
# This section contains time to live (ttl) settings.
|
|
|
|
|
|
# The length of time to persist search cache entries (in seconds).
|
|
|
cache_ttl = 300
|
|
|
|
|
|
# How long jobs are saved for by default.
|
|
|
default_save_ttl = 604800
|
|
|
|
|
|
# How long searches should be stored on disk once failed.
|
|
|
failed_job_ttl = 86400
|
|
|
|
|
|
# How long should searches run for a search head live on the indexers.
|
|
|
remote_ttl = 600
|
|
|
|
|
|
# How long searches should be stored on disk once completed.
|
|
|
ttl = 600
|
|
|
|
|
|
|
|
|
# Timeout value for checking search marker files like hotbucketmarker or backfill
|
|
|
# marker.
|
|
|
check_search_marker_done_interval = 60
|
|
|
|
|
|
# Time interval of sleeping between subsequent search marker files checks.
|
|
|
check_search_marker_sleep_interval = 1
|
|
|
|
|
|
# How long srtemp sub-directories should be kept before they are deleted
|
|
|
srtemp_dir_ttl = 86400
|
|
|
|
|
|
############################################################################
|
|
|
# Distributed search throttling
|
|
|
############################################################################
|
|
|
# This section contains settings for distributed search throttling (peers
|
|
|
# side) information.
|
|
|
[search_throttling::per_cpu]
|
|
|
max_concurrent = 12
|
|
|
|
|
|
[search_throttling::physical_ram]
|
|
|
min_memory_per_search = 134217728
|
|
|
|
|
|
############################################################################
|
|
|
# OTHER COMMAND SETTINGS
|
|
|
############################################################################
|
|
|
# This section contains the stanzas for the SPL commands, except for the
|
|
|
# search command, which is in a separate section.
|
|
|
|
|
|
|
|
|
[anomalousvalue]
|
|
|
maxresultrows = 50000
|
|
|
|
|
|
# Maximum number of distinct values for a field.
|
|
|
maxvalues = 0
|
|
|
# Maximum size in bytes of any single value
|
|
|
# (truncated to this size if larger).
|
|
|
maxvaluesize = 0
|
|
|
|
|
|
|
|
|
[associate]
|
|
|
maxfields = 10000
|
|
|
maxvalues = 0
|
|
|
maxvaluesize = 0
|
|
|
|
|
|
|
|
|
[autoregress]
|
|
|
maxp = 10000
|
|
|
maxrange = 1000
|
|
|
|
|
|
[collect]
|
|
|
# Setting for multivalue field representations in collect command.
|
|
|
format_multivalue_collect = false
|
|
|
# Setting for collect command to add quotation marks based on major breakers.
|
|
|
collect_ignore_minor_breakers = false
|
|
|
|
|
|
[concurrency]
|
|
|
# Maximum concurrency level to keep record of.
|
|
|
max_count = 10000000
|
|
|
|
|
|
|
|
|
[correlate]
|
|
|
maxfields = 1000
|
|
|
|
|
|
|
|
|
[ctable]
|
|
|
# This stanza contains settings for the contingency/ctable/counttable
|
|
|
# command.
|
|
|
maxvalues = 1000
|
|
|
|
|
|
[dbinspect]
|
|
|
maxresultrows = 50000
|
|
|
|
|
|
[discretize]
|
|
|
# This stanza contains settings for the bin/bucket/discretize command.
|
|
|
maxbins = 50000
|
|
|
# if maxbins not specified or = 0, defaults to searchresults::maxresultrows
|
|
|
|
|
|
|
|
|
[findkeywords]
|
|
|
maxevents = 50000
|
|
|
|
|
|
|
|
|
[geostats]
|
|
|
# At the lowest level of the tree, i.e. ZL=0 (when we are zoomed out to
|
|
|
# the world level ), what is the size of each gridcell in terms of latitude
|
|
|
# and longitude (degrees)?
|
|
|
# Valid values for zl_0_gridcell_latspan are from 0 to 180.0, and
|
|
|
# for zl_0_gridcell_longspan are from 0 to 360.0.
|
|
|
# Rest of the zoom level gridcell sizes are auto-tuning, i.e. will
|
|
|
# reduce by a factor of 2 at each additional level.
|
|
|
zl_0_gridcell_latspan = 22.5
|
|
|
zl_0_gridcell_longspan = 45.0
|
|
|
# Configures the filtering/search strategy for events on the map.
|
|
|
# Currently experimental.
|
|
|
filterstrategy = 2
|
|
|
# How many levels of clustering will be done in geostats.
|
|
|
maxzoomlevel = 9
|
|
|
|
|
|
|
|
|
[inputcsv]
|
|
|
# Maximum number of retries for creating a tmp directory (with random
|
|
|
# name in SPLUNK_HOME/var/run/splunk).
|
|
|
mkdir_max_retries = 100
|
|
|
|
|
|
|
|
|
[join]
|
|
|
# The join command subsearch is restricted by two settings, 'subsearch_maxout'
|
|
|
# and the 'maxresultrows' setting in the [searchresults] stanza.
|
|
|
subsearch_maxout = 50000
|
|
|
subsearch_maxtime = 60
|
|
|
|
|
|
[kmeans]
|
|
|
maxdatapoints = 100000000
|
|
|
maxkvalue = 1000
|
|
|
maxkrange = 100
|
|
|
|
|
|
|
|
|
[lookup]
|
|
|
# Maximum size of static lookup file to use a in-memory index for.
|
|
|
max_memtable_bytes = 26214400
|
|
|
# Maximum size of static lookup file to use when using a lookup()
|
|
|
# eval function in the ingest context
|
|
|
# Defaults to 10MB
|
|
|
ingest_max_memtable_bytes = 10485760
|
|
|
# Period of time after which we should refresh in-memory lookup
|
|
|
# tables being used with lookup() eval function at ingest time.
|
|
|
ingest_lookup_refresh_period_secs = 60
|
|
|
# Maximum reverse lookup matches (for search expansion).
|
|
|
max_reverse_matches = 50
|
|
|
# Default setting for if non-memory file lookups (for large files)
|
|
|
# should batch queries.
|
|
|
# Can be overridden using a lookup table's stanza in transforms.conf.
|
|
|
batch_index_query = true
|
|
|
# When doing batch request, what's the most matches to retrieve?
|
|
|
# If more than this limit of matches would otherwise be retrieved,
|
|
|
# we will fall back to non-batch mode matching.
|
|
|
batch_response_limit = 5000000
|
|
|
# Maximum number of lookup error messages that should be logged.
|
|
|
max_lookup_messages = 20
|
|
|
# time to live for an indexed csv
|
|
|
indexed_csv_ttl = 300
|
|
|
# keep alive token file period
|
|
|
indexed_csv_keep_alive_timeout = 30
|
|
|
# max time for the CSV indexing
|
|
|
indexed_csv_inprogress_max_timeout = 300
|
|
|
# whether we want to error on invalid lookups or let them proceed.
|
|
|
input_errors_fatal = false
|
|
|
# Should KV Store lookups be indexed at time of bundle replication?
|
|
|
enable_splunkd_kv_lookup_indexing = true
|
|
|
|
|
|
|
|
|
[metadata]
|
|
|
maxresultrows = 10000
|
|
|
# The most metadata results to fetch from each indexer.
|
|
|
maxcount = 100000
|
|
|
bucket_localize_max_lookahead = 10
|
|
|
|
|
|
[metric_alerts]
|
|
|
condition_evaluation_interval = 1
|
|
|
search_delay = 15s+
|
|
|
search_ttl = 2p
|
|
|
honor_action = false
|
|
|
|
|
|
[msearch]
|
|
|
chunk_size = 1000
|
|
|
target_per_timeseries = 5
|
|
|
|
|
|
[mvcombine]
|
|
|
max_mem_usage_mb = 500
|
|
|
|
|
|
|
|
|
[mvexpand]
|
|
|
max_mem_usage_mb = 500
|
|
|
|
|
|
|
|
|
[outputlookup]
|
|
|
# Specifies if the outputlookup command should check if the user
|
|
|
# has write permission on the lookup file.
|
|
|
# The permission is set in the .meta file.
|
|
|
outputlookup_check_permission = false
|
|
|
# Specifies the context where the lookup file will be created for the first time.
|
|
|
create_context = app
|
|
|
|
|
|
[rare]
|
|
|
maxresultrows = 50000
|
|
|
# Maximum distinct value vectors to keep track of.
|
|
|
maxvalues = 0
|
|
|
maxvaluesize = 0
|
|
|
|
|
|
[rest]
|
|
|
allow_reload = false
|
|
|
|
|
|
[set]
|
|
|
maxresultrows = 50000
|
|
|
|
|
|
|
|
|
[sort]
|
|
|
# maximum number of concurrent files to open
|
|
|
maxfiles = 64
|
|
|
|
|
|
|
|
|
[spath]
|
|
|
# Number of characters to read from an XML or JSON event when
|
|
|
# auto extracting.
|
|
|
extraction_cutoff = 5000
|
|
|
extract_all = true
|
|
|
|
|
|
|
|
|
[stats]
|
|
|
max_keymap_rows = 1000000
|
|
|
maxresultrows = 50000
|
|
|
maxvalues = 0
|
|
|
maxvaluesize = 0
|
|
|
# For streamstats's maximum window size.
|
|
|
max_stream_window = 10000
|
|
|
# For rdigest, used to approximate order statistics (median, percentiles).
|
|
|
rdigest_k = 100
|
|
|
rdigest_maxnodes = 1
|
|
|
tdigest_k = 50
|
|
|
tdigest_max_buffer_size = 1000
|
|
|
tmpfile_compression = lz4
|
|
|
tmpfile_compression_level = 0
|
|
|
perc_digest_type = tdigest
|
|
|
list_maxsize = 100
|
|
|
min_chunk_size_kb = 64
|
|
|
max_chunk_size_kb = 4096
|
|
|
chunk_size_double_every = 100
|
|
|
# Determines whether to return results for searches with time-sensitive
|
|
|
# aggregations and missing or invalid timestamps in input events.
|
|
|
check_for_invalid_time = false
|
|
|
|
|
|
[sistats]
|
|
|
max_keymap_rows = 1000000
|
|
|
maxvalues = 0
|
|
|
maxvaluesize = 0
|
|
|
rdigest_k = 100
|
|
|
rdigest_maxnodes = 1
|
|
|
tdigest_k = 50
|
|
|
tdigest_max_buffer_size = 1000
|
|
|
perc_digest_type = tdigest
|
|
|
max_valuemap_bytes = 100000
|
|
|
|
|
|
[top]
|
|
|
maxresultrows = 50000
|
|
|
# Maximum distinct value vectors to keep track of.
|
|
|
maxvalues = 0
|
|
|
maxvaluesize = 0
|
|
|
|
|
|
|
|
|
[transactions]
|
|
|
# Maximum number of open transaction or events in open.
|
|
|
# Transaction before transaction eviction happens.
|
|
|
maxopentxn = 5000
|
|
|
maxopenevents = 100000
|
|
|
|
|
|
|
|
|
[tscollect]
|
|
|
# Default value of 'squashcase' arg if not specified by the command.
|
|
|
squashcase = false
|
|
|
# Default value of 'keepresults' arg if not specified by the command.
|
|
|
keepresults = false
|
|
|
# The max allowed size of tsidx files to create in megabytes.
|
|
|
# ’0’ implies no limit
|
|
|
optimize_max_size_mb = 256
|
|
|
|
|
|
|
|
|
[tstats]
|
|
|
# Whether we apply role-based search filters when users run tstats
|
|
|
# on normal index data (never applied on data from tscollect or
|
|
|
# data model acceleration).
|
|
|
apply_search_filter = true
|
|
|
# Default value of 'summariesonly' arg if not specified by the command.
|
|
|
summariesonly = false
|
|
|
# Default value of 'allow_old_summaries' arg if not specified
|
|
|
# by the command.
|
|
|
allow_old_summaries = false
|
|
|
# By default we retrieve up to ten million events at once from a
|
|
|
# TSIDX file when answering queries.
|
|
|
chunk_size = 10000000
|
|
|
# By default, do not include non-numeric values when applying
|
|
|
# searches that filter on numeric values.
|
|
|
include_events_omitted_when_filtering_numeric_values = false
|
|
|
# Number of search pipelines created per batch search
|
|
|
batch_search_max_pipeline = 1
|
|
|
# Should tstats use bloomfilters to eliminate buckets
|
|
|
use_bloomfilter = true
|
|
|
update_datamodel_usage_stats = true
|
|
|
|
|
|
[mstats]
|
|
|
time_bin_limit = 1000000
|
|
|
# Should mstats use bloomfilters to eliminate buckets
|
|
|
use_bloomfilter = true
|
|
|
|
|
|
[typeahead]
|
|
|
maxcount = 1000
|
|
|
max_servers = 2
|
|
|
fetch_multiplier = 50
|
|
|
use_cache = true
|
|
|
cache_ttl_sec = 300
|
|
|
min_prefix_length = 1
|
|
|
max_concurrent_per_user = 3
|
|
|
banned_segments =
|
|
|
|
|
|
|
|
|
[typer]
|
|
|
# In eventtyping, pay attention to first N characters of any
|
|
|
# attribute (e.g., _raw), including individual tokens. Can be
|
|
|
# overridden by supplying the typer operator with the argument.
|
|
|
# maxlen (e.g. "|typer maxlen=300").
|
|
|
maxlen = 10000
|
|
|
|
|
|
|
|
|
[xyseries]
|
|
|
max_mem_usage_mb = 200
|
|
|
|
|
|
|
|
|
############################################################################
|
|
|
# GENERAL
|
|
|
############################################################################
|
|
|
# This section contains the stanzas for a variety of general settings.
|
|
|
|
|
|
[auto_summarizer]
|
|
|
cache_timeout = 600
|
|
|
maintenance_period = 1800
|
|
|
return_actions_with_normalized_ids = fromcontext
|
|
|
normalized_summaries = true
|
|
|
detailed_dashboard = true
|
|
|
shc_accurate_access_counts = false
|
|
|
disable_transparent_mode_federation = false
|
|
|
|
|
|
|
|
|
[http_input]
|
|
|
# The max number of tokens reported by logging input metrics.
|
|
|
max_number_of_tokens = 10000
|
|
|
# The interval (in seconds) of logging input metrics report.
|
|
|
metrics_report_interval = 60
|
|
|
# The max request content length (800MB, to match HTTP server).
|
|
|
max_content_length = 838860800
|
|
|
# The max number of ACK channels.
|
|
|
max_number_of_ack_channel = 1000000
|
|
|
# The max number of acked requests pending query.
|
|
|
max_number_of_acked_requests_pending_query = 10000000
|
|
|
# The max number of acked requests pending query per ACK channel.
|
|
|
max_number_of_acked_requests_pending_query_per_ack_channel = 1000000
|
|
|
|
|
|
|
|
|
[indexpreview]
|
|
|
# Maximum number of bytes to read from each file during preview.
|
|
|
max_preview_bytes = 2000000
|
|
|
# Maximum number of results to emit per call to preview data generator.
|
|
|
max_results_perchunk = 2500
|
|
|
# Loosely-applied maximum on number of preview data objects held in memory.
|
|
|
soft_preview_queue_size = 100
|
|
|
|
|
|
|
|
|
[inputproc]
|
|
|
# Threshold size (in mb) to trigger fishbucket rolling to a new db.
|
|
|
file_tracking_db_threshold_mb = 500
|
|
|
# Approximate ceiling on source types & fingerprints in learned app.
|
|
|
learned_sourcetypes_limit = 1000
|
|
|
|
|
|
# Maximum size (in mb) of heap allowed to be created by Splunk modular
|
|
|
# input MonitorNoHandle.
|
|
|
monitornohandle_max_heap_mb = 0
|
|
|
|
|
|
|
|
|
[kv]
|
|
|
# When non-zero, the point at which kv should stop creating new columns.
|
|
|
maxcols = 512
|
|
|
# Maximum number of keys auto kv can generate.
|
|
|
# Set this value to 0 to not impose any limit on auto kv limit and indexed kv limit.
|
|
|
limit = 100
|
|
|
# Maximum number of key-value pairs that can be extracted at index time.
|
|
|
# Set this value to 0 to not impose any limit on indexed kv limit.
|
|
|
indexed_kv_limit = 200
|
|
|
# Truncate _raw to to this size and then do auto KV.
|
|
|
maxchars = 10240
|
|
|
|
|
|
max_extractor_time = 1000
|
|
|
avg_extractor_time = 500
|
|
|
|
|
|
|
|
|
[kvstore]
|
|
|
# The max number of accelerations that can be assigned to a single collection.
|
|
|
# Valid values range from 0 to 50
|
|
|
max_accelerations_per_collection = 10
|
|
|
# The max number of fields that can be part of an acceleration.
|
|
|
# Valid values range from 0 to 30
|
|
|
max_fields_per_acceleration = 10
|
|
|
# The max number of rows that will be returned per query.
|
|
|
max_rows_per_query = 50000
|
|
|
# The max number of queries that can be run as part of the same batch.
|
|
|
max_queries_per_batch = 1000
|
|
|
# The max size of a query result in MB.
|
|
|
max_size_per_result_mb = 50
|
|
|
# The max size of a batch save operation in MB.
|
|
|
max_size_per_batch_save_mb = 50
|
|
|
# The max number of documents of a batch save operation.
|
|
|
max_documents_per_batch_save = 1000
|
|
|
# The max size of a batched query result in MB.
|
|
|
max_size_per_batch_result_mb = 100
|
|
|
# The max number of rows in memory before flushing them to CSV projection.
|
|
|
max_rows_in_memory_per_dump = 200
|
|
|
# The max number of threads to use for outputlookup.
|
|
|
max_threads_per_outputlookup = 1
|
|
|
|
|
|
[kvstore_migration]
|
|
|
# The interval in seconds at which the status of KV Store migration or
|
|
|
# KV Store upgrade is polled for search head cluster members.
|
|
|
periodic_timer_interval = 10
|
|
|
# The maximum number of intervals that a search head cluster member's
|
|
|
# failed status can remain unchanged during KV Store migration or upgrade.
|
|
|
max_failed_status_unchanged_count = 30
|
|
|
|
|
|
[input_channels]
|
|
|
max_inactive = auto
|
|
|
lowater_inactive = auto
|
|
|
inactive_eligibility_age_seconds = 330
|
|
|
|
|
|
[ldap]
|
|
|
# Maximum number of users we will attempt to precache from LDAP after
|
|
|
# reloading auth.
|
|
|
max_users_to_precache = 1000
|
|
|
# Controls whether we allow login when we find multiple entries with the
|
|
|
# same value for the username attribute.
|
|
|
allow_multiple_matching_users = true
|
|
|
|
|
|
|
|
|
[metrics]
|
|
|
# The number of series to include in the per_x_thruput reports in
|
|
|
# metrics.log.
|
|
|
maxseries = 10
|
|
|
|
|
|
# 30 seconds metrics logging interval
|
|
|
[tcpin_connections]
|
|
|
interval = 30
|
|
|
[thruput:thruput]
|
|
|
interval = 30
|
|
|
[thruput:index_thruput]
|
|
|
interval = 30
|
|
|
[queue]
|
|
|
interval = 30
|
|
|
[dutycycle]
|
|
|
interval = 30
|
|
|
[search_concurrency]
|
|
|
interval = 30
|
|
|
[searchscheduler]
|
|
|
interval = 30
|
|
|
[executor]
|
|
|
interval = 30
|
|
|
[jobs]
|
|
|
interval = 30
|
|
|
[search_pool]
|
|
|
interval = 30
|
|
|
[smartbus]
|
|
|
interval = 30
|
|
|
[pipeline_lb_cpu-stashparsing]
|
|
|
interval = 120
|
|
|
[pipeline_agg_cpu-stashparsing]
|
|
|
interval = 120
|
|
|
[pipeline_lb_cpu-parsing]
|
|
|
interval = 120
|
|
|
[pipeline_mp_cpu-parsing]
|
|
|
interval = 120
|
|
|
[pipeline_agg_cpu-merging]
|
|
|
interval = 120
|
|
|
[pipeline_msp_cpu-typing]
|
|
|
interval = 120
|
|
|
[pipeline:regexextractionprocessor-stashparsing]
|
|
|
interval = 120
|
|
|
[pipeline:regexextractionprocessor-typing]
|
|
|
interval = 120
|
|
|
[pipeline:regexextractionprocessor-ruleset]
|
|
|
interval = 120
|
|
|
|
|
|
[metrics:tcpin_connections]
|
|
|
# Keep each connection metrics.
|
|
|
aggregate_metrics = false
|
|
|
|
|
|
# Keep _tcp_Bps, _tcp_KBps, _tcp_avg_thruput, _tcp_Kprocessed that can
|
|
|
# be derived from kb.
|
|
|
suppress_derived_info = false
|
|
|
|
|
|
|
|
|
[pdf]
|
|
|
# The max number of rows that the pdfgen rendering engine (not PDF Report
|
|
|
# Server app) will render for any individual table or event listing.
|
|
|
max_rows_per_table = 1000
|
|
|
|
|
|
# The number of seconds after which the pdfgen render endpoint will timeout
|
|
|
# if it has not yet finished rendering the PDF output.
|
|
|
render_endpoint_timeout = 3600
|
|
|
|
|
|
# The number of seconds after which the Chromium engine will timeout if the
|
|
|
# engine still needs to render the dashboard output.
|
|
|
# This setting does not impact the render_chromium_screenshot_delay.
|
|
|
render_chromium_timeout = 30
|
|
|
|
|
|
# The number of seconds after which the Chromium engine takes a screenshot
|
|
|
# of a dashboard to render before exporting the dashboard.
|
|
|
# This setting does not impact the render_chromium_timeout setting.
|
|
|
render_chromium_screenshot_delay = 0
|
|
|
|
|
|
|
|
|
[restapi]
|
|
|
# Maximum result rows to be return by /events or /results getters from
|
|
|
# REST API.
|
|
|
maxresultrows = 50000
|
|
|
|
|
|
# Regex constraint on time_format and output_time_format for search
|
|
|
# endpoints.
|
|
|
time_format_reject = [<>!]
|
|
|
|
|
|
# Truncate the properties over this length in the contents dictionary
|
|
|
# of a job entry from the jobs endpoint.
|
|
|
# 0 means don't truncate.
|
|
|
jobscontentmaxcount = 0
|
|
|
|
|
|
# Determines whether we want to hard error for REST command searches
|
|
|
restprocessor_errors_fatal = false
|
|
|
|
|
|
# The maximum number of persistent processes that EAI custom REST handlers can create to serve REST API calls in persistent mode.
|
|
|
max_persistent_connections = 3000
|
|
|
|
|
|
|
|
|
[realtime]
|
|
|
# Default options for indexer support of real-time searches.
|
|
|
# These can all be overridden for a single search via REST API arguments.
|
|
|
|
|
|
# Size of queue for each real-time search.
|
|
|
queue_size = 10000
|
|
|
|
|
|
# Should indexer block if a queue is full?
|
|
|
blocking = false
|
|
|
|
|
|
# Maximum time to block if the queue is full.
|
|
|
# Meaningless if blocking = false.
|
|
|
max_blocking_secs = 60
|
|
|
|
|
|
# Should the indexer prefilter events for efficiency?
|
|
|
indexfilter = true
|
|
|
|
|
|
# Should real-time windowed searches backfill with historical data by default?
|
|
|
default_backfill = true
|
|
|
|
|
|
# Should real-time windowed searches sort events to be in descending time order?
|
|
|
enforce_time_order = true
|
|
|
|
|
|
# Should we use indexedRealtime by default?
|
|
|
indexed_realtime_use_by_default = false
|
|
|
|
|
|
# Number of seconds to wait for disk flushes to finish with
|
|
|
# indexed/continuous/pseudo realtime search.
|
|
|
indexed_realtime_disk_sync_delay = 60
|
|
|
|
|
|
# Minimum seconds to wait between component index searches during an
|
|
|
# indexed realtime search.
|
|
|
indexed_realtime_default_span = 1
|
|
|
|
|
|
# Max number of seconds allowed to fall behind realtime before we drop data
|
|
|
# and reset back to the default span from realtime.
|
|
|
indexed_realtime_maximum_span = 0
|
|
|
|
|
|
# Frequency to fetch updated bucket list.
|
|
|
indexed_realtime_update_interval = 30
|
|
|
|
|
|
# This limits the frequency that we will trigger alerts during a
|
|
|
# realtime search.
|
|
|
alerting_period_ms = 0
|
|
|
|
|
|
[rex]
|
|
|
match_limit = 100000
|
|
|
depth_limit = 1000
|
|
|
|
|
|
[reversedns]
|
|
|
# Max percent of time allowed for reverse dns lookups for incoming
|
|
|
# forwarder connections before WARN is logged in splunkd.log.
|
|
|
# Sanity check diagnostic for slow lookups.
|
|
|
rdnsMaxDutyCycle = 10
|
|
|
|
|
|
|
|
|
[scheduler]
|
|
|
# Uses a separate thread to fetch scheduled and auto summarize saved searches
|
|
|
# asynchronously.
|
|
|
async_saved_search_fetch = true
|
|
|
# The interval at which scheduled and auto summarize saved searches
|
|
|
# will be fetched asynchronously.
|
|
|
async_saved_search_interval = 30
|
|
|
|
|
|
# The interval at which scheduled saved searches will be evaluated for
|
|
|
# admission rules asynchronously.
|
|
|
async_admission_eval_interval = 600
|
|
|
|
|
|
saved_searches_disabled = false
|
|
|
|
|
|
|
|
|
# The maximum number of searches the scheduler can run, as a percentage
|
|
|
# of the maximum number of concurrent searches.
|
|
|
max_searches_perc = 50
|
|
|
|
|
|
# Fraction of concurrent scheduler searches to use for auto summarization.
|
|
|
auto_summary_perc = 50
|
|
|
|
|
|
# Every search should run as soon possible after its next scheduled time.
|
|
|
# However, each is penalized by its average runtime thus allowing
|
|
|
# shorter-running searches to run sooner and not potentially starve.
|
|
|
#
|
|
|
# However, since many searches run in fractions of a second and the
|
|
|
# priority type is integral and based on seconds, adding a raw runtime is
|
|
|
# too small to alter the result. Therefore, we scale the runtime.
|
|
|
priority_runtime_factor = 10
|
|
|
|
|
|
# A potential issue with the priority_runtime_factor is that now
|
|
|
# longer-running searches may get starved. To balance this out, make a
|
|
|
# search's priority lower (better) the more times it has been skipped.
|
|
|
#
|
|
|
# The adjustment should be normalized by the search's period, i.e., an
|
|
|
# infrequent search that has been skipped should get a lower (better) score
|
|
|
# than a frequent search that has been skipped the same number of times.
|
|
|
#
|
|
|
# Eventually, this adjustment will outweigh any worse priority due to a long
|
|
|
# runtime. The priority_skipped_factor controls how quickly this happens.
|
|
|
priority_skipped_factor = 1
|
|
|
|
|
|
# The amount of time, in seconds, to delay retrying a scheduled search that
|
|
|
# failed to dispatch (usually due to hitting concurrency limits).
|
|
|
dispatch_retry_delay = 0
|
|
|
|
|
|
# The maximum number of minutes to defer running continuous scheduled searches
|
|
|
# while waiting for the KV Store to come up in order to load historical data.
|
|
|
# This is used to prevent gaps in continuous scheduled searches when splunkd
|
|
|
# was down.
|
|
|
#
|
|
|
# Use [<int>]<unit> to specify a duration; a missing <int> defaults to 1.
|
|
|
# Relevant units are: s, sec, second, secs, seconds, m, min, minute, mins,
|
|
|
# minutes.
|
|
|
# For example: "60s" = 60 seconds, "5m" = 5 minutes.
|
|
|
search_history_load_timeout = 2m
|
|
|
|
|
|
# The number of runtimes kept for each search that are used to calculate the
|
|
|
# historical average runtime during search prioritization.
|
|
|
search_history_max_runtimes = 10
|
|
|
|
|
|
# The maximum amount of time to run missed continuous scheduled searches for
|
|
|
# once Splunk comes back up in the event it was down.
|
|
|
#
|
|
|
# Use [<int>]<unit> to specify a duration; a missing <int> defaults to 1.
|
|
|
# Relevant units are: min, minute, mins, minutes, h, hr, hour, hrs, hours, d,
|
|
|
# day, days, w, week, weeks, mon, month, months.
|
|
|
# For example: "5m" = 5 minutes, "1h" = 1 hour.
|
|
|
#
|
|
|
# A value of 0 means no lookback.
|
|
|
max_continuous_scheduled_search_lookback = 24h
|
|
|
|
|
|
# The amount of time to "look back" when reporting introspection statistics.
|
|
|
# For example: what is the number of dispatched searches in the last 60 minutes?
|
|
|
#
|
|
|
# Use [<int>]<unit> to specify a duration; a missing <int> defaults to 1.
|
|
|
# Relevant units are: m, min, minute, mins, minutes, h, hr, hour, hrs, hours,
|
|
|
# d, day, days, w, week, weeks.
|
|
|
# For example: "5m" = 5 minutes, "1h" = 1 hour.
|
|
|
introspection_lookback = 1h
|
|
|
|
|
|
# Maximum number of results to load when triggering an action.
|
|
|
max_action_results = 50000
|
|
|
|
|
|
action_execution_threads = 10
|
|
|
|
|
|
actions_queue_size = 500
|
|
|
|
|
|
actions_queue_timeout = 30
|
|
|
|
|
|
alerts_max_count = 50000
|
|
|
|
|
|
alerts_max_history = 7d
|
|
|
|
|
|
alerts_expire_period = 120
|
|
|
|
|
|
persistence_period = 30
|
|
|
|
|
|
# Maximum number of lock files to keep around for each scheduled search.
|
|
|
# Effective only if search head pooling is enabled, the most recent files
|
|
|
# are kept.
|
|
|
max_lock_files = 5
|
|
|
|
|
|
# The lock file reaper should clean lock files that are this old (in seconds).
|
|
|
max_lock_file_ttl = 86400
|
|
|
|
|
|
max_per_result_alerts = 500
|
|
|
|
|
|
scheduled_view_timeout = 60m
|
|
|
|
|
|
# Scheduler timeout for printing a throttled warning message
|
|
|
# if we're hitting scheduler concurrency limits.
|
|
|
concurrency_message_throttle_time = 10m
|
|
|
|
|
|
# By default the scheduler should not run jobs on itself in search head
|
|
|
# pooling mode. It should dispatch to pool members.
|
|
|
shp_dispatch_to_member = true
|
|
|
# In 6.3 and beyond Search Head Clustering has implemented role quota
|
|
|
# enforcement. Set this to true to enable this feature.
|
|
|
shc_role_quota_enforcement = false
|
|
|
shc_syswide_quota_enforcement = false
|
|
|
|
|
|
|
|
|
[search_metrics]
|
|
|
# Add more detail to the per-search metrics.
|
|
|
debug_metrics = false
|
|
|
|
|
|
|
|
|
[show_source]
|
|
|
# Maximum events retrievable by show source.
|
|
|
max_count = 10000
|
|
|
max_timebefore = 1day
|
|
|
max_timeafter = 1day
|
|
|
distributed = true
|
|
|
# Maximum events we will request in the distributed show source.
|
|
|
# Likely all of these will not be used.
|
|
|
distributed_search_limit = 30000
|
|
|
|
|
|
|
|
|
[slc]
|
|
|
# Maximum number of clusters to create.
|
|
|
maxclusters = 10000
|
|
|
|
|
|
|
|
|
[slow_peer_disconnect]
|
|
|
# Settings for the heuristic that will detect and disconnect slow peers
|
|
|
# towards the end of a search that has returned a large volume of data.
|
|
|
|
|
|
disabled = true
|
|
|
# is this feature enabled.
|
|
|
# Defaults to true
|
|
|
|
|
|
batch_search_activation_fraction = 0.9
|
|
|
# The fraction of peers that must have completed before we start disconnecting.
|
|
|
# This is only applicable to batch search because the slow peers will not hold
|
|
|
# back the fast peers.
|
|
|
# Defaults to 0.9
|
|
|
|
|
|
packets_per_data_point = 500
|
|
|
# Rate statistics will be sampled once every packets_per_data_point packets.
|
|
|
|
|
|
sensitivity = 0.3
|
|
|
# Sensitivity of the heuristic to newer values. For larger values of
|
|
|
# sensitivity the heuristic will give more weight to newer statistic.
|
|
|
|
|
|
grace_period_before_disconnect = 0.10
|
|
|
# If the heuristic consistently claims that the peer is slow for at least
|
|
|
# <grace_period_before_disconnect>*life_time_of_collector seconds then only
|
|
|
# will we disconnect the peer.
|
|
|
|
|
|
threshold_data_volume = 100
|
|
|
# The volume of uncompressed data that must have accumulated in KB from
|
|
|
# a peer before we consider them in the heuristic.
|
|
|
|
|
|
threshold_connection_life_time = 5
|
|
|
# All peers will be given an initial grace period of at least these many
|
|
|
# seconds before we consider them in the heuristic.
|
|
|
|
|
|
bound_on_disconnect_threshold_as_fraction_of_mean = 0.2
|
|
|
# If network is too homogenous resulting in very low standard deviations
|
|
|
# this value may be tweaked to ensure that the thresholds we set are not
|
|
|
# too close to the mean. If threshold is an upper bound
|
|
|
# threshold >= mean*(1+bound_on_threshold) and if the threshold is a
|
|
|
# lower bound threshold <= mean*(1-bound_on_threshold).
|
|
|
# The actual threshold is computed during the search based on the mean
|
|
|
# and std. deviations of network statistics.
|
|
|
|
|
|
|
|
|
[summarize]
|
|
|
poll_buckets_until_maxtime = false
|
|
|
bucket_refresh_interval = 30
|
|
|
bucket_refresh_interval_cluster = 120
|
|
|
auto_finalize_secs_after_maxtime = 300
|
|
|
|
|
|
|
|
|
[system_checks]
|
|
|
insufficient_search_capabilities = enabled
|
|
|
orphan_searches = enabled
|
|
|
installed_files_integrity = enabled
|
|
|
installed_files_integrity_interval = 12h
|
|
|
# See limits.conf.spec file for details.
|
|
|
|
|
|
|
|
|
[thruput]
|
|
|
# Throughput limiting at index time.
|
|
|
maxKBps = 0
|
|
|
|
|
|
|
|
|
[viewstates]
|
|
|
# Is the viewstate reaper enabled?
|
|
|
enable_reaper = true
|
|
|
# How often does the reaper run?
|
|
|
reaper_freq = 86400
|
|
|
# How many viewstates does the reaper consider "acceptable"?
|
|
|
reaper_soft_warn_level = 1000
|
|
|
# Teaper eligibility age.
|
|
|
ttl = 86400
|
|
|
|
|
|
[scheduled_views]
|
|
|
enable_reaper = true
|
|
|
reaper_freq = 86400
|
|
|
|
|
|
############################################################################
|
|
|
# OPTIMIZATION
|
|
|
############################################################################
|
|
|
# This section contains global and specific optimization settings
|
|
|
|
|
|
|
|
|
[search_optimization]
|
|
|
enabled = true
|
|
|
|
|
|
[search_optimization::search_expansion]
|
|
|
enabled = true
|
|
|
|
|
|
[search_optimization::replace_append_with_union]
|
|
|
enabled = true
|
|
|
|
|
|
[search_optimization::merge_union]
|
|
|
enabled = true
|
|
|
|
|
|
[search_optimization::insert_redistribute_command]
|
|
|
enabled = true
|
|
|
|
|
|
[search_optimization::predicate_split]
|
|
|
enabled = true
|
|
|
|
|
|
[search_optimization::predicate_push]
|
|
|
enabled = true
|
|
|
|
|
|
[search_optimization::predicate_merge]
|
|
|
enabled = true
|
|
|
inputlookup_merge = true
|
|
|
merge_to_base_search = true
|
|
|
|
|
|
[search_optimization::projection_elimination]
|
|
|
enabled = true
|
|
|
|
|
|
[search_optimization::required_field_values]
|
|
|
enabled = true
|
|
|
fields = eventtype, tag
|
|
|
|
|
|
[search_optimization::search_flip_normalization]
|
|
|
enabled = true
|
|
|
|
|
|
[search_optimization::reverse_calculated_fields]
|
|
|
enabled = true
|
|
|
|
|
|
[search_optimization::search_sort_normalization]
|
|
|
enabled = true
|
|
|
|
|
|
[search_optimization::eval_merge]
|
|
|
enabled = true
|
|
|
|
|
|
[search_optimization::replace_table_with_fields]
|
|
|
enabled = true
|
|
|
|
|
|
[search_optimization::replace_stats_cmds_with_tstats]
|
|
|
enabled = true
|
|
|
detect_search_time_field_collisions = true
|
|
|
|
|
|
[search_optimization::replace_datamodel_stats_cmds_with_tstats]
|
|
|
enabled = true
|
|
|
|
|
|
[search_optimization::replace_chart_cmds_with_tstats]
|
|
|
enabled = true
|
|
|
detect_search_time_field_collisions = true
|
|
|
|
|
|
[search_optimization::set_required_fields]
|
|
|
stats = false
|
|
|
|
|
|
[directives]
|
|
|
required_tags = enabled
|
|
|
required_eventtypes = enabled
|
|
|
read_summary = enabled
|
|
|
|
|
|
[parallelreduce]
|
|
|
# The maximum number of valid indexers that can be used as intermediate
|
|
|
# reducers in the reducing phase of a parallel reduce operation.
|
|
|
maxReducersPerPhase = 20
|
|
|
# The percentage of valid indexers that can be selected from the search peers
|
|
|
# as intermediate reducers for a parallel reduce search operation.
|
|
|
winningRate = 50
|
|
|
# Timeout value setting to ensure that adequate time is provided for indexers and intermediate indexers to get paired
|
|
|
rdinPairingTimeout = 30
|
|
|
# The percentage of search queries to run as prjob in total traffic.
|
|
|
autoAppliedPercentage = 0
|
|
|
# Use parallel reduce processing to improve the performance of qualifying
|
|
|
# ad-hoc searches.
|
|
|
autoAppliedToAdhocSearches = false
|
|
|
# The maximum amount of preview cache memory usage allowed for parallel reduce
|
|
|
# search, in MB.
|
|
|
maxPreviewMemUsageMb = 100
|
|
|
# Feature flag: preview for parallel reduce
|
|
|
enablePreview = true
|
|
|
# The list of commands that will not utilize parallel reduce
|
|
|
disabledCommandList = addinfo
|
|
|
|
|
|
[rollup]
|
|
|
# Threshold on lowest rollup interval allowed
|
|
|
minSpanAllowed = 300
|
|
|
|
|
|
[mcollect]
|
|
|
# Sets the output format from mcollect to use single value format always.
|
|
|
always_use_single_value_output = true
|
|
|
|
|
|
[auth]
|
|
|
enable_install_apps = false
|
|
|
|
|
|
[segmenter]
|
|
|
use_segmenter_v2 = true
|
|
|
|
|
|
|
|
|
|
|
|
[watchdog]
|
|
|
stack_files_ttl = 7d
|
|
|
stack_files_removal_period = 1h
|
|
|
|
|
|
|
|
|
[ingest_actions]
|
|
|
rfs.provider.rawdata_limit_mb = 1024
|
|
|
rfs.provider.max_workers = 4
|
|
|
rfsS3DestinationOff = false
|
|
|
|
|
|
|
|
|
[spl2]
|
|
|
origin = all
|
|
|
|