You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

1620 lines
51 KiB

This file contains ambiguous Unicode characters!

This file contains ambiguous Unicode characters that may be confused with others in your current locale. If your use case is intentional and legitimate, you can safely ignore this warning. Use the Escape button to highlight these characters.

# Version 9.2.2.20240415
#
############################################################################
# OVERVIEW
############################################################################
# DO NOT EDIT THIS FILE!
# Changes to default files will be lost on update and are difficult to
# manage and support.
#
# Please make any changes to system defaults by overriding them in
# apps or $SPLUNK_HOME/etc/system/local
# (See "Configuration file precedence" in the web documentation).
#
# To override a specific setting, copy the name of the stanza and
# setting to the file where you wish to override it.
#
# This file configures various limits to the Splunk's search commands.
# CAUTION: Do not alter the settings in limits.conf unless you know what
# you are doing.
#
# Improperly configured limits may result in splunkd crashes and/or
# memory overuse.
#
############################################################################
# GLOBAL SETTINGS
############################################################################
[default]
max_mem_usage_mb = 200
[searchresults]
maxresultrows = 50000
# Maximum number of times to try in the atomic write operation
# (1 = no retries)
tocsv_maxretry = 5
# Retry period is 1/2 second (500 milliseconds)
tocsv_retryperiod_ms = 500
compression_level = 1
[search_info]
# These setting control logging of error messages to info.csv
# All messages will be logged to search.log regardless of these settings.
# maximum number of error messages to log in info.csv
# Set to 0 to remove limit, may affect search performance
max_infocsv_messages = 20
# log level = DEBUG | INFO | WARN | ERROR
infocsv_log_level = INFO
# Log warnings if search returns no results because user has no
# permissions to search on queried indexes.
show_warn_on_filtered_indexes = false
# Log level of messages when search returns no results because user has
# no permissions to search on queried indexes.
filteredindexes_log_level = DEBUG
[subsearch]
# Maximum number of results to return from a subsearch.
maxout = 10000
# Maximum number of seconds to run a subsearch before finalizing.
maxtime = 60
# Time to cache a given subsearch's results.
ttl = 300
############################################################################
# SEARCH COMMAND
############################################################################
# This section contains the settings for the search command.
# The settings are organized in subsections by type of setting.
[search]
############################################################################
# Batch search
############################################################################
# This section contains settings for batch search.
# Allow batch mode which searches in non-time order for certain classes
# of searches.
allow_batch_mode = true
# When batch mode attempts to retry the search on a peer that failed wait
# at least this many seconds.
batch_retry_min_interval = 5
# When batch mode attempts to retry the search on a peer that failed wait
# at most this many seconds.
batch_retry_max_interval = 300
# After a retry attempt fails increase the time to wait before trying
# again by this scaling factor.
batch_retry_scaling = 1.5
# When in batch mode what is the max number of index values to read in
# at one time.
batch_search_max_index_values = 10000000
# Number of search pipelines created per batch search
batch_search_max_pipeline = 1
# Default size of the aggregator queue to which all the search pipelines
# dump the search results on the indexer.
batch_search_max_results_aggregator_queue_size = 100000000
# Default size of the serialized results queue where all the serialized
# results are kept before transmission.
batch_search_max_serialized_results_queue_size = 100000000
############################################################################
# Bundles
############################################################################
# This section contains settings for bundles and bundle replication.
# Avoid loading remote bundles in splunkd.
load_remote_bundles = false
# Bundle replication file ttl.
replication_file_ttl = 600
# The minimum bundle replication period.
replication_period_sec = 60
# Whether bundle replication is synchronous (and thus blocking searches).
sync_bundle_replication = auto
# Bundle status expiry time
bundle_status_expiry_time = 1hr
############################################################################
# Concurrency
############################################################################
# This section contains settings for search concurrency limits.
# If 'total_search_concurrency_limit = auto', the total limit of concurrent
# historical searches is
# max_hist_searches = max_searches_per_cpu x number_of_cpus + base_max_searches.
# The maximum number of concurrent historical searches in the search head.
total_search_concurrency_limit = auto
# The base number of concurrent historical searches.
base_max_searches = 6
# Max real-time searches = max_rt_search_multiplier x max historical searches.
max_rt_search_multiplier = 1
# The maximum number of concurrent historical searches per CPU.
max_searches_per_cpu = 1
# Whether maximum number of concurrent searches are enforced cluster-wide
# for admission of adhoc searches
shc_adhoc_quota_enforcement = off
############################################################################
# Distributed search
############################################################################
# This section contains settings for distributed search connection
# information.
# Limit on the skew permitted when adding a search peer.
# Peers with a skew larger than this will be rejected.
addpeer_skew_limit = 600
# Defaults to download all remote logs other than saved search logs and
# oneshot search logs.
fetch_remote_search_log = disabledSavedSearches
# Maximum size of the chunk queue.
max_chunk_queue_size = 10000000
# Search results combiner maximum in-memory buffer size (in events).
max_combiner_memevents = 50000
# Absolute value of largest time skew we will tolerate between the search
# head and the peer (in seconds).
max_tolerable_skew = 60
# Maximum number of worker threads in Round Robin policy.
max_workers_searchparser = 5
# The minimum number of results blobs to keep for consumption by the
# search head.
results_queue_min_size = 10
# Corresponds to the size of the results queue in the dispatch fetch level
result_queue_max_size = 100000000
# If all currently active peers have finished with the search wait this
# many seconds before giving up on peers we are attempting to reconnect
# to for a retry.
results_queue_read_timeout_sec = 900
# by default we do not enable throttling
remote_search_requests_throttling_type = disabled
# By default, send user capabilities to the search peers.
remote_search_requests_send_capabilities_list = true
# By default, allow remote search execution even if the capability list is missing.
remote_search_requests_reject_if_capabilities_list_absent = false
############################################################################
# Field stats
############################################################################
# This section contains settings for field statistics.
# How often to update the field summary statistics, as a ratio to the
# elapsed run time so far.
fieldstats_update_freq = 0
# Maximum period for updating field summary statistics in seconds.
fieldstats_update_maxperiod = 60
# The minimum frequency of a field displayed in the /summary endpoint.
min_freq = 0.01
############################################################################
# History
############################################################################
# This section contains settings for search history.
# Enable search history?
enable_history = true
# Max number of searches to store in history
# for each user/app, if search_history_storage_mode is csv;
# for each user, if search_history_storage_mode is kvstore.
max_history_length = 500
# Max time search history records to store in history (for each user).
max_history_storage_retention_time = 90d
# History storage
search_history_storage_mode = csv
############################################################################
# Memory tracker
############################################################################
# This section contains settings for the memory tracker.
# If memory tracker is disabled, search won't be terminated even if it
# exceeds the memory limit.
# By default memory tracking is disabled.
enable_memory_tracker = false
# Default value for percentage memory usage for the splunk search
# process is set to 25%.
search_process_memory_usage_percentage_threshold = 25
# Default value for memory usage for the Splunk search process is set to 4GB.
search_process_memory_usage_threshold = 4000
############################################################################
# Meta search
############################################################################
# This section contains settings for meta search.
# Allow inexact metasearch?
allow_inexact_metasearch = false
############################################################################
# Misc
############################################################################
# This section contains miscellaneous search settings.
# Determines if the saved searches handler uses a removable cache
use_removable_search_cache = true
# Specifies after how long a paused search should be auto canceled,
# in seconds.
# 0 means do not auto cancel the paused search.
auto_cancel_after_pause = 0
dispatch_dir_warning_size = 5000
# Enable concatenation of successively occurring evals into a single
# comma separated eval during generation of data model searches.
enable_datamodel_meval = true
# Determines whether or not scoped conditional expansion of knowledge
# objects occurs during search string expansion. This only applies on
# the search head.
# NOTE: Do not change unless instructed to do so by Splunk Support.
enable_conditional_expansion = true
# If true, always dispatch saved searches as the requesting user.
# The default for dispatchAs in the savedsearches.conf.spec.in file
# is 'owner'.
force_saved_search_dispatch_as_user = false
# Max length of custom job id when passing spawning new job.
max_id_length = 150
# Specifies the maximum length of a generated or custom search job ID before
# the Splunk software shortens the directory name. The search job ID itself
# remains the same.
max_id_length_before_hash = 230
# Specifies whether the Splunk software reruns all or elements of a currently
# running search process when there are indexer failures in an indexer
# clustering environment.
search_retry = false
# Sets how long, in seconds, 'search_retry' waits to get updated
# indexer information.
search_retry_waiting_time = 70
# Maximum number of attempts made to retry a historical search before failing
# Only applied when search_retry is set to true
search_retry_max_historical = 15
# Stack size of the search executing thread.
stack_size = 4194304
# Use precomputed summaries if possible?
summary_mode = all
# Track indextime range of searches (shown in job inspector).
track_indextime_range = true
# By default use bloom filter.
use_bloomfilter = true
# By default use metadata elimination.
use_metadata_elimination = true
# Serialization format and compression algorithm used for search results
results_serial_format = srs
results_compression_algorithm = zstd
# Record search telemetry in search_telemetry.json in the dispatch dir
# The setting in telemetry.conf controls whether the data is sent back,
# this setting controls whether we generate search telemetry data in the
# dispatch dir. Search telemetry data is also put into _introspection.
record_search_telemetry = true
# Number of files to use as a threshold at which to stop adding more
# files to var/run/splunk/search_telemetry for indexing search telemetry
# data. This setting applies only to telemetry on the search head.
search_telemetry_file_limit = 500
# Limit, in bytes, for each of the constituent components of the search
# telemetry json representation
search_telemetry_component_limit = 10000
# track of the number of events of each sourcetype that match a search
track_matching_sourcetypes = true
# timeout to launch a search job
search_launch_timeout_seconds = 180
# timeout to initialize startup configuration, in milliseconds
search_startup_config_timeout_ms = 3000
# maximum number of tracked search result sourcetypes to add to audit.log
max_audit_sourcetypes = 100
use_search_evaluator_v2 = true
# The maximum number of field metadata displayed in the /jobs/fieldmeta endpoint.
max_fieldmeta_cnt_ui = 1000
############################################################################
# Parsing
############################################################################
# This section contains settings related to parsing searches.
# Max recursion depth for macros.
# Considered a search exception if macro expansion does not stop after
# this many levels.
max_macro_depth = 100
# Max recursion depth for subsearch.
# Considered a search exception if subsearch does not stop after
# this many levels.
max_subsearch_depth = 8
# The minimum length of a prefix before a * to ask the index about.
min_prefix_len = 1
# When true, always search the lexicon for both field::val and val for
# field=val searches, unless INDEXED=true is set for the field in
# in fields.conf (in which case only field::val is searched)
always_include_indexedfield_lispy = true
# When set to true, we will scope every indexed field=val statement
# with sourcetype and convert it to the indexed form (field::val)
indexed_fields_expansion = true
############################################################################
# Preview
############################################################################
# This section contains settings for previews.
# The maximum time to spend generating previews, as a fraction of total
# search time.
preview_duty_cycle = 0.25
############################################################################
# Quota or queued searches
############################################################################
# This section contains settings for quota or queued searches.
# Default setting for allowing async jobs to be queued if quota violation.
default_allow_queue = true
# The maximum number of times to retry to dispatch a search when the
# quota has been reached.
dispatch_quota_retry = 4
# Milliseconds between retrying to dispatch a search if a quota has been
# reached. We retry the given number of times, with each successive wait
# 2x longer than the previous.
dispatch_quota_sleep_ms = 100
# Enforce cumulative role based quotas
enable_cumulative_quota = false
# how often to retry queued jobs (in seconds)
queued_job_check_freq = 1
############################################################################
# Reading chunk controls
############################################################################
# This section contains settings for reading chunk controls.
# max_results_perchunk, min_results_perchunk, and target_time_perchunk
# are multiplied by this for a long running search.
chunk_multiplier = 5
# Time in seconds until a search is considered "long running”.
long_search_threshold = 2
# Maximum raw size of results for each call to search (in dispatch).
# 0 = no limit, not affected by chunk_multiplier.
max_rawsize_perchunk = 100000000
# Maximum results per call to search (in dispatch).
# Must be <= maxresultrows.
max_results_perchunk = 2500
# Minimum results per call to search (in dispatch).
# Must be <= max_results_perchunk.
min_results_perchunk = 100
# Target duration of a particular call to fetch search results in ms.
target_time_perchunk = 2000
############################################################################
# Real-time
############################################################################
# This section contains settings for real-time searches.
# For real-time searches in the UI, maximum number of events stored
# (as a FIFO buffer).
realtime_buffer = 10000
############################################################################
# Remote storage
############################################################################
# This section contains settings for remote storage.
# Maximum number of remote buckets to localize as a look-ahead on searches.
bucket_localize_max_lookahead = 5
# Specifies which bucket prediction algorithm to use.
bucket_predictor = consec_not_needed
############################################################################
# Results storage
############################################################################
# This section contains settings for storing final search results.
# The maximum number of end results to store globally
# (when status_buckets=0).
max_count = 500000
# By default, no timeline information is retained. UI will supply the
# status_buckets as needed.
status_buckets = 0
# Truncate report output to max_count?
truncate_report = false
# Do we write multi-file results to results_dir?
write_multifile_results_out = true
# For event searches should we read final results from the timeliner
read_final_results_from_timeliner = true
# Field filters are turned off by default in the preview release.
field_filters = false
############################################################################
# Search process
############################################################################
# This section contains settings for search process configurations.
# Even if the search process has not been idle for the above time, check its
# internal caches for stale data after this many searches.
idle_process_cache_search_count = 8
# If a search process is idle for this many seconds, take the opportunity
# to scan its internal caches for stale data.
idle_process_cache_timeout = 0.5
# Periodically we'll check if we have too many idle search processes.
# This controls how often that happens (in seconds).
idle_process_reaper_period = auto
# Inside a search process, keep up to this many compiled regex artifacts
# before checking for stale ones. Normally the above idle_process_cache_*
# settings will check for stale entries before this limit is hit.
idle_process_regex_cache_hiwater = 2500
# When running a search, scan at most this many idle processes before
# launching a new one.
launcher_max_idle_checks = auto
# Number of server threads dedicated to managing communication with
# search processes.
# Negative number means automatically pick a sensible value
launcher_threads = -1
# Maximum number of preforked search processes that are idle
# and wait for next search execution
max_idle_process_count = auto
# Memory (RSS) limit of a search process that can be idle and reusable, in KB
# Number 0 is set to use the system default, 1024*1024 KB (1GB)
# Negative number means no limit
max_idle_process_memory = auto
# The number of search processes constructing a pool to run searches, which is
# dependent on system resources (CPU and memory) available to Splunk server
# Number 0 is set to automatically pick a sensible value
# Negative number means no limit
max_search_process_pool = 2048
# When reaping idle search processes, allow one to be reaped if it is
# not using the most recent configuration bundle, and its bundle has not
# been used in at least this many seconds.
max_old_bundle_idle_time = auto
# On UNIX we can run more that one search per process.
# Set this to a number greater than one to enable.
max_searches_per_process = 500
# When running more than one search per process, limit the number of new
# searches that can be started before allowing time to service the ones
# that are already running
max_searches_started_per_cycle = 30
# When running more than one search per process, don't allow a process to
# accumulate more than this number of seconds running searches. Note that a
# search can run longer than this without being terminated, it only prevents
# the process from being used for another search
max_time_per_process = auto
# When running more than one search per process, do not reuse a process
# if it is older than this number of seconds. This is different than
# max_time_per_process because it includes time the process spent idle.
process_max_age = 7200.0
# Don't reuse a process that last served a different user unless it has
# been idle this long (in seconds).
process_min_age_before_user_change = auto
search_process_mode = auto
# Whether to increase oom_score of search processes to make splunk more stable.
search_process_configure_oom_score_adj = true
# The value added to search process, ranging between 0 and 1000.
# Only applies when 'search_process_configure_oom_score_adj' is set to true.
search_process_set_oom_score_adj = 700
############################################################################
# search_messages.log
############################################################################
# Specifies whether splunkd promotes user-facing search messages
# from $SPLUNK_HOME/var/run/splunk/dispatch/<sid>/info.csv to
# $SPLUNK_HOME/var/log/splunk/search_messages.log.
log_search_messages = true
# When 'log_search_messages = true', this setting specifies the lowest
# severity of message that splunkd logs to search_messages.log.
# Splunkd ignores all messages with a lower severity.
# Possible values in ascending order: DEBUG, INFO, WARN, ERROR
search_messages_severity = WARN
############################################################################
# Search reuse
############################################################################
# This section contains settings for search reuse.
############################################################################
# Splunk Analytics for Hadoop
############################################################################
# This section contains settings for use with Splunk Analytics for Hadoop.
# The maximum time to spend doing reduce, as a fraction of total search time.
reduce_duty_cycle = 0.25
# The frequency with which try to reduce intermediate data when there is
# a non-streaming and non-stateful streaming command. (0 = never)
reduce_freq = 10
############################################################################
# Status
############################################################################
# This section contains settings for search status.
# The number of search job metadata to cache in RAM.
status_cache_size = 10000
############################################################################
# Timelines
############################################################################
# This section contains settings for timelines.
# Size of thread pool for remote event download framework.
remote_event_download_initialize_pool = 5
remote_event_download_finalize_pool = 5
remote_event_download_local_pool = 5
# Allow timeline to be map/reduced?
remote_timeline = true
# Whether to fetch all events accessible through the timeline from the
# remote peers before the job is considered done.
remote_timeline_fetchall = 1
# Minimum number of peers required to utilize remote timelining.
remote_timeline_min_peers = 1
# How often to touch remote artifacts to keep them from being reaped
# when search has not finished? (in seconds).
remote_timeline_touchperiod = 300
# Timeouts for fetching remote timeline events.
remote_timeline_connection_timeout = 5
remote_timeline_send_timeout = 10
remote_timeline_receive_timeout = 10
# In ms
search_keepalive_frequency = 30000
# Maximum number of uninterrupted keepalives before the connection is closed.
search_keepalive_max = 100
# Enable timeline preview
timeline_events_preview = false
############################################################################
# TTL
############################################################################
# This section contains time to live (ttl) settings.
# The length of time to persist search cache entries (in seconds).
cache_ttl = 300
# How long jobs are saved for by default.
default_save_ttl = 604800
# How long searches should be stored on disk once failed.
failed_job_ttl = 86400
# How long should searches run for a search head live on the indexers.
remote_ttl = 600
# How long searches should be stored on disk once completed.
ttl = 600
# Timeout value for checking search marker files like hotbucketmarker or backfill
# marker.
check_search_marker_done_interval = 60
# Time interval of sleeping between subsequent search marker files checks.
check_search_marker_sleep_interval = 1
# How long srtemp sub-directories should be kept before they are deleted
srtemp_dir_ttl = 86400
############################################################################
# Distributed search throttling
############################################################################
# This section contains settings for distributed search throttling (peers
# side) information.
[search_throttling::per_cpu]
max_concurrent = 12
[search_throttling::physical_ram]
min_memory_per_search = 134217728
############################################################################
# OTHER COMMAND SETTINGS
############################################################################
# This section contains the stanzas for the SPL commands, except for the
# search command, which is in a separate section.
[anomalousvalue]
maxresultrows = 50000
# Maximum number of distinct values for a field.
maxvalues = 0
# Maximum size in bytes of any single value
# (truncated to this size if larger).
maxvaluesize = 0
[associate]
maxfields = 10000
maxvalues = 0
maxvaluesize = 0
[autoregress]
maxp = 10000
maxrange = 1000
[collect]
# Setting for multivalue field representations in collect command.
format_multivalue_collect = false
# Setting for collect command to add quotation marks based on major breakers.
collect_ignore_minor_breakers = false
[concurrency]
# Maximum concurrency level to keep record of.
max_count = 10000000
[correlate]
maxfields = 1000
[ctable]
# This stanza contains settings for the contingency/ctable/counttable
# command.
maxvalues = 1000
[dbinspect]
maxresultrows = 50000
[discretize]
# This stanza contains settings for the bin/bucket/discretize command.
maxbins = 50000
# if maxbins not specified or = 0, defaults to searchresults::maxresultrows
[findkeywords]
maxevents = 50000
[geostats]
# At the lowest level of the tree, i.e. ZL=0 (when we are zoomed out to
# the world level ), what is the size of each gridcell in terms of latitude
# and longitude (degrees)?
# Valid values for zl_0_gridcell_latspan are from 0 to 180.0, and
# for zl_0_gridcell_longspan are from 0 to 360.0.
# Rest of the zoom level gridcell sizes are auto-tuning, i.e. will
# reduce by a factor of 2 at each additional level.
zl_0_gridcell_latspan = 22.5
zl_0_gridcell_longspan = 45.0
# Configures the filtering/search strategy for events on the map.
# Currently experimental.
filterstrategy = 2
# How many levels of clustering will be done in geostats.
maxzoomlevel = 9
[inputcsv]
# Maximum number of retries for creating a tmp directory (with random
# name in SPLUNK_HOME/var/run/splunk).
mkdir_max_retries = 100
[join]
# The join command subsearch is restricted by two settings, 'subsearch_maxout'
# and the 'maxresultrows' setting in the [searchresults] stanza.
subsearch_maxout = 50000
subsearch_maxtime = 60
[kmeans]
maxdatapoints = 100000000
maxkvalue = 1000
maxkrange = 100
[lookup]
# Maximum size of static lookup file to use a in-memory index for.
max_memtable_bytes = 26214400
# Maximum size of static lookup file to use when using a lookup()
# eval function in the ingest context
# Defaults to 10MB
ingest_max_memtable_bytes = 10485760
# Period of time after which we should refresh in-memory lookup
# tables being used with lookup() eval function at ingest time.
ingest_lookup_refresh_period_secs = 60
# Maximum reverse lookup matches (for search expansion).
max_reverse_matches = 50
# Default setting for if non-memory file lookups (for large files)
# should batch queries.
# Can be overridden using a lookup table's stanza in transforms.conf.
batch_index_query = true
# When doing batch request, what's the most matches to retrieve?
# If more than this limit of matches would otherwise be retrieved,
# we will fall back to non-batch mode matching.
batch_response_limit = 5000000
# Maximum number of lookup error messages that should be logged.
max_lookup_messages = 20
# time to live for an indexed csv
indexed_csv_ttl = 300
# keep alive token file period
indexed_csv_keep_alive_timeout = 30
# max time for the CSV indexing
indexed_csv_inprogress_max_timeout = 300
# whether we want to error on invalid lookups or let them proceed.
input_errors_fatal = false
# Should KV Store lookups be indexed at time of bundle replication?
enable_splunkd_kv_lookup_indexing = true
[metadata]
maxresultrows = 10000
# The most metadata results to fetch from each indexer.
maxcount = 100000
bucket_localize_max_lookahead = 10
[metric_alerts]
condition_evaluation_interval = 1
search_delay = 15s+
search_ttl = 2p
honor_action = false
[msearch]
chunk_size = 1000
target_per_timeseries = 5
[mvcombine]
max_mem_usage_mb = 500
[mvexpand]
max_mem_usage_mb = 500
[outputlookup]
# Specifies if the outputlookup command should check if the user
# has write permission on the lookup file.
# The permission is set in the .meta file.
outputlookup_check_permission = false
# Specifies the context where the lookup file will be created for the first time.
create_context = app
[rare]
maxresultrows = 50000
# Maximum distinct value vectors to keep track of.
maxvalues = 0
maxvaluesize = 0
[rest]
allow_reload = false
[set]
maxresultrows = 50000
[sort]
# maximum number of concurrent files to open
maxfiles = 64
[spath]
# Number of characters to read from an XML or JSON event when
# auto extracting.
extraction_cutoff = 5000
extract_all = true
[stats]
max_keymap_rows = 1000000
maxresultrows = 50000
maxvalues = 0
maxvaluesize = 0
# For streamstats's maximum window size.
max_stream_window = 10000
# For rdigest, used to approximate order statistics (median, percentiles).
rdigest_k = 100
rdigest_maxnodes = 1
tdigest_k = 50
tdigest_max_buffer_size = 1000
tmpfile_compression = lz4
tmpfile_compression_level = 0
perc_digest_type = tdigest
list_maxsize = 100
min_chunk_size_kb = 64
max_chunk_size_kb = 4096
chunk_size_double_every = 100
# Determines whether to return results for searches with time-sensitive
# aggregations and missing or invalid timestamps in input events.
check_for_invalid_time = false
[sistats]
max_keymap_rows = 1000000
maxvalues = 0
maxvaluesize = 0
rdigest_k = 100
rdigest_maxnodes = 1
tdigest_k = 50
tdigest_max_buffer_size = 1000
perc_digest_type = tdigest
max_valuemap_bytes = 100000
[top]
maxresultrows = 50000
# Maximum distinct value vectors to keep track of.
maxvalues = 0
maxvaluesize = 0
[transactions]
# Maximum number of open transaction or events in open.
# Transaction before transaction eviction happens.
maxopentxn = 5000
maxopenevents = 100000
[tscollect]
# Default value of 'squashcase' arg if not specified by the command.
squashcase = false
# Default value of 'keepresults' arg if not specified by the command.
keepresults = false
# The max allowed size of tsidx files to create in megabytes.
# 0 implies no limit
optimize_max_size_mb = 256
[tstats]
# Whether we apply role-based search filters when users run tstats
# on normal index data (never applied on data from tscollect or
# data model acceleration).
apply_search_filter = true
# Default value of 'summariesonly' arg if not specified by the command.
summariesonly = false
# Default value of 'allow_old_summaries' arg if not specified
# by the command.
allow_old_summaries = false
# By default we retrieve up to ten million events at once from a
# TSIDX file when answering queries.
chunk_size = 10000000
# By default, do not include non-numeric values when applying
# searches that filter on numeric values.
include_events_omitted_when_filtering_numeric_values = false
# Number of search pipelines created per batch search
batch_search_max_pipeline = 1
# Should tstats use bloomfilters to eliminate buckets
use_bloomfilter = true
update_datamodel_usage_stats = true
[mstats]
time_bin_limit = 1000000
# Should mstats use bloomfilters to eliminate buckets
use_bloomfilter = true
[typeahead]
maxcount = 1000
max_servers = 2
fetch_multiplier = 50
use_cache = true
cache_ttl_sec = 300
min_prefix_length = 1
max_concurrent_per_user = 3
banned_segments =
[typer]
# In eventtyping, pay attention to first N characters of any
# attribute (e.g., _raw), including individual tokens. Can be
# overridden by supplying the typer operator with the argument.
# maxlen (e.g. "|typer maxlen=300").
maxlen = 10000
[xyseries]
max_mem_usage_mb = 200
############################################################################
# GENERAL
############################################################################
# This section contains the stanzas for a variety of general settings.
[auto_summarizer]
cache_timeout = 600
maintenance_period = 1800
return_actions_with_normalized_ids = fromcontext
normalized_summaries = true
detailed_dashboard = true
shc_accurate_access_counts = false
disable_transparent_mode_federation = false
[http_input]
# The max number of tokens reported by logging input metrics.
max_number_of_tokens = 10000
# The interval (in seconds) of logging input metrics report.
metrics_report_interval = 60
# The max request content length (800MB, to match HTTP server).
max_content_length = 838860800
# The max number of ACK channels.
max_number_of_ack_channel = 1000000
# The max number of acked requests pending query.
max_number_of_acked_requests_pending_query = 10000000
# The max number of acked requests pending query per ACK channel.
max_number_of_acked_requests_pending_query_per_ack_channel = 1000000
[indexpreview]
# Maximum number of bytes to read from each file during preview.
max_preview_bytes = 2000000
# Maximum number of results to emit per call to preview data generator.
max_results_perchunk = 2500
# Loosely-applied maximum on number of preview data objects held in memory.
soft_preview_queue_size = 100
[inputproc]
# Threshold size (in mb) to trigger fishbucket rolling to a new db.
file_tracking_db_threshold_mb = 500
# Approximate ceiling on source types & fingerprints in learned app.
learned_sourcetypes_limit = 1000
# Maximum size (in mb) of heap allowed to be created by Splunk modular
# input MonitorNoHandle.
monitornohandle_max_heap_mb = 0
[kv]
# When non-zero, the point at which kv should stop creating new columns.
maxcols = 512
# Maximum number of keys auto kv can generate.
# Set this value to 0 to not impose any limit on auto kv limit and indexed kv limit.
limit = 100
# Maximum number of key-value pairs that can be extracted at index time.
# Set this value to 0 to not impose any limit on indexed kv limit.
indexed_kv_limit = 200
# Truncate _raw to to this size and then do auto KV.
maxchars = 10240
max_extractor_time = 1000
avg_extractor_time = 500
[kvstore]
# The max number of accelerations that can be assigned to a single collection.
# Valid values range from 0 to 50
max_accelerations_per_collection = 10
# The max number of fields that can be part of an acceleration.
# Valid values range from 0 to 30
max_fields_per_acceleration = 10
# The max number of rows that will be returned per query.
max_rows_per_query = 50000
# The max number of queries that can be run as part of the same batch.
max_queries_per_batch = 1000
# The max size of a query result in MB.
max_size_per_result_mb = 50
# The max size of a batch save operation in MB.
max_size_per_batch_save_mb = 50
# The max number of documents of a batch save operation.
max_documents_per_batch_save = 1000
# The max size of a batched query result in MB.
max_size_per_batch_result_mb = 100
# The max number of rows in memory before flushing them to CSV projection.
max_rows_in_memory_per_dump = 200
# The max number of threads to use for outputlookup.
max_threads_per_outputlookup = 1
[kvstore_migration]
# The interval in seconds at which the status of KV Store migration or
# KV Store upgrade is polled for search head cluster members.
periodic_timer_interval = 10
# The maximum number of intervals that a search head cluster member's
# failed status can remain unchanged during KV Store migration or upgrade.
max_failed_status_unchanged_count = 30
[input_channels]
max_inactive = auto
lowater_inactive = auto
inactive_eligibility_age_seconds = 330
[ldap]
# Maximum number of users we will attempt to precache from LDAP after
# reloading auth.
max_users_to_precache = 1000
# Controls whether we allow login when we find multiple entries with the
# same value for the username attribute.
allow_multiple_matching_users = true
[metrics]
# The number of series to include in the per_x_thruput reports in
# metrics.log.
maxseries = 10
# 30 seconds metrics logging interval
[tcpin_connections]
interval = 30
[thruput:thruput]
interval = 30
[thruput:index_thruput]
interval = 30
[queue]
interval = 30
[dutycycle]
interval = 30
[search_concurrency]
interval = 30
[searchscheduler]
interval = 30
[executor]
interval = 30
[jobs]
interval = 30
[search_pool]
interval = 30
[smartbus]
interval = 30
[pipeline_lb_cpu-stashparsing]
interval = 120
[pipeline_agg_cpu-stashparsing]
interval = 120
[pipeline_lb_cpu-parsing]
interval = 120
[pipeline_mp_cpu-parsing]
interval = 120
[pipeline_agg_cpu-merging]
interval = 120
[pipeline_msp_cpu-typing]
interval = 120
[pipeline:regexextractionprocessor-stashparsing]
interval = 120
[pipeline:regexextractionprocessor-typing]
interval = 120
[pipeline:regexextractionprocessor-ruleset]
interval = 120
[metrics:tcpin_connections]
# Keep each connection metrics.
aggregate_metrics = false
# Keep _tcp_Bps, _tcp_KBps, _tcp_avg_thruput, _tcp_Kprocessed that can
# be derived from kb.
suppress_derived_info = false
[pdf]
# The max number of rows that the pdfgen rendering engine (not PDF Report
# Server app) will render for any individual table or event listing.
max_rows_per_table = 1000
# The number of seconds after which the pdfgen render endpoint will timeout
# if it has not yet finished rendering the PDF output.
render_endpoint_timeout = 3600
# The number of seconds after which the Chromium engine will timeout if the
# engine still needs to render the dashboard output.
# This setting does not impact the render_chromium_screenshot_delay.
render_chromium_timeout = 30
# The number of seconds after which the Chromium engine takes a screenshot
# of a dashboard to render before exporting the dashboard.
# This setting does not impact the render_chromium_timeout setting.
render_chromium_screenshot_delay = 0
[restapi]
# Maximum result rows to be return by /events or /results getters from
# REST API.
maxresultrows = 50000
# Regex constraint on time_format and output_time_format for search
# endpoints.
time_format_reject = [<>!]
# Truncate the properties over this length in the contents dictionary
# of a job entry from the jobs endpoint.
# 0 means don't truncate.
jobscontentmaxcount = 0
# Determines whether we want to hard error for REST command searches
restprocessor_errors_fatal = false
# The maximum number of persistent processes that EAI custom REST handlers can create to serve REST API calls in persistent mode.
max_persistent_connections = 3000
[realtime]
# Default options for indexer support of real-time searches.
# These can all be overridden for a single search via REST API arguments.
# Size of queue for each real-time search.
queue_size = 10000
# Should indexer block if a queue is full?
blocking = false
# Maximum time to block if the queue is full.
# Meaningless if blocking = false.
max_blocking_secs = 60
# Should the indexer prefilter events for efficiency?
indexfilter = true
# Should real-time windowed searches backfill with historical data by default?
default_backfill = true
# Should real-time windowed searches sort events to be in descending time order?
enforce_time_order = true
# Should we use indexedRealtime by default?
indexed_realtime_use_by_default = false
# Number of seconds to wait for disk flushes to finish with
# indexed/continuous/pseudo realtime search.
indexed_realtime_disk_sync_delay = 60
# Minimum seconds to wait between component index searches during an
# indexed realtime search.
indexed_realtime_default_span = 1
# Max number of seconds allowed to fall behind realtime before we drop data
# and reset back to the default span from realtime.
indexed_realtime_maximum_span = 0
# Frequency to fetch updated bucket list.
indexed_realtime_update_interval = 30
# This limits the frequency that we will trigger alerts during a
# realtime search.
alerting_period_ms = 0
[rex]
match_limit = 100000
depth_limit = 1000
[reversedns]
# Max percent of time allowed for reverse dns lookups for incoming
# forwarder connections before WARN is logged in splunkd.log.
# Sanity check diagnostic for slow lookups.
rdnsMaxDutyCycle = 10
[scheduler]
# Uses a separate thread to fetch scheduled and auto summarize saved searches
# asynchronously.
async_saved_search_fetch = true
# The interval at which scheduled and auto summarize saved searches
# will be fetched asynchronously.
async_saved_search_interval = 30
# The interval at which scheduled saved searches will be evaluated for
# admission rules asynchronously.
async_admission_eval_interval = 600
saved_searches_disabled = false
# The maximum number of searches the scheduler can run, as a percentage
# of the maximum number of concurrent searches.
max_searches_perc = 50
# Fraction of concurrent scheduler searches to use for auto summarization.
auto_summary_perc = 50
# Every search should run as soon possible after its next scheduled time.
# However, each is penalized by its average runtime thus allowing
# shorter-running searches to run sooner and not potentially starve.
#
# However, since many searches run in fractions of a second and the
# priority type is integral and based on seconds, adding a raw runtime is
# too small to alter the result. Therefore, we scale the runtime.
priority_runtime_factor = 10
# A potential issue with the priority_runtime_factor is that now
# longer-running searches may get starved. To balance this out, make a
# search's priority lower (better) the more times it has been skipped.
#
# The adjustment should be normalized by the search's period, i.e., an
# infrequent search that has been skipped should get a lower (better) score
# than a frequent search that has been skipped the same number of times.
#
# Eventually, this adjustment will outweigh any worse priority due to a long
# runtime. The priority_skipped_factor controls how quickly this happens.
priority_skipped_factor = 1
# The amount of time, in seconds, to delay retrying a scheduled search that
# failed to dispatch (usually due to hitting concurrency limits).
dispatch_retry_delay = 0
# The maximum number of minutes to defer running continuous scheduled searches
# while waiting for the KV Store to come up in order to load historical data.
# This is used to prevent gaps in continuous scheduled searches when splunkd
# was down.
#
# Use [<int>]<unit> to specify a duration; a missing <int> defaults to 1.
# Relevant units are: s, sec, second, secs, seconds, m, min, minute, mins,
# minutes.
# For example: "60s" = 60 seconds, "5m" = 5 minutes.
search_history_load_timeout = 2m
# The number of runtimes kept for each search that are used to calculate the
# historical average runtime during search prioritization.
search_history_max_runtimes = 10
# The maximum amount of time to run missed continuous scheduled searches for
# once Splunk comes back up in the event it was down.
#
# Use [<int>]<unit> to specify a duration; a missing <int> defaults to 1.
# Relevant units are: min, minute, mins, minutes, h, hr, hour, hrs, hours, d,
# day, days, w, week, weeks, mon, month, months.
# For example: "5m" = 5 minutes, "1h" = 1 hour.
#
# A value of 0 means no lookback.
max_continuous_scheduled_search_lookback = 24h
# The amount of time to "look back" when reporting introspection statistics.
# For example: what is the number of dispatched searches in the last 60 minutes?
#
# Use [<int>]<unit> to specify a duration; a missing <int> defaults to 1.
# Relevant units are: m, min, minute, mins, minutes, h, hr, hour, hrs, hours,
# d, day, days, w, week, weeks.
# For example: "5m" = 5 minutes, "1h" = 1 hour.
introspection_lookback = 1h
# Maximum number of results to load when triggering an action.
max_action_results = 50000
action_execution_threads = 10
actions_queue_size = 500
actions_queue_timeout = 30
alerts_max_count = 50000
alerts_max_history = 7d
alerts_expire_period = 120
persistence_period = 30
# Maximum number of lock files to keep around for each scheduled search.
# Effective only if search head pooling is enabled, the most recent files
# are kept.
max_lock_files = 5
# The lock file reaper should clean lock files that are this old (in seconds).
max_lock_file_ttl = 86400
max_per_result_alerts = 500
scheduled_view_timeout = 60m
# Scheduler timeout for printing a throttled warning message
# if we're hitting scheduler concurrency limits.
concurrency_message_throttle_time = 10m
# By default the scheduler should not run jobs on itself in search head
# pooling mode. It should dispatch to pool members.
shp_dispatch_to_member = true
# In 6.3 and beyond Search Head Clustering has implemented role quota
# enforcement. Set this to true to enable this feature.
shc_role_quota_enforcement = false
shc_syswide_quota_enforcement = false
[search_metrics]
# Add more detail to the per-search metrics.
debug_metrics = false
[show_source]
# Maximum events retrievable by show source.
max_count = 10000
max_timebefore = 1day
max_timeafter = 1day
distributed = true
# Maximum events we will request in the distributed show source.
# Likely all of these will not be used.
distributed_search_limit = 30000
[slc]
# Maximum number of clusters to create.
maxclusters = 10000
[slow_peer_disconnect]
# Settings for the heuristic that will detect and disconnect slow peers
# towards the end of a search that has returned a large volume of data.
disabled = true
# is this feature enabled.
# Defaults to true
batch_search_activation_fraction = 0.9
# The fraction of peers that must have completed before we start disconnecting.
# This is only applicable to batch search because the slow peers will not hold
# back the fast peers.
# Defaults to 0.9
packets_per_data_point = 500
# Rate statistics will be sampled once every packets_per_data_point packets.
sensitivity = 0.3
# Sensitivity of the heuristic to newer values. For larger values of
# sensitivity the heuristic will give more weight to newer statistic.
grace_period_before_disconnect = 0.10
# If the heuristic consistently claims that the peer is slow for at least
# <grace_period_before_disconnect>*life_time_of_collector seconds then only
# will we disconnect the peer.
threshold_data_volume = 100
# The volume of uncompressed data that must have accumulated in KB from
# a peer before we consider them in the heuristic.
threshold_connection_life_time = 5
# All peers will be given an initial grace period of at least these many
# seconds before we consider them in the heuristic.
bound_on_disconnect_threshold_as_fraction_of_mean = 0.2
# If network is too homogenous resulting in very low standard deviations
# this value may be tweaked to ensure that the thresholds we set are not
# too close to the mean. If threshold is an upper bound
# threshold >= mean*(1+bound_on_threshold) and if the threshold is a
# lower bound threshold <= mean*(1-bound_on_threshold).
# The actual threshold is computed during the search based on the mean
# and std. deviations of network statistics.
[summarize]
poll_buckets_until_maxtime = false
bucket_refresh_interval = 30
bucket_refresh_interval_cluster = 120
auto_finalize_secs_after_maxtime = 300
[system_checks]
insufficient_search_capabilities = enabled
orphan_searches = enabled
installed_files_integrity = enabled
installed_files_integrity_interval = 12h
# See limits.conf.spec file for details.
[thruput]
# Throughput limiting at index time.
maxKBps = 0
[viewstates]
# Is the viewstate reaper enabled?
enable_reaper = true
# How often does the reaper run?
reaper_freq = 86400
# How many viewstates does the reaper consider "acceptable"?
reaper_soft_warn_level = 1000
# Teaper eligibility age.
ttl = 86400
[scheduled_views]
enable_reaper = true
reaper_freq = 86400
############################################################################
# OPTIMIZATION
############################################################################
# This section contains global and specific optimization settings
[search_optimization]
enabled = true
[search_optimization::search_expansion]
enabled = true
[search_optimization::replace_append_with_union]
enabled = true
[search_optimization::merge_union]
enabled = true
[search_optimization::insert_redistribute_command]
enabled = true
[search_optimization::predicate_split]
enabled = true
[search_optimization::predicate_push]
enabled = true
[search_optimization::predicate_merge]
enabled = true
inputlookup_merge = true
merge_to_base_search = true
[search_optimization::projection_elimination]
enabled = true
[search_optimization::required_field_values]
enabled = true
fields = eventtype, tag
[search_optimization::search_flip_normalization]
enabled = true
[search_optimization::reverse_calculated_fields]
enabled = true
[search_optimization::search_sort_normalization]
enabled = true
[search_optimization::eval_merge]
enabled = true
[search_optimization::replace_table_with_fields]
enabled = true
[search_optimization::replace_stats_cmds_with_tstats]
enabled = true
detect_search_time_field_collisions = true
[search_optimization::replace_datamodel_stats_cmds_with_tstats]
enabled = true
[search_optimization::replace_chart_cmds_with_tstats]
enabled = true
detect_search_time_field_collisions = true
[search_optimization::set_required_fields]
stats = false
[directives]
required_tags = enabled
required_eventtypes = enabled
read_summary = enabled
[parallelreduce]
# The maximum number of valid indexers that can be used as intermediate
# reducers in the reducing phase of a parallel reduce operation.
maxReducersPerPhase = 20
# The percentage of valid indexers that can be selected from the search peers
# as intermediate reducers for a parallel reduce search operation.
winningRate = 50
# Timeout value setting to ensure that adequate time is provided for indexers and intermediate indexers to get paired
rdinPairingTimeout = 30
# The percentage of search queries to run as prjob in total traffic.
autoAppliedPercentage = 0
# Use parallel reduce processing to improve the performance of qualifying
# ad-hoc searches.
autoAppliedToAdhocSearches = false
# The maximum amount of preview cache memory usage allowed for parallel reduce
# search, in MB.
maxPreviewMemUsageMb = 100
# Feature flag: preview for parallel reduce
enablePreview = true
# The list of commands that will not utilize parallel reduce
disabledCommandList = addinfo
[rollup]
# Threshold on lowest rollup interval allowed
minSpanAllowed = 300
[mcollect]
# Sets the output format from mcollect to use single value format always.
always_use_single_value_output = true
[auth]
enable_install_apps = false
[segmenter]
use_segmenter_v2 = true
[watchdog]
stack_files_ttl = 7d
stack_files_removal_period = 1h
[ingest_actions]
rfs.provider.rawdata_limit_mb = 1024
rfs.provider.max_workers = 4
rfsS3DestinationOff = false
[spl2]
origin = all

Powered by BW's shoe-string budget.