-
-
Save dima-dmytruk23/68f74bc03b3c9638ca78fca0632824f5 to your computer and use it in GitHub Desktop.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
{"log":"+ '[' '' = true ']'\n","stream":"stderr","time":"2024-12-19T22:25:32.186271171Z"} | |
{"log":"+ exec /usr/bin/rpk redpanda start --overprovisioned\n","stream":"stderr","time":"2024-12-19T22:25:32.186284146Z"} | |
{"log":"WARNING: This is a setup for development purposes only; in this mode your clusters may run unrealistically fast and data can be corrupted any time your computer shuts down uncleanly.\n","stream":"stderr","time":"2024-12-19T22:25:32.210487473Z"} | |
{"log":"We'd love to hear about your experience with Redpanda:\n","stream":"stdout","time":"2024-12-19T22:25:32.21121656Z"} | |
{"log":"https://redpanda.com/feedback\n","stream":"stdout","time":"2024-12-19T22:25:32.211220107Z"} | |
{"log":"Starting redpanda...\n","stream":"stdout","time":"2024-12-19T22:25:32.211221359Z"} | |
{"log":"Running:\n","stream":"stdout","time":"2024-12-19T22:25:32.211245935Z"} | |
{"log":"/opt/redpanda/bin/redpanda redpanda --redpanda-cfg /etc/redpanda/redpanda.yaml --lock-memory=false --unsafe-bypass-fsync=true --reserve-memory=0M --overprovisioned\n","stream":"stdout","time":"2024-12-19T22:25:32.211249312Z"} | |
{"log":"\n","stream":"stdout","time":"2024-12-19T22:25:32.294960408Z"} | |
{"log":"\n","stream":"stdout","time":"2024-12-19T22:25:32.294966409Z"} | |
{"log":"Welcome to the Redpanda community!\n","stream":"stdout","time":"2024-12-19T22:25:32.294967812Z"} | |
{"log":"\n","stream":"stdout","time":"2024-12-19T22:25:32.294968954Z"} | |
{"log":"Documentation: https://docs.redpanda.com - Product documentation site\n","stream":"stdout","time":"2024-12-19T22:25:32.294970056Z"} | |
{"log":"GitHub Discussion: https://github.com/redpanda-data/redpanda/discussions - Longer, more involved discussions\n","stream":"stdout","time":"2024-12-19T22:25:32.294971489Z"} | |
{"log":"GitHub Issues: https://github.com/redpanda-data/redpanda/issues - Report and track issues with the codebase\n","stream":"stdout","time":"2024-12-19T22:25:32.294972671Z"} | |
{"log":"Support: https://support.redpanda.com - Contact the support team privately\n","stream":"stdout","time":"2024-12-19T22:25:32.294973833Z"} | |
{"log":"Product Feedback: https://redpanda.com/feedback - Let us know how we can improve your experience\n","stream":"stdout","time":"2024-12-19T22:25:32.294974986Z"} | |
{"log":"Slack: https://redpanda.com/slack - Chat about all things Redpanda. Join the conversation!\n","stream":"stdout","time":"2024-12-19T22:25:32.294976148Z"} | |
{"log":"Twitter: https://twitter.com/redpandadata - All the latest Redpanda news!\n","stream":"stdout","time":"2024-12-19T22:25:32.29497759Z"} | |
{"log":"\n","stream":"stdout","time":"2024-12-19T22:25:32.294979324Z"} | |
{"log":"\n","stream":"stdout","time":"2024-12-19T22:25:32.294980806Z"} | |
{"log":"INFO 2024-12-19 22:25:32,302 seastar - Reactor backend: linux-aio\n","stream":"stderr","time":"2024-12-19T22:25:32.302738551Z"} | |
{"log":"WARN 2024-12-19 22:25:32,305 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.30522359Z"} | |
{"log":"WARN 2024-12-19 22:25:32,305 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.305233859Z"} | |
{"log":"WARN 2024-12-19 22:25:32,305 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.305311495Z"} | |
{"log":"WARN 2024-12-19 22:25:32,305 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.305317336Z"} | |
{"log":"WARN 2024-12-19 22:25:32,305 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.305814929Z"} | |
{"log":"WARN 2024-12-19 22:25:32,305 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.305865534Z"} | |
{"log":"WARN 2024-12-19 22:25:32,306 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.306253281Z"} | |
{"log":"WARN 2024-12-19 22:25:32,306 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.306278017Z"} | |
{"log":"WARN 2024-12-19 22:25:32,306 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.306420995Z"} | |
{"log":"WARN 2024-12-19 22:25:32,306 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.306438939Z"} | |
{"log":"WARN 2024-12-19 22:25:32,306 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.306442475Z"} | |
{"log":"WARN 2024-12-19 22:25:32,306 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.306444219Z"} | |
{"log":"WARN 2024-12-19 22:25:32,306 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.306739933Z"} | |
{"log":"WARN 2024-12-19 22:25:32,306 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.30675936Z"} | |
{"log":"WARN 2024-12-19 22:25:32,306 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.307026641Z"} | |
{"log":"WARN 2024-12-19 22:25:32,307 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.307070533Z"} | |
{"log":"WARN 2024-12-19 22:25:32,307 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.307634631Z"} | |
{"log":"WARN 2024-12-19 22:25:32,307 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.307651562Z"} | |
{"log":"WARN 2024-12-19 22:25:32,307 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.307752381Z"} | |
{"log":"WARN 2024-12-19 22:25:32,307 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.307773932Z"} | |
{"log":"WARN 2024-12-19 22:25:32,308 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.308190783Z"} | |
{"log":"WARN 2024-12-19 22:25:32,308 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.30829052Z"} | |
{"log":"WARN 2024-12-19 22:25:32,308 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.309036128Z"} | |
{"log":"WARN 2024-12-19 22:25:32,309 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.309055825Z"} | |
{"log":"WARN 2024-12-19 22:25:32,309 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.309130866Z"} | |
{"log":"WARN 2024-12-19 22:25:32,309 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.309170981Z"} | |
{"log":"WARN 2024-12-19 22:25:32,310 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.310932133Z"} | |
{"log":"WARN 2024-12-19 22:25:32,310 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.311407855Z"} | |
{"log":"WARN 2024-12-19 22:25:32,310 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.311413135Z"} | |
{"log":"WARN 2024-12-19 22:25:32,310 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.311415129Z"} | |
{"log":"WARN 2024-12-19 22:25:32,311 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.311416882Z"} | |
{"log":"WARN 2024-12-19 22:25:32,311 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.311418445Z"} | |
{"log":"WARN 2024-12-19 22:25:32,311 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.311865593Z"} | |
{"log":"WARN 2024-12-19 22:25:32,311 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.311885661Z"} | |
{"log":"WARN 2024-12-19 22:25:32,312 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.312411276Z"} | |
{"log":"WARN 2024-12-19 22:25:32,312 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.31242416Z"} | |
{"log":"WARN 2024-12-19 22:25:32,312 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.312625828Z"} | |
{"log":"WARN 2024-12-19 22:25:32,312 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.312653661Z"} | |
{"log":"WARN 2024-12-19 22:25:32,312 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.312661575Z"} | |
{"log":"WARN 2024-12-19 22:25:32,312 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.31266421Z"} | |
{"log":"WARN 2024-12-19 22:25:32,312 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.312681072Z"} | |
{"log":"WARN 2024-12-19 22:25:32,312 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.312685921Z"} | |
{"log":"WARN 2024-12-19 22:25:32,312 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.312793453Z"} | |
{"log":"WARN 2024-12-19 22:25:32,312 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.312813771Z"} | |
{"log":"WARN 2024-12-19 22:25:32,313 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.313783679Z"} | |
{"log":"WARN 2024-12-19 22:25:32,313 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.313797104Z"} | |
{"log":"WARN 2024-12-19 22:25:32,313 seastar - Creation of perf_event based stall detector failed: falling back to posix timer: std::__1::system_error (error system:1, perf_event_open() failed: Operation not permitted)\n","stream":"stderr","time":"2024-12-19T22:25:32.313835356Z"} | |
{"log":"WARN 2024-12-19 22:25:32,313 cpu_profiler - Creation of perf_event based cpu profiler failed: falling back to posix timer: perf_event_open() failed: Operation not permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.313868999Z"} | |
{"log":"INFO 2024-12-19 22:25:32,318 [shard 0:main] main - application.cc:441 - Redpanda v24.2.12 - e9dc86e4946ceb2b16418c617e679d78dc1edd9c\n","stream":"stderr","time":"2024-12-19T22:25:32.318635846Z"} | |
{"log":"INFO 2024-12-19 22:25:32,318 [shard 0:main] main - application.cc:449 - kernel=6.8.0-49-generic, nodename=redpanda1, machine=x86_64\n","stream":"stderr","time":"2024-12-19T22:25:32.318646526Z"} | |
{"log":"INFO 2024-12-19 22:25:32,318 [shard 0:main] main - application.cc:364 - System resources: { cpus: 24, available memory: 3.891GiB, reserved memory: 0.000bytes}\n","stream":"stderr","time":"2024-12-19T22:25:32.318654642Z"} | |
{"log":"INFO 2024-12-19 22:25:32,318 [shard 0:main] main - application.cc:372 - File handle limit: 1048576/1048576\n","stream":"stderr","time":"2024-12-19T22:25:32.31865889Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] cluster - config_manager.cc:510 - Can't load config cache: std::__1::__fs::filesystem::filesystem_error (error system:2, filesystem error: open failed: No such file or directory [\"/var/lib/redpanda/data/config_cache.yaml\"])\n","stream":"stderr","time":"2024-12-19T22:25:32.321289812Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] cluster - config_manager.cc:450 - Can't load config bootstrap file: std::__1::__fs::filesystem::filesystem_error (error system:2, filesystem error: open failed: No such file or directory [\"/etc/redpanda/.bootstrap.yaml\"])\n","stream":"stderr","time":"2024-12-19T22:25:32.321310832Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:855 - Cluster configuration properties:\n","stream":"stderr","time":"2024-12-19T22:25:32.321564537Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:856 - (use `rpk cluster config edit` to change)\n","stream":"stderr","time":"2024-12-19T22:25:32.321568084Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.abort_index_segment_size:50000\u0009- Capacity (in number of txns) of an abort index segment\n","stream":"stderr","time":"2024-12-19T22:25:32.321777296Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.abort_timed_out_transactions_interval_ms:10000\u0009- How often look for the inactive transactions and abort them\n","stream":"stderr","time":"2024-12-19T22:25:32.321779811Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.admin_api_require_auth:0\u0009- Whether admin API clients must provide HTTP Basic authentication headers\n","stream":"stderr","time":"2024-12-19T22:25:32.321781274Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.aggregate_metrics:0\u0009- Enable aggregations of metrics returned by the prometheus '/metrics' endpoint. Metric aggregation is performed by summing the values of samples by labels. Aggregations are performed where it makes sense by the shard and/or partition labels.\n","stream":"stderr","time":"2024-12-19T22:25:32.321782696Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.alive_timeout_ms:5000\u0009- Time from the last node status heartbeat after which a node will be considered offline and not alive\n","stream":"stderr","time":"2024-12-19T22:25:32.321784219Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.alter_topic_cfg_timeout_ms:5000\u0009- Time to wait for entries replication in controller log when executing alter configuration request\n","stream":"stderr","time":"2024-12-19T22:25:32.321785652Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.append_chunk_size:16384\u0009- Size of direct write operations to disk in bytes\n","stream":"stderr","time":"2024-12-19T22:25:32.321787004Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.audit_client_max_buffer_size:16777216\u0009- Maximum number of bytes the internal audit client will allocate for audit log records. Disable and re-enable auditing for changes to take affect\n","stream":"stderr","time":"2024-12-19T22:25:32.321788317Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.audit_enabled:0\u0009- Enable/Disable audit logging.\n","stream":"stderr","time":"2024-12-19T22:25:32.32178987Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.audit_enabled_event_types:{management, authenticate, admin}\u0009- List of event classes that will be audited, options are: [management, produce, consume, describe, heartbeat, authenticate, admin, schema_registry]. Please refer to the documentation to know exactly which request(s) map to a particular audit event type.\n","stream":"stderr","time":"2024-12-19T22:25:32.321791513Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.audit_excluded_principals:{}\u0009- List of user principals to exclude from auditing\n","stream":"stderr","time":"2024-12-19T22:25:32.321793116Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.audit_excluded_topics:{}\u0009- List of topics to exclude from auditing\n","stream":"stderr","time":"2024-12-19T22:25:32.321798616Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.audit_log_num_partitions:12\u0009- Number of partitions for the internal audit log topic. Attempt to create topic is only performed if it doesn't already exist, disable and re-enable auditing for changes to take affect\n","stream":"stderr","time":"2024-12-19T22:25:32.321800069Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.audit_log_replication_factor:{nullopt}\u0009- Replication factor of the internal audit log topic. Attempt to create topic is only performed if it doesn't already exist, disable and re-enable auditing for changes to take affect. If unset, defaults to `default_topic_replication`\n","stream":"stderr","time":"2024-12-19T22:25:32.321801542Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.audit_queue_drain_interval_ms:500\u0009- Frequency in which per shard audit logs are batched to client for write to audit log. Longer intervals allow for greater change for coalescing duplicates (great for high throughput auditing scenarios) but increase the risk of data loss during hard shutdowns.\n","stream":"stderr","time":"2024-12-19T22:25:32.321803094Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.audit_queue_max_buffer_size_per_shard:1048576\u0009- Maximum amount of memory allowed in the audit buffer per shard Once this value is reached, any request handlers that cannot enqueue audit messages will return a non retryable error to the client. Note that this only will occur when handling requests that are currently enabled for auditing.\n","stream":"stderr","time":"2024-12-19T22:25:32.321804657Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.auto_create_topics_enabled:1\u0009- Allow topic auto creation\n","stream":"stderr","time":"2024-12-19T22:25:32.321806411Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_access_key:{nullopt}\u0009- AWS access key\n","stream":"stderr","time":"2024-12-19T22:25:32.321807743Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_api_endpoint:{nullopt}\u0009- Optional API endpoint\n","stream":"stderr","time":"2024-12-19T22:25:32.321809116Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_api_endpoint_port:443\u0009- TLS port override\n","stream":"stderr","time":"2024-12-19T22:25:32.321810438Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_attempt_cluster_restore_on_bootstrap:0\u0009- If set to `true`, when a cluster is started for the first time and there is cluster metadata in the configured cloud storage bucket, Redpanda automatically starts a cluster restore from that metadata. If using an automated method for deployment where it's not easy to predictably determine that a restore is needed, we recommend setting to `true`. Take care to ensure that in such deployments, a cluster bootstrap with a given bucket means that any previous cluster using that bucket is fully destroyed; otherwise tiered storage subsystems may interfere with each other.\n","stream":"stderr","time":"2024-12-19T22:25:32.321813504Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_azure_adls_endpoint:{nullopt}\u0009- Azure Data Lake Storage v2 endpoint override. Use when Hierarchical Namespaces are enabled on your storage account and you have set up a custom endpoint.\n","stream":"stderr","time":"2024-12-19T22:25:32.321815488Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_azure_adls_port:{nullopt}\u0009- Azure Data Lake Storage v2 port override. Also see cloud_storage_azure_adls_endpoint.\n","stream":"stderr","time":"2024-12-19T22:25:32.32181697Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_azure_container:{nullopt}\u0009- The name of the Azure container to use with Tiered Storage. Note that the container must belong to 'cloud_storage_azure_storage_account'\n","stream":"stderr","time":"2024-12-19T22:25:32.321820227Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_azure_hierarchical_namespace_enabled:{nullopt}\u0009- Force Redpanda to use or not use an \"Azure Data Lake Storage Gen2 hierarchical namespace\" compliant client. When this property is not set, each node infers at startup if HNS is enabled. When set to True, this property disables the check and assumes HNS is enabled. When set to False, this property disables the check and assumes HNS is not enabled. This setting should be used only in emergencies where Redpanda fails to detect the correct HNS status.\n","stream":"stderr","time":"2024-12-19T22:25:32.32182196Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_azure_managed_identity_id:{nullopt}\u0009- The managed identity ID to use with Azure Managed Identities. This takes affect when the cloud_storage_credential_source configuration option is set to azure_vm_instance_metadata.\n","stream":"stderr","time":"2024-12-19T22:25:32.321823793Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_azure_shared_key:{nullopt}\u0009- The shared key to be used for Azure Shared Key authentication with the configured Azure storage account (see 'cloud_storage_azure_storage_account)'. Note that Redpanda expects this string to be Base64 encoded.\n","stream":"stderr","time":"2024-12-19T22:25:32.321825296Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_azure_storage_account:{nullopt}\u0009- The name of the Azure storage account to use with Tiered Storage\n","stream":"stderr","time":"2024-12-19T22:25:32.321826769Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_backend:unknown\u0009- Optional cloud storage backend variant used to select API capabilities. If not supplied, will be inferred from other configuration parameters.\n","stream":"stderr","time":"2024-12-19T22:25:32.321828182Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_background_jobs_quota:5000\u0009- The number of total requests that the cloud storage background jobs are allowed to make during one background housekeeping run. This is a per shard limit.\n","stream":"stderr","time":"2024-12-19T22:25:32.321829564Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_bucket:{nullopt}\u0009- AWS bucket that should be used to store data\n","stream":"stderr","time":"2024-12-19T22:25:32.321831127Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_cache_check_interval:5000\u0009- Minimum time between trims of tiered storage cache. If a fetch operation requires trimming the cache, and the most recent trim was within this period, then trimming will be delayed until this period has elapsed\n","stream":"stderr","time":"2024-12-19T22:25:32.32183246Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_cache_chunk_size:16777216\u0009- Size of chunks of segments downloaded into cloud storage cache. Reduces space usage by only downloading the necessary chunk from a segment.\n","stream":"stderr","time":"2024-12-19T22:25:32.321833982Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_cache_max_objects:100000\u0009- Maximum number of objects that may be held in the tiered storage cache. This applies simultaneously with `cloud_storage_cache_size`, and which ever limit is hit first will drive trimming of the cache.\n","stream":"stderr","time":"2024-12-19T22:25:32.321835395Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_cache_num_buckets:0\u0009- Divide cloud storage cache across specified number of buckets. This only works for objects with randomized prefixes. The names will not be changed if the value is set to zero.\n","stream":"stderr","time":"2024-12-19T22:25:32.321840725Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_cache_size:0\u0009- Max size of archival cache\n","stream":"stderr","time":"2024-12-19T22:25:32.321842228Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_cache_size_percent:{20}\u0009- The maximum size of the archival cache as a percentage of unreserved disk space (see disk_reservation_percent). The default value for this option is tuned for a shared disk configuration. When using a dedicated cache disk consider increasing the value.\n","stream":"stderr","time":"2024-12-19T22:25:32.32184354Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_cache_trim_carryover_bytes:0\u0009- The cache performs a recursive directory inspection during the cache trim. The information obtained during the inspection can be carried over to the next trim operation. This parameter sets a limit on the memory occupied by objects that can be carried over from one trim to next, and allows cache to quickly unblock readers before starting the directory inspection (deprecated)\n","stream":"stderr","time":"2024-12-19T22:25:32.321845254Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_cache_trim_threshold_percent_objects:{nullopt}\u0009- Trim is triggered when the cache reaches this percent of the maximum object count. If this is unset, the default behavioris to start trim when the cache is about 100% full.\n","stream":"stderr","time":"2024-12-19T22:25:32.321846907Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_cache_trim_threshold_percent_size:{nullopt}\u0009- Trim is triggered when the cache reaches this percent of the maximum cache size. If this is unset, the default behavioris to start trim when the cache is about 100% full.\n","stream":"stderr","time":"2024-12-19T22:25:32.321848389Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_cache_trim_walk_concurrency:1\u0009- The maximum number of concurrent tasks launched for directory walk during cache trimming. A higher number allows cache trimming to run faster but can cause latency spikes due to increased pressure on I/O subsystem and syscall threads.\n","stream":"stderr","time":"2024-12-19T22:25:32.321849872Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_chunk_eviction_strategy:eager\u0009- Selects a strategy for evicting unused cache chunks.\n","stream":"stderr","time":"2024-12-19T22:25:32.321851395Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_chunk_prefetch:0\u0009- Number of chunks to prefetch ahead of every downloaded chunk\n","stream":"stderr","time":"2024-12-19T22:25:32.321852728Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_cluster_metadata_num_consumer_groups_per_upload:1000\u0009- Number of groups to upload in a single snapshot object during consumer offsets upload. Setting a lower value will mean a larger number of smaller snapshots are uploaded.\n","stream":"stderr","time":"2024-12-19T22:25:32.32185406Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_cluster_metadata_retries:5\u0009- Number of attempts metadata operations may be retried.\n","stream":"stderr","time":"2024-12-19T22:25:32.321855703Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_cluster_metadata_upload_interval_ms:3600000\u0009- Time interval to wait between cluster metadata uploads.\n","stream":"stderr","time":"2024-12-19T22:25:32.321858929Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_cluster_metadata_upload_timeout_ms:60000\u0009- Timeout for cluster metadata uploads.\n","stream":"stderr","time":"2024-12-19T22:25:32.321860342Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_credentials_host:{nullopt}\u0009- The hostname to connect to for retrieving role based credentials. Derived from cloud_storage_credentials_source if not set. Only required when using IAM role based access.\n","stream":"stderr","time":"2024-12-19T22:25:32.321861684Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_credentials_source:config_file\u0009- The source of credentials to connect to cloud services\n","stream":"stderr","time":"2024-12-19T22:25:32.321863177Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_crl_file:{nullopt}\u0009- Path to certificate revocation list for cloud_storage_trust_file.\n","stream":"stderr","time":"2024-12-19T22:25:32.321866563Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_disable_archiver_manager:1\u0009- Use legacy upload mode and do not start archiver_manager.\n","stream":"stderr","time":"2024-12-19T22:25:32.321868116Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_disable_chunk_reads:0\u0009- Disable chunk reads and switch back to legacy mode where full segments are downloaded.\n","stream":"stderr","time":"2024-12-19T22:25:32.32186995Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_disable_metadata_consistency_checks:1\u0009- Disable all metadata consistency checks. This will allow redpanda to replay logs with inconsistent tiered-storage metadata. Normally, this option should be disabled.\n","stream":"stderr","time":"2024-12-19T22:25:32.321871873Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_disable_read_replica_loop_for_tests:0\u0009- Begins the read replica sync loop in tiered-storage-enabled topic partitions. The property exists to simplify testing and shouldn't be set in production.\n","stream":"stderr","time":"2024-12-19T22:25:32.321874278Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_disable_remote_labels_for_tests:0\u0009- If 'true', Redpanda disables remote labels and falls back on the hash-based object naming scheme for new topics. This property exists to simplify testing and shouldn't be set in production.\n","stream":"stderr","time":"2024-12-19T22:25:32.321876252Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_disable_tls:0\u0009- Disable TLS for all S3 connections\n","stream":"stderr","time":"2024-12-19T22:25:32.321878065Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_disable_upload_consistency_checks:0\u0009- Disable all upload consistency checks. This will allow redpanda to upload logs with gaps and replicate metadata with consistency violations. Normally, this options should be disabled.\n","stream":"stderr","time":"2024-12-19T22:25:32.321879658Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_disable_upload_loop_for_tests:0\u0009- Begins the upload loop in tiered-storage-enabled topic partitions. The property exists to simplify testing and shouldn't be set in production.\n","stream":"stderr","time":"2024-12-19T22:25:32.321881441Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_enable_compacted_topic_reupload:1\u0009- Enable re-uploading data for compacted topics\n","stream":"stderr","time":"2024-12-19T22:25:32.321885359Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_enable_remote_read:0\u0009- Default remote read config value for new topics\n","stream":"stderr","time":"2024-12-19T22:25:32.321887172Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_enable_remote_write:0\u0009- Default remote write value for new topics\n","stream":"stderr","time":"2024-12-19T22:25:32.321888815Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_enable_scrubbing:0\u0009- Enable scrubbing of cloud storage partitions. The scrubber validates the integrity of data and metadata uploaded to cloud storage\n","stream":"stderr","time":"2024-12-19T22:25:32.321890548Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_enable_segment_merging:1\u0009- Enables adjacent segment merging. The segments are reuploaded if there is an opportunity for that and if it will improve the tiered-storage performance\n","stream":"stderr","time":"2024-12-19T22:25:32.321892662Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_enabled:0\u0009- Enable archival storage\n","stream":"stderr","time":"2024-12-19T22:25:32.321894616Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_full_scrub_interval_ms:43200000\u0009- Time interval between a final scrub and thte next scrub\n","stream":"stderr","time":"2024-12-19T22:25:32.32189653Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_garbage_collect_timeout_ms:30000\u0009- Timeout for running the cloud storage garbage collection (ms)\n","stream":"stderr","time":"2024-12-19T22:25:32.321898433Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_graceful_transfer_timeout_ms:{5000}\u0009- Time limit on waiting for uploads to complete before a leadership transfer. If this is null, leadership transfers will proceed without waiting.\n","stream":"stderr","time":"2024-12-19T22:25:32.321899856Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_housekeeping_interval_ms:300000\u0009- Interval for cloud storage housekeeping tasks\n","stream":"stderr","time":"2024-12-19T22:25:32.321901279Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_hydrated_chunks_per_segment_ratio:0.7\u0009- The maximum number of chunks per segment that can be hydrated at a time. Above this number, unused chunks will be trimmed.\n","stream":"stderr","time":"2024-12-19T22:25:32.321902601Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_hydration_timeout_ms:600000\u0009- Duration to wait for a hydration request to be fulfilled, if hydration is not completed within this time, the consumer will be notified with a timeout error.\n","stream":"stderr","time":"2024-12-19T22:25:32.321904024Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_idle_threshold_rps:10\u0009- The cloud storage request rate threshold for idle state detection. If the average request rate for the configured period is lower than this threshold the cloud storage is considered being idle.\n","stream":"stderr","time":"2024-12-19T22:25:32.321905446Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_idle_timeout_ms:10000\u0009- Timeout used to detect idle state of the cloud storage API. If the average cloud storage request rate is below this threshold for a configured amount of time the cloud storage is considered idle and the housekeeping jobs are started.\n","stream":"stderr","time":"2024-12-19T22:25:32.321907059Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_initial_backoff_ms:100\u0009- Initial backoff time for exponential backoff algorithm (ms)\n","stream":"stderr","time":"2024-12-19T22:25:32.321910416Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_manifest_cache_size:1048576\u0009- Amount of memory that can be used to handle tiered-storage metadata\n","stream":"stderr","time":"2024-12-19T22:25:32.321911808Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_manifest_max_upload_interval_sec:{60000}\u0009- Wait at least this long between partition manifest uploads. Actual time between uploads may be greater than this interval. If this is null, metadata will be updated after each segment upload.\n","stream":"stderr","time":"2024-12-19T22:25:32.321913171Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_manifest_upload_timeout_ms:10000\u0009- Manifest upload timeout (ms)\n","stream":"stderr","time":"2024-12-19T22:25:32.321914654Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_materialized_manifest_ttl_ms:10000\u0009- The time interval that determines how long the materialized manifest can stay in cache under contention. This parameter is used for performance tuning. When the spillover manifest is materialized and stored in cache and the cache needs to evict it it will use 'cloud_storage_materialized_manifest_ttl_ms' value as a timeout. The cursor that uses the spillover manifest uses this value as a TTL interval after which it stops referencing the manifest making it available for eviction. This only affects spillover manifests under contention.\n","stream":"stderr","time":"2024-12-19T22:25:32.321915986Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_max_concurrent_hydrations_per_shard:{nullopt}\u0009- Maximum concurrent segment hydrations of remote data per CPU core. If unset, value of `cloud_storage_max_connections / 2` is used, which means that half of available S3 bandwidth could be used to download data from S3. If the cloud storage cache is empty every new segment reader will require a download. This will lead to 1:1 mapping between number of partitions scanned by the fetch request and number of parallel downloads. If this value is too large the downloads can affect other workloads. In case of any problem caused by the tiered-storage reads this value can be lowered. This will only affect segment hydrations (downloads) but won't affect cached segments. If fetch request is reading from the tiered-storage cache its concurrency will only be limited by available memory.\n","stream":"stderr","time":"2024-12-19T22:25:32.321927377Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_max_connection_idle_time_ms:5000\u0009- Max https connection idle time (ms)\n","stream":"stderr","time":"2024-12-19T22:25:32.321929802Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_max_connections:20\u0009- Max number of simultaneous connections to S3 per shard (includes connections used for both uploads and downloads)\n","stream":"stderr","time":"2024-12-19T22:25:32.321931155Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_max_materialized_segments_per_shard:{nullopt}\u0009- Maximum concurrent readers of remote data per CPU core. If unset, value of `topic_partitions_per_shard` multiplied by 2 is used.\n","stream":"stderr","time":"2024-12-19T22:25:32.321932577Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_max_partition_readers_per_shard:{nullopt}\u0009- Maximum partition readers per shard (deprecated)\n","stream":"stderr","time":"2024-12-19T22:25:32.321935954Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_max_segment_readers_per_shard:{nullopt}\u0009- Maximum concurrent I/O cursors of materialized remote segments per CPU core. If unset, value of `topic_partitions_per_shard` is used, i.e. one segment reader per partition if the shard is at its maximum partition capacity. These readers are cachedacross Kafka consume requests and store a readahead buffer.\n","stream":"stderr","time":"2024-12-19T22:25:32.321937376Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_max_segments_pending_deletion_per_partition:5000\u0009- The per-partition limit for the number of segments pending deletion from the cloud. Segments can be deleted due to retention or compaction. If this limit is breached and deletion fails, then segments will be orphaned in the cloud and will have to be removed manually\n","stream":"stderr","time":"2024-12-19T22:25:32.321939109Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_max_throughput_per_shard:{1073741824}\u0009- Max throughput used by tiered-storage per shard in bytes per second. This value is an upper bound of the throughput available to the tiered-storage subsystem. This parameter is intended to be used as a safeguard and in tests when we need to set precise throughput value independent of actual storage media. Please use 'cloud_storage_throughput_limit_percent' instead of this parameter in the production environment.\n","stream":"stderr","time":"2024-12-19T22:25:32.321940692Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_metadata_sync_timeout_ms:10000\u0009- Timeout for SI metadata synchronization\n","stream":"stderr","time":"2024-12-19T22:25:32.321942396Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_min_chunks_per_segment_threshold:5\u0009- The minimum number of chunks per segment for trimming to be enabled. If the number of chunks in a segment is below this threshold, the segment is small enough that all chunks in it can be hydrated at any given time\n","stream":"stderr","time":"2024-12-19T22:25:32.321943718Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_partial_scrub_interval_ms:3600000\u0009- Time interval between two partial scrubs of the same partition\n","stream":"stderr","time":"2024-12-19T22:25:32.321945241Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_readreplica_manifest_sync_timeout_ms:30000\u0009- Timeout to check if new data is available for partition in S3 for read replica\n","stream":"stderr","time":"2024-12-19T22:25:32.321946614Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_reconciliation_interval_ms:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.321947976Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_recovery_temporary_retention_bytes_default:1073741824\u0009- Retention in bytes for topics created during automated recovery\n","stream":"stderr","time":"2024-12-19T22:25:32.321949379Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_recovery_topic_validation_depth:10\u0009- Number of segment metadata to validate, from newest to oldest, when `cloud_storage_recovery_topic_validation_mode` is `check_manifest_and_segment_metadata`\n","stream":"stderr","time":"2024-12-19T22:25:32.321950791Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_recovery_topic_validation_mode:check_manifest_existence\u0009- Validation mode performed before recovering a topic from cloud storage\n","stream":"stderr","time":"2024-12-19T22:25:32.321952234Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_region:{nullopt}\u0009- AWS region that houses the bucket used for storage\n","stream":"stderr","time":"2024-12-19T22:25:32.32195557Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_roles_operation_timeout_ms:30000\u0009- Timeout for IAM role related operations (ms)\n","stream":"stderr","time":"2024-12-19T22:25:32.321956953Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_scrubbing_interval_jitter_ms:600000\u0009- Jitter applied to the cloud storage scrubbing interval.\n","stream":"stderr","time":"2024-12-19T22:25:32.321958305Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_secret_key:{nullopt}\u0009- AWS secret key\n","stream":"stderr","time":"2024-12-19T22:25:32.321959648Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_segment_max_upload_interval_sec:{3600000}\u0009- Time that segment can be kept locally without uploading it to the remote storage (sec)\n","stream":"stderr","time":"2024-12-19T22:25:32.32196094Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_segment_size_min:{nullopt}\u0009- Smallest acceptable segment size in the cloud storage. Default: cloud_storage_segment_size_target/2\n","stream":"stderr","time":"2024-12-19T22:25:32.321962313Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_segment_size_target:{nullopt}\u0009- Desired segment size in the cloud storage. Default: segment.bytes\n","stream":"stderr","time":"2024-12-19T22:25:32.321963686Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_segment_upload_timeout_ms:30000\u0009- Log segment upload timeout (ms)\n","stream":"stderr","time":"2024-12-19T22:25:32.321965118Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_spillover_manifest_max_segments:{nullopt}\u0009- Maximum number of elements in the spillover manifest that can be offloaded to the cloud storage. This property is similar to 'cloud_storage_spillover_manifest_size' but it triggers spillover based on number of segments instead of the size of the manifest in bytes. The property exists to simplify testing and shouldn't be set in the production environment\n","stream":"stderr","time":"2024-12-19T22:25:32.321966451Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_spillover_manifest_size:{65536}\u0009- The size of the manifest which can be offloaded to the cloud. If the size of the local manifest stored in redpanda exceeds cloud_storage_spillover_manifest_size x2 the spillover mechanism will split the manifest into two parts and one of them will be uploaded to S3.\n","stream":"stderr","time":"2024-12-19T22:25:32.321968254Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_throughput_limit_percent:{50}\u0009- Max throughput used by tiered-storage per node expressed as a percentage of the disk bandwidth. If the server has several disks Redpanda will take into account only the one which is used to store tiered-storage cache. Note that even if the tiered-storage is allowed to use full bandwidth of the disk (100%) it won't necessary use it in full. The actual usage depend on your workload and the state of the tiered-storage cache. This parameter is a safeguard that prevents tiered-storage from using too many system resources and not a performance tuning knob.\n","stream":"stderr","time":"2024-12-19T22:25:32.321969937Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_topic_purge_grace_period_ms:30000\u0009- Grace period during which the purger will refuse to purge the topic.\n","stream":"stderr","time":"2024-12-19T22:25:32.321973474Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_trust_file:{nullopt}\u0009- Path to certificate that should be used to validate server certificate during TLS handshake\n","stream":"stderr","time":"2024-12-19T22:25:32.321974907Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_upload_ctrl_d_coeff:0\u0009- derivative coefficient for upload PID controller.\n","stream":"stderr","time":"2024-12-19T22:25:32.321976389Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_upload_ctrl_max_shares:1000\u0009- maximum number of IO and CPU shares that archival upload can use\n","stream":"stderr","time":"2024-12-19T22:25:32.321977722Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_upload_ctrl_min_shares:100\u0009- minimum number of IO and CPU shares that archival upload can use\n","stream":"stderr","time":"2024-12-19T22:25:32.321979054Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_upload_ctrl_p_coeff:-2\u0009- proportional coefficient for upload PID controller\n","stream":"stderr","time":"2024-12-19T22:25:32.321980357Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_upload_ctrl_update_interval_ms:60000\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.321981659Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_upload_loop_initial_backoff_ms:100\u0009- Initial backoff interval when there is nothing to upload for a partition (ms)\n","stream":"stderr","time":"2024-12-19T22:25:32.321982962Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_upload_loop_max_backoff_ms:10000\u0009- Max backoff interval when there is nothing to upload for a partition (ms)\n","stream":"stderr","time":"2024-12-19T22:25:32.321984324Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_url_style:{nullopt}\u0009- Specifies the addressing style to use for Amazon S3 requests. This configuration determines how S3 bucket URLs are formatted. You can choose between: `virtual_host`, (e.g. `\u003cbucket-name\u003e.s3.amazonaws.com`), `path`, (e.g. `s3.amazonaws.com/\u003cbucket-name\u003e`), and `null`. Path style is supported for backward compatibility with legacy systems. When this property is not set (`null`), the client tries to use `virtual_host` addressing. If the initial request fails, the client automatically tries the `path` style. If neither addressing style works, Redpanda terminates the startup, requiring manual configuration to proceed.\n","stream":"stderr","time":"2024-12-19T22:25:32.321985747Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cluster_id:{nullopt}\u0009- Cluster identifier\n","stream":"stderr","time":"2024-12-19T22:25:32.321988051Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.compacted_log_segment_size:268435456\u0009- How large in bytes should each compacted log segment be (default 256MiB)\n","stream":"stderr","time":"2024-12-19T22:25:32.321989374Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.compaction_ctrl_backlog_size:{nullopt}\u0009- target backlog size for compaction controller. if not set compaction target compaction backlog would be equal to \n","stream":"stderr","time":"2024-12-19T22:25:32.321990736Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.compaction_ctrl_d_coeff:0.2\u0009- derivative coefficient for compaction PID controller.\n","stream":"stderr","time":"2024-12-19T22:25:32.321992129Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.compaction_ctrl_i_coeff:0\u0009- integral coefficient for compaction PID controller.\n","stream":"stderr","time":"2024-12-19T22:25:32.321995145Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.compaction_ctrl_max_shares:1000\u0009- maximum number of IO and CPU shares that compaction process can use\n","stream":"stderr","time":"2024-12-19T22:25:32.321996557Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.compaction_ctrl_min_shares:10\u0009- minimum number of IO and CPU shares that compaction process can use\n","stream":"stderr","time":"2024-12-19T22:25:32.32199792Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.compaction_ctrl_p_coeff:-12.5\u0009- proportional coefficient for compaction PID controller. This has to be negative since compaction backlog should decrease when number of compaction shares increases\n","stream":"stderr","time":"2024-12-19T22:25:32.321999262Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.compaction_ctrl_update_interval_ms:30000\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.322000675Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.controller_backend_housekeeping_interval_ms:1000\u0009- Interval between iterations of controller backend housekeeping loop\n","stream":"stderr","time":"2024-12-19T22:25:32.322001947Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.controller_log_accummulation_rps_capacity_acls_and_users_operations:{nullopt}\u0009- Maximum capacity of rate limit accumulationin controller acls and users operations limit\n","stream":"stderr","time":"2024-12-19T22:25:32.32200345Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.controller_log_accummulation_rps_capacity_configuration_operations:{nullopt}\u0009- Maximum capacity of rate limit accumulationin controller configuration operations limit\n","stream":"stderr","time":"2024-12-19T22:25:32.322004883Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.controller_log_accummulation_rps_capacity_move_operations:{nullopt}\u0009- Maximum capacity of rate limit accumulationin controller move operations limit\n","stream":"stderr","time":"2024-12-19T22:25:32.322006255Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.controller_log_accummulation_rps_capacity_node_management_operations:{nullopt}\u0009- Maximum capacity of rate limit accumulationin controller node management operations limit\n","stream":"stderr","time":"2024-12-19T22:25:32.322007688Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.controller_log_accummulation_rps_capacity_topic_operations:{nullopt}\u0009- Maximum capacity of rate limit accumulationin controller topic operations limit\n","stream":"stderr","time":"2024-12-19T22:25:32.322009061Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.controller_snapshot_max_age_sec:60000\u0009- Max time that will pass before we make an attempt to create a controller snapshot, after a new controller command appears\n","stream":"stderr","time":"2024-12-19T22:25:32.322010433Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.coproc_max_batch_size:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.322011856Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.coproc_max_inflight_bytes:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.322013108Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.coproc_max_ingest_bytes:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.322014351Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.coproc_offset_flush_interval_ms:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.322017356Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.core_balancing_continuous:0\u0009- If set to 'true', move partitions between cores in runtime to maintain balanced partition distribution.\n","stream":"stderr","time":"2024-12-19T22:25:32.322018689Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.core_balancing_debounce_timeout:10000\u0009- Interval, in milliseconds, between trigger and invocation of core balancing.\n","stream":"stderr","time":"2024-12-19T22:25:32.322020192Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.core_balancing_on_core_count_change:1\u0009- If set to 'true', and if after a restart the number of cores changes, Redpanda will move partitions between cores to maintain balanced partition distribution.\n","stream":"stderr","time":"2024-12-19T22:25:32.322021554Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cpu_profiler_enabled:0\u0009- Enables cpu profiling for Redpanda\n","stream":"stderr","time":"2024-12-19T22:25:32.322023027Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.cpu_profiler_sample_period_ms:100\u0009- The sample period for the CPU profiler\n","stream":"stderr","time":"2024-12-19T22:25:32.322024309Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.create_topic_timeout_ms:2000\u0009- Timeout (ms) to wait for new topic creation\n","stream":"stderr","time":"2024-12-19T22:25:32.322025602Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.data_transforms_binary_max_size:10485760\u0009- The maximum size for a deployable WebAssembly binary that the broker can store.\n","stream":"stderr","time":"2024-12-19T22:25:32.322026914Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.data_transforms_commit_interval_ms:3000\u0009- The interval at which Data Transforms commits progress.\n","stream":"stderr","time":"2024-12-19T22:25:32.322028267Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.data_transforms_enabled:0\u0009- Enables WebAssembly powered Data Transforms directly in the broker\n","stream":"stderr","time":"2024-12-19T22:25:32.322029639Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.data_transforms_logging_buffer_capacity_bytes:512000\u0009- Buffer capacity for transform logs, per shard. Buffer occupancy is calculated as the total size of buffered (i.e. emitted but not yet produced) log messages.\n","stream":"stderr","time":"2024-12-19T22:25:32.322030982Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.data_transforms_logging_flush_interval_ms:500\u0009- Flush interval for transform logs. When a timer expires, pending logs are collected and published to the transform_logs topic.\n","stream":"stderr","time":"2024-12-19T22:25:32.322032414Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.data_transforms_logging_line_max_bytes:1024\u0009- Transform log lines will be truncate to this length. Truncation occurs after any character escaping.\n","stream":"stderr","time":"2024-12-19T22:25:32.322033927Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.data_transforms_per_core_memory_reservation:20971520\u0009- The amount of memory to reserve per core for Data Transform WebAssembly Virtual Machines. Memory is reserved on boot. The maximum number of functions that can be deployed to a cluster is equal to data_transforms_per_core_memory_reservation / data_transforms_per_function_memory_limit\n","stream":"stderr","time":"2024-12-19T22:25:32.32203533Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.data_transforms_per_function_memory_limit:2097152\u0009- The amount of memory to give an instance of a Data Transform WebAssembly Virtual Machine. The maximum number of functions that can be deployed to a cluster is equal to data_transforms_per_core_memory_reservation / data_transforms_per_function_memory_limit\n","stream":"stderr","time":"2024-12-19T22:25:32.322038596Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.data_transforms_read_buffer_memory_percentage:45\u0009- The percentage of available memory in the transform subsystem to use for read buffers\n","stream":"stderr","time":"2024-12-19T22:25:32.322040189Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.data_transforms_runtime_limit_ms:3000\u0009- The maximum amount of runtime for startup time of a data transform, and the time it takes for a single record to be transformed.\n","stream":"stderr","time":"2024-12-19T22:25:32.322041572Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.data_transforms_write_buffer_memory_percentage:45\u0009- The percentage of available memory in the transform subsystem to use for write buffers\n","stream":"stderr","time":"2024-12-19T22:25:32.322042994Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.debug_load_slice_warning_depth:{nullopt}\u0009- The recursion depth after which debug logging will be enabled automatically for the log reader.\n","stream":"stderr","time":"2024-12-19T22:25:32.322044407Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.default_num_windows:10\u0009- Default number of quota tracking windows\n","stream":"stderr","time":"2024-12-19T22:25:32.322045789Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.default_topic_partitions:1\u0009- Default number of partitions per topic\n","stream":"stderr","time":"2024-12-19T22:25:32.322047222Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.default_topic_replications:1\u0009- Default replication factor for new topics\n","stream":"stderr","time":"2024-12-19T22:25:32.322048555Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.default_window_sec:1000\u0009- Default quota tracking window size in milliseconds\n","stream":"stderr","time":"2024-12-19T22:25:32.322049847Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.disable_batch_cache:0\u0009- Disable batch cache in log manager\n","stream":"stderr","time":"2024-12-19T22:25:32.32205115Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.disable_cluster_recovery_loop_for_tests:0\u0009- Disables the cluster recovery loop. The property exists to simplify testing and shouldn't be set in production.\n","stream":"stderr","time":"2024-12-19T22:25:32.322052432Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.disable_metrics:0\u0009- Disable registering metrics exposed on the internal metrics endpoint (/metrics)\n","stream":"stderr","time":"2024-12-19T22:25:32.322053815Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.disable_public_metrics:0\u0009- Disable registering metrics exposed on the public metrics endpoint (/public_metrics)\n","stream":"stderr","time":"2024-12-19T22:25:32.322055157Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.disk_reservation_percent:25\u0009- The percentage of total disk capacity that Redpanda will avoid using. This applies both when cloud cache and log data share a disk, as well as when cloud cache uses a dedicated disk. It is recommended to not run disks near capacity to avoid blocking I/O due to low disk space, as well as avoiding performance issues associated with SSD garbage collection.\n","stream":"stderr","time":"2024-12-19T22:25:32.32205651Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.election_timeout_ms:1500\u0009- Election timeout expressed in milliseconds\n","stream":"stderr","time":"2024-12-19T22:25:32.322060006Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.enable_admin_api:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.322061359Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.enable_auto_rebalance_on_node_add:0\u0009- Enable automatic partition rebalancing when new nodes are added\n","stream":"stderr","time":"2024-12-19T22:25:32.322062771Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.enable_cluster_metadata_upload_loop:1\u0009- Enables the cluster metadata upload loop.\n","stream":"stderr","time":"2024-12-19T22:25:32.322064124Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.enable_controller_log_rate_limiting:0\u0009- Enables limiting of controller log write rate\n","stream":"stderr","time":"2024-12-19T22:25:32.322065446Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.enable_coproc:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.322066759Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.enable_idempotence:1\u0009- Enable idempotent producer\n","stream":"stderr","time":"2024-12-19T22:25:32.322067991Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.enable_leader_balancer:1\u0009- Enable automatic leadership rebalancing\n","stream":"stderr","time":"2024-12-19T22:25:32.322069274Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.enable_metrics_reporter:1\u0009- Enable cluster metrics reporter\n","stream":"stderr","time":"2024-12-19T22:25:32.322070556Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.enable_mpx_extensions:0\u0009- Enable Redpanda extensions for MPX.\n","stream":"stderr","time":"2024-12-19T22:25:32.322071848Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.enable_pid_file:1\u0009- Enable pid file. You probably don't want to change this.\n","stream":"stderr","time":"2024-12-19T22:25:32.322073131Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.enable_rack_awareness:0\u0009- Enables rack-aware replica assignment\n","stream":"stderr","time":"2024-12-19T22:25:32.322074433Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.enable_sasl:0\u0009- Enable SASL authentication for Kafka connections, authorization is required. see also `kafka_enable_authorization`\n","stream":"stderr","time":"2024-12-19T22:25:32.322075746Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.enable_schema_id_validation:none\u0009- Enable Server Side Schema ID Validation.\n","stream":"stderr","time":"2024-12-19T22:25:32.322077098Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.enable_transactions:1\u0009- Enable transactions\n","stream":"stderr","time":"2024-12-19T22:25:32.322078411Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.enable_usage:0\u0009- Enables the usage tracking mechanism, storing windowed history of kafka/cloud_storage metrics over time\n","stream":"stderr","time":"2024-12-19T22:25:32.322079693Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.features_auto_enable:1\u0009- Whether new feature flags may auto-activate after upgrades (true) or must wait for manual activation via the admin API (false)\n","stream":"stderr","time":"2024-12-19T22:25:32.322081196Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.fetch_max_bytes:57671680\u0009- Maximum number of bytes returned in fetch request\n","stream":"stderr","time":"2024-12-19T22:25:32.322084412Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.fetch_read_strategy:non_polling\u0009- The strategy used to fulfill fetch requests\n","stream":"stderr","time":"2024-12-19T22:25:32.322085825Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.fetch_reads_debounce_timeout:10\u0009- Time to wait for next read in fetch request when requested min bytes wasn't reached\n","stream":"stderr","time":"2024-12-19T22:25:32.322087147Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.fetch_session_eviction_timeout_ms:60000\u0009- Minimum time before which unused session will get evicted from sessions. Maximum time after which inactive session will be deleted is two time given configuration valuecache\n","stream":"stderr","time":"2024-12-19T22:25:32.32208851Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.find_coordinator_timeout_ms:2000\u0009- Time to wait for a response from tx_registry\n","stream":"stderr","time":"2024-12-19T22:25:32.322089952Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.full_raft_configuration_recovery_pattern:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.322091275Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.group_initial_rebalance_delay:0\u0009- Extra delay (ms) added to rebalance phase to wait for new members\n","stream":"stderr","time":"2024-12-19T22:25:32.322092537Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.group_max_session_timeout_ms:300000\u0009- The maximum allowed session timeout for registered consumers. Longer timeouts give consumers more time to process messages in between heartbeats at the cost of a longer time to detect failures. \n","stream":"stderr","time":"2024-12-19T22:25:32.32209388Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.group_min_session_timeout_ms:6000\u0009- The minimum allowed session timeout for registered consumers. Shorter timeouts result in quicker failure detection at the cost of more frequent consumer heartbeating, which can overwhelm broker resources.\n","stream":"stderr","time":"2024-12-19T22:25:32.322095533Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.group_new_member_join_timeout:30000\u0009- Timeout for new member joins\n","stream":"stderr","time":"2024-12-19T22:25:32.322097016Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.group_offset_retention_check_ms:600000\u0009- How often the system should check for expired group offsets.\n","stream":"stderr","time":"2024-12-19T22:25:32.322098298Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.group_offset_retention_sec:{604800000}\u0009- Consumer group offset retention seconds. Offset retention can be disabled by setting this value to null.\n","stream":"stderr","time":"2024-12-19T22:25:32.32209964Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.group_topic_partitions:3\u0009- Number of partitions in the internal group membership topic\n","stream":"stderr","time":"2024-12-19T22:25:32.322101023Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.health_manager_tick_interval:180000\u0009- How often the health manager runs\n","stream":"stderr","time":"2024-12-19T22:25:32.322102346Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.health_monitor_max_metadata_age:10000\u0009- Max age of metadata cached in the health monitor of non controller node\n","stream":"stderr","time":"2024-12-19T22:25:32.322103648Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.health_monitor_tick_interval:10000\u0009- How often health monitor refresh cluster state\n","stream":"stderr","time":"2024-12-19T22:25:32.322106684Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.http_authentication:{BASIC}\u0009- A list of supported HTTP authentication mechanisms. `BASIC` and `OIDC` are allowed.\n","stream":"stderr","time":"2024-12-19T22:25:32.322108116Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.id_allocator_batch_size:1000\u0009- Id allocator allocates messages in batches (each batch is a one log record) and then serves requests from memory without touching the log until the batch is exhausted.\n","stream":"stderr","time":"2024-12-19T22:25:32.322109509Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.id_allocator_log_capacity:100\u0009- Capacity of the id_allocator log in number of batches. Once it reached id_allocator_stm truncates log's prefix.\n","stream":"stderr","time":"2024-12-19T22:25:32.322111122Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.id_allocator_replication:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.322112525Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.initial_retention_local_target_bytes_default:{nullopt}\u0009- Initial local retention size target for partitions of topics with cloud storage write enabled. If no initial local target retention is configured all locally retained data will be delivered to learner when joining partition replica set\n","stream":"stderr","time":"2024-12-19T22:25:32.322113807Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.initial_retention_local_target_ms_default:{nullopt}\u0009- Initial local retention time target for partitions of topics with cloud storage write enabled. If no initial local target retention is configured all locally retained data will be delivered to learner when joining partition replica set\n","stream":"stderr","time":"2024-12-19T22:25:32.32211536Z"} | |
{"log":"INFO 2024-12-19 22:25:32,321 [shard 0:main] main - application.cc:808 - redpanda.internal_topic_replication_factor:3\u0009- Target replication factor for internal topics\n","stream":"stderr","time":"2024-12-19T22:25:32.322116863Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.join_retry_timeout_ms:5000\u0009- Time between cluster join retries in milliseconds\n","stream":"stderr","time":"2024-12-19T22:25:32.322118175Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_admin_topic_api_rate:{nullopt}\u0009- Target quota rate (partition mutations per default_window_sec)\n","stream":"stderr","time":"2024-12-19T22:25:32.322119478Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_batch_max_bytes:1048576\u0009- Maximum size of a batch processed by server. If batch is compressed the limit applies to compressed batch size\n","stream":"stderr","time":"2024-12-19T22:25:32.32212082Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_client_group_byte_rate_quota:{}\u0009- Per-group target produce quota byte rate (bytes per second). Client is considered part of the group if client_id contains clients_prefix\n","stream":"stderr","time":"2024-12-19T22:25:32.322122203Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_client_group_fetch_byte_rate_quota:{}\u0009- Per-group target fetch quota byte rate (bytes per second). Client is considered part of the group if client_id contains clients_prefix\n","stream":"stderr","time":"2024-12-19T22:25:32.322123756Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_connection_rate_limit:{nullopt}\u0009- Maximum connections per second for one core\n","stream":"stderr","time":"2024-12-19T22:25:32.322126942Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_connection_rate_limit_overrides:{}\u0009- Overrides for specific ips for maximum connections per second for one core\n","stream":"stderr","time":"2024-12-19T22:25:32.322128324Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_connections_max:{nullopt}\u0009- Maximum number of Kafka client connections per broker\n","stream":"stderr","time":"2024-12-19T22:25:32.322129707Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_connections_max_overrides:{}\u0009- Per-IP overrides of kafka connection count limit, list of \u003cip\u003e:\u003ccount\u003e strings\n","stream":"stderr","time":"2024-12-19T22:25:32.322131009Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_connections_max_per_ip:{nullopt}\u0009- Maximum number of Kafka client connections from each IP address, per broker\n","stream":"stderr","time":"2024-12-19T22:25:32.322132482Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_enable_authorization:{nullopt}\u0009- Enable authorization for Kafka connections. Values:- `nil`: Ignored. Authorization is enabled with `enable_sasl: true`; `true`: authorization is required; `false`: authorization is disabled. See also: `enable_sasl` and `kafka_api[].authentication_method`\n","stream":"stderr","time":"2024-12-19T22:25:32.322133865Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_enable_describe_log_dirs_remote_storage:1\u0009- Whether to include tiered storage as a special remote:// directory in DescribeLogDirs Kafka API requests.\n","stream":"stderr","time":"2024-12-19T22:25:32.322135367Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_enable_partition_reassignment:1\u0009- Enable the Kafka partition reassignment API\n","stream":"stderr","time":"2024-12-19T22:25:32.32213674Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_group_recovery_timeout_ms:30000\u0009- Kafka group recovery timeout expressed in milliseconds\n","stream":"stderr","time":"2024-12-19T22:25:32.322138042Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_max_bytes_per_fetch:67108864\u0009- Limit fetch responses to this many bytes, even if total of partition bytes limits is higher\n","stream":"stderr","time":"2024-12-19T22:25:32.322139475Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_memory_batch_size_estimate_for_fetch:1048576\u0009- The size of the batch used to estimate memory consumption for Fetch requests, in bytes. Smaller sizes allow more concurrent fetch requests per shard, larger sizes prevent running out of memory because of too many concurrent fetch requests.\n","stream":"stderr","time":"2024-12-19T22:25:32.322140858Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_memory_share_for_fetch:0.5\u0009- The share of kafka subsystem memory that can be used for fetch read buffers, as a fraction of kafka subsystem memory amount\n","stream":"stderr","time":"2024-12-19T22:25:32.322142371Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_mtls_principal_mapping_rules:{nullopt}\u0009- Principal Mapping Rules for mTLS Authentication on the Kafka API\n","stream":"stderr","time":"2024-12-19T22:25:32.322143793Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_nodelete_topics:{_redpanda.audit_log, __consumer_offsets, _schemas}\u0009- Prevents the topics in the list from being deleted via the kafka api\n","stream":"stderr","time":"2024-12-19T22:25:32.322145126Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_noproduce_topics:{}\u0009- Prevents the topics in the list from having message produced to them via the kafka api\n","stream":"stderr","time":"2024-12-19T22:25:32.322148242Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_qdc_depth_alpha:0.8\u0009- Smoothing factor for kafka queue depth control depth tracking.\n","stream":"stderr","time":"2024-12-19T22:25:32.322149634Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_qdc_depth_update_ms:7000\u0009- Update frequency for kafka queue depth control.\n","stream":"stderr","time":"2024-12-19T22:25:32.322150957Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_qdc_enable:0\u0009- Enable kafka queue depth control.\n","stream":"stderr","time":"2024-12-19T22:25:32.322152279Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_qdc_idle_depth:10\u0009- Queue depth when idleness is detected in kafka queue depth control.\n","stream":"stderr","time":"2024-12-19T22:25:32.322153672Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_qdc_latency_alpha:0.002\u0009- Smoothing parameter for kafka queue depth control latency tracking.\n","stream":"stderr","time":"2024-12-19T22:25:32.322155004Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_qdc_max_depth:100\u0009- Maximum queue depth used in kafka queue depth control.\n","stream":"stderr","time":"2024-12-19T22:25:32.322156317Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_qdc_max_latency_ms:80\u0009- Max latency threshold for kafka queue depth control depth tracking.\n","stream":"stderr","time":"2024-12-19T22:25:32.322157609Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_qdc_min_depth:1\u0009- Minimum queue depth used in kafka queue depth control.\n","stream":"stderr","time":"2024-12-19T22:25:32.322158902Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_qdc_window_count:12\u0009- Number of windows used in kafka queue depth control latency tracking.\n","stream":"stderr","time":"2024-12-19T22:25:32.322160224Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_qdc_window_size_ms:1500\u0009- Window size for kafka queue depth control latency tracking.\n","stream":"stderr","time":"2024-12-19T22:25:32.322162188Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_quota_balancer_min_shard_throughput_bps:256\u0009- The lowest value of the throughput quota a shard can get in the process of quota balancing, in bytes/s. 0 means there is no minimum.\n","stream":"stderr","time":"2024-12-19T22:25:32.32216351Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_quota_balancer_min_shard_throughput_ratio:0.01\u0009- The lowest value of the throughput quota a shard can get in the process of quota balancing, expressed as a ratio of default shard quota. 0 means there is no minimum, 1 means no quota can be taken away by the balancer.\n","stream":"stderr","time":"2024-12-19T22:25:32.322164963Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_quota_balancer_node_period_ms:0\u0009- Intra-node throughput quota balancer invocation period, in milliseconds. Value of 0 disables the balancer and makes all the throughput quotas immutable.\n","stream":"stderr","time":"2024-12-19T22:25:32.322166536Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_quota_balancer_window_ms:5000\u0009- Time window used to average current throughput measurement for quota balancer, in milliseconds\n","stream":"stderr","time":"2024-12-19T22:25:32.322168139Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_request_max_bytes:104857600\u0009- Maximum size of a single request processed via Kafka API\n","stream":"stderr","time":"2024-12-19T22:25:32.322172677Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_rpc_server_stream_recv_buf:{nullopt}\u0009- Userspace receive buffer max size in bytes\n","stream":"stderr","time":"2024-12-19T22:25:32.32217406Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_rpc_server_tcp_recv_buf:{nullopt}\u0009- Kafka server TCP receive buffer size in bytes.\n","stream":"stderr","time":"2024-12-19T22:25:32.322175362Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_rpc_server_tcp_send_buf:{nullopt}\u0009- Kafka server TCP transmit buffer size in bytes.\n","stream":"stderr","time":"2024-12-19T22:25:32.322176655Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_sasl_max_reauth_ms:{nullopt}\u0009- The maximum time between Kafka client reauthentications. If a client has not reauthenticated a connection within this time frame, that connection is torn down. Without this, a connection could live long after the client's credentials are expired or revoked. Session expiry is disabled if the value is null.\n","stream":"stderr","time":"2024-12-19T22:25:32.322177987Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_schema_id_validation_cache_capacity:128\u0009- Per-shard capacity of the cache for validating schema IDs.\n","stream":"stderr","time":"2024-12-19T22:25:32.32217952Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_tcp_keepalive_probe_interval_seconds:60000\u0009- TCP keepalive probe interval in seconds for kafka connections. This describes the timeout between unacknowledged tcp keepalives. Refers to the TCP_KEEPINTVL socket option. When changed applies to new connections only.\n","stream":"stderr","time":"2024-12-19T22:25:32.322180853Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_tcp_keepalive_probes:3\u0009- TCP keepalive unacknowledged probes until the connection is considered dead for kafka connections. Refers to the TCP_KEEPCNT socket option. When changed applies to new connections only.\n","stream":"stderr","time":"2024-12-19T22:25:32.322182476Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_tcp_keepalive_timeout:120000\u0009- TCP keepalive idle timeout in seconds for kafka connections. This describes the timeout between tcp keepalive probes that the remote sitesuccessfully acknowledged. Refers to the TCP_KEEPIDLE socket option. When changed applies to new connections only.\n","stream":"stderr","time":"2024-12-19T22:25:32.322183908Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_throughput_control:{}\u0009- List of throughput control groups that define exclusions from node-wide throughput limits. Each group consists of: (\"name\" (optional) - any unique group name, \"client_id\" - regex to match client_id). A connection is assigned the first matching group, then the connection is excluded from throughput control.\n","stream":"stderr","time":"2024-12-19T22:25:32.322185421Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_throughput_controlled_api_keys:{produce, fetch}\u0009- List of Kafka API keys that are subject to cluster-wide and node-wide throughput limit control\n","stream":"stderr","time":"2024-12-19T22:25:32.322187054Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_throughput_limit_node_in_bps:{nullopt}\u0009- Node wide throughput ingress limit - maximum kafka traffic throughput allowed on the ingress side of each node, in bytes/s. Default is no limit.\n","stream":"stderr","time":"2024-12-19T22:25:32.322188457Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_throughput_limit_node_out_bps:{nullopt}\u0009- Node wide throughput egress limit - maximum kafka traffic throughput allowed on the egress side of each node, in bytes/s. Default is no limit.\n","stream":"stderr","time":"2024-12-19T22:25:32.322191613Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_throughput_replenish_threshold:{nullopt}\u0009- Threshold for refilling the token bucket. Will be clamped between 1 and kafka_throughput_limit_node_*_bps.\n","stream":"stderr","time":"2024-12-19T22:25:32.322193086Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_throughput_throttling_v2:1\u0009- Use throughput throttling based on a shared token bucket instead of balancing quota between shards\n","stream":"stderr","time":"2024-12-19T22:25:32.322194619Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kvstore_flush_interval:10\u0009- Key-value store flush interval (ms)\n","stream":"stderr","time":"2024-12-19T22:25:32.322195991Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kvstore_max_segment_size:16777216\u0009- Key-value maximum segment size (bytes)\n","stream":"stderr","time":"2024-12-19T22:25:32.322197274Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.leader_balancer_idle_timeout:120000\u0009- Leadership rebalancing idle timeout\n","stream":"stderr","time":"2024-12-19T22:25:32.322198586Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.leader_balancer_mode:random_hill_climbing\u0009- Leader balancer mode\n","stream":"stderr","time":"2024-12-19T22:25:32.322199888Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.leader_balancer_mute_timeout:300000\u0009- Leadership rebalancing mute timeout\n","stream":"stderr","time":"2024-12-19T22:25:32.322201171Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.leader_balancer_transfer_limit_per_shard:512\u0009- Per shard limit for in progress leadership transfers\n","stream":"stderr","time":"2024-12-19T22:25:32.322202473Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.legacy_group_offset_retention_enabled:0\u0009- Group offset retention is enabled by default in versions of Redpanda \u003e= 23.1. To enable offset retention after upgrading from an older version set this option to true.\n","stream":"stderr","time":"2024-12-19T22:25:32.322203816Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.legacy_permit_unsafe_log_operation:1\u0009- Permits the use of strings that may induct log injection/modification\n","stream":"stderr","time":"2024-12-19T22:25:32.322205319Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.legacy_unsafe_log_warning_interval_sec:300000\u0009- Interval, in seconds, of how often a message informing the operator that unsafe strings are permitted\n","stream":"stderr","time":"2024-12-19T22:25:32.322206661Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.log_cleanup_policy:delete\u0009- Default topic cleanup policy\n","stream":"stderr","time":"2024-12-19T22:25:32.322208044Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.log_compaction_interval_ms:10000\u0009- How often do we trigger background compaction\n","stream":"stderr","time":"2024-12-19T22:25:32.322209336Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.log_compaction_use_sliding_window:1\u0009- Use sliding window compaction.\n","stream":"stderr","time":"2024-12-19T22:25:32.322210779Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.log_compression_type:producer\u0009- Default topic compression type\n","stream":"stderr","time":"2024-12-19T22:25:32.322213815Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.log_disable_housekeeping_for_tests:0\u0009- Disables the housekeeping loop for local storage. The property exists to simplify testing and shouldn't be set in production.\n","stream":"stderr","time":"2024-12-19T22:25:32.322215157Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.log_message_timestamp_alert_after_ms:7200000\u0009- Threshold in milliseconds for alerting on messages with a timestamp after the broker's time, meaning they are in the future relative to the broker's clock.\n","stream":"stderr","time":"2024-12-19T22:25:32.32221658Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.log_message_timestamp_alert_before_ms:{nullopt}\u0009- Threshold in milliseconds for alerting on messages with a timestamp before the broker's time, meaning they are in the past relative to the broker's clock. null to disable this check\n","stream":"stderr","time":"2024-12-19T22:25:32.322218022Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.log_message_timestamp_type:CreateTime\u0009- Default topic messages timestamp type\n","stream":"stderr","time":"2024-12-19T22:25:32.322219485Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.log_retention_ms:604800000\u0009- delete segments older than this - default 1 week\n","stream":"stderr","time":"2024-12-19T22:25:32.322220788Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.log_segment_ms:{1209600000}\u0009- Default log segment lifetime in ms for topics which do not set segment.ms\n","stream":"stderr","time":"2024-12-19T22:25:32.32222209Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.log_segment_ms_max:31536000000\u0009- Upper bound on topic segment.ms: higher values will be clamped to this value\n","stream":"stderr","time":"2024-12-19T22:25:32.322223453Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.log_segment_ms_min:600000\u0009- Lower bound on topic segment.ms: lower values will be clamped to this value\n","stream":"stderr","time":"2024-12-19T22:25:32.322224945Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.log_segment_size:134217728\u0009- Default log segment size in bytes for topics which do not set segment.bytes\n","stream":"stderr","time":"2024-12-19T22:25:32.322227791Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.log_segment_size_jitter_percent:5\u0009- Random variation to the segment size limit used for each partition\n","stream":"stderr","time":"2024-12-19T22:25:32.322229294Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.log_segment_size_max:{nullopt}\u0009- Upper bound on topic segment.bytes: higher values will be clamped to this limit\n","stream":"stderr","time":"2024-12-19T22:25:32.322230676Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.log_segment_size_min:{1}\u0009- Lower bound on topic segment.bytes: lower values will be clamped to this limit\n","stream":"stderr","time":"2024-12-19T22:25:32.322232019Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.lz4_decompress_reusable_buffers_disabled:0\u0009- Disable reusable preallocated buffers for LZ4 decompression\n","stream":"stderr","time":"2024-12-19T22:25:32.322233381Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.max_compacted_log_segment_size:5368709120\u0009- Max compacted segment size after consolidation\n","stream":"stderr","time":"2024-12-19T22:25:32.322234704Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.max_concurrent_producer_ids:18446744073709551615\u0009- Max number of the active sessions (producers). When the threshold is passed Redpanda terminates old sessions. When an idle producer corresponding to the terminated session wakes up and produces - it leads to its batches being rejected with out of order sequence error.\n","stream":"stderr","time":"2024-12-19T22:25:32.32223786Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.max_in_flight_pandaproxy_requests_per_shard:500\u0009- Maximum number of in flight HTTP requests permitted in pandaproxy per shard. Any additional requests above this limit will be rejected with a 429 error\n","stream":"stderr","time":"2024-12-19T22:25:32.322239503Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.max_in_flight_schema_registry_requests_per_shard:500\u0009- Maximum number of in flight HTTP requests permitted in schema registry per shard. Any additional requests above this limit will be rejected with a 429 error\n","stream":"stderr","time":"2024-12-19T22:25:32.322241036Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.max_kafka_throttle_delay_ms:30000\u0009- Fail-safe maximum throttle delay on kafka requests\n","stream":"stderr","time":"2024-12-19T22:25:32.322242649Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.max_transactions_per_coordinator:18446744073709551615\u0009- Max number of the active txn sessions (producers). When the threshold is passed Redpanda terminates old sessions. When an idle producer corresponding to the terminated session wakes up and produces - it leads to its batches being rejected with invalid producer epoch or invalid_producer_id_mapping (it depends on the txn execution phase).\n","stream":"stderr","time":"2024-12-19T22:25:32.322244031Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.max_version:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.322245664Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.members_backend_retry_ms:5000\u0009- Time between members backend reconciliation loop retries \n","stream":"stderr","time":"2024-12-19T22:25:32.322246907Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.memory_abort_on_alloc_failure:1\u0009- If true, the redpanda process will terminate immediately when an allocation cannot be satisfied due to memory exhaustion. If false, an exception is thrown instead.\n","stream":"stderr","time":"2024-12-19T22:25:32.322248249Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.memory_enable_memory_sampling:1\u0009- If true, memory allocations will be sampled and tracked. A sampled live set of allocations can then be retrieved from the Admin API. Additionally, we will periodically log the top-n allocation sites\n","stream":"stderr","time":"2024-12-19T22:25:32.322249712Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.metadata_dissemination_interval_ms:3000\u0009- Interval for metadata dissemination batching\n","stream":"stderr","time":"2024-12-19T22:25:32.322251165Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.metadata_dissemination_retries:30\u0009- Number of attempts of looking up a topic's meta data like shard before failing a request\n","stream":"stderr","time":"2024-12-19T22:25:32.322252517Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.metadata_dissemination_retry_delay_ms:320\u0009- Delay before retry a topic lookup in a shard or other meta tables\n","stream":"stderr","time":"2024-12-19T22:25:32.32225402Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.metadata_status_wait_timeout_ms:2000\u0009- Maximum time to wait in metadata request for cluster health to be refreshed\n","stream":"stderr","time":"2024-12-19T22:25:32.322257356Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.metrics_reporter_report_interval:86400000\u0009- cluster metrics reporter report interval\n","stream":"stderr","time":"2024-12-19T22:25:32.322258759Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.metrics_reporter_tick_interval:60000\u0009- Cluster metrics reporter tick interval\n","stream":"stderr","time":"2024-12-19T22:25:32.322260101Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.metrics_reporter_url:https://m.rp.vectorized.io/v2\u0009- cluster metrics reporter url\n","stream":"stderr","time":"2024-12-19T22:25:32.322261404Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.min_version:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.322262726Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.minimum_topic_replications:1\u0009- Minimum permitted value of replication factor for new topics\n","stream":"stderr","time":"2024-12-19T22:25:32.322263979Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.node_isolation_heartbeat_timeout:3000\u0009- How long after the last heartbeat request a node will wait before considering itself to be isolated\n","stream":"stderr","time":"2024-12-19T22:25:32.322265331Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.node_management_operation_timeout_ms:5000\u0009- Timeout for executing node management operations\n","stream":"stderr","time":"2024-12-19T22:25:32.322266704Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.node_status_interval:100\u0009- Time interval between two node status messages. Node status messages establish liveness status outside of the Raft protocol.\n","stream":"stderr","time":"2024-12-19T22:25:32.322268026Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.node_status_reconnect_max_backoff_ms:15000\u0009- Maximum backoff (in ms) to reconnect to an unresponsive peer during node status liveness checks.\n","stream":"stderr","time":"2024-12-19T22:25:32.322269399Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.oidc_clock_skew_tolerance:0\u0009- The amount of seconds to allow for when validating the exp, nbf, and iat claims in the token.\n","stream":"stderr","time":"2024-12-19T22:25:32.322270882Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.oidc_discovery_url:https://auth.prd.cloud.redpanda.com/.well-known/openid-configuration\u0009- The URL pointing to the well-known discovery endpoint for the OIDC provider.\n","stream":"stderr","time":"2024-12-19T22:25:32.322272244Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.oidc_keys_refresh_interval:3600000\u0009- The frequency of refreshing the JSON Web Keys (JWKS) used to validate access tokens.\n","stream":"stderr","time":"2024-12-19T22:25:32.322273657Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.oidc_principal_mapping:$.sub\u0009- Rule for mapping JWT Payload claim to a Redpanda User Principal\n","stream":"stderr","time":"2024-12-19T22:25:32.322274999Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.oidc_token_audience:redpanda\u0009- A string representing the intended recipient of the token.\n","stream":"stderr","time":"2024-12-19T22:25:32.322276342Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.partition_autobalancing_concurrent_moves:50\u0009- Number of partitions that can be reassigned at once\n","stream":"stderr","time":"2024-12-19T22:25:32.322277664Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.partition_autobalancing_max_disk_usage_percent:80\u0009- Disk usage threshold that triggers moving partitions from the node\n","stream":"stderr","time":"2024-12-19T22:25:32.322281061Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.partition_autobalancing_min_size_threshold:{nullopt}\u0009- Minimum size of partition that is going to be prioritized when rebalancing cluster due to disk size threshold being breached. By default this value is calculated automaticaly\n","stream":"stderr","time":"2024-12-19T22:25:32.322282473Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.partition_autobalancing_mode:node_add\u0009- Partition autobalancing mode\n","stream":"stderr","time":"2024-12-19T22:25:32.322283966Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.partition_autobalancing_movement_batch_size_bytes:5368709120\u0009- Total size of partitions that autobalancer is going to move in one batch (deprecated, use partition_autobalancing_concurrent_moves to limit the autobalancer concurrency)\n","stream":"stderr","time":"2024-12-19T22:25:32.322285289Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.partition_autobalancing_node_availability_timeout_sec:900000\u0009- Node unavailability timeout that triggers moving partitions from the node\n","stream":"stderr","time":"2024-12-19T22:25:32.322286892Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.partition_autobalancing_tick_interval_ms:30000\u0009- Partition autobalancer tick interval\n","stream":"stderr","time":"2024-12-19T22:25:32.322288274Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.partition_autobalancing_tick_moves_drop_threshold:0.2\u0009- If the number of scheduled tick moves drops by this ratio, a new tick is scheduled immediately. Valid values are (0, 1]. For example, with a value of 0.2 and 100 scheduled moves in a tick, a new tick is scheduled when the inprogress moves are \u003c 80.\n","stream":"stderr","time":"2024-12-19T22:25:32.322289657Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.partition_autobalancing_topic_aware:1\u0009- If true, Redpanda will prioritize balancing topic-wise number of partitions on each node, as opposed to balancing the total number of partitions. This should give better balancing results if topics with diverse partition sizes and load profiles are present in the cluster.\n","stream":"stderr","time":"2024-12-19T22:25:32.32229122Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.partition_manager_shutdown_watchdog_timeout:30000\u0009- A threshold value to detect partitions which shutdown might have been stuck. After this threshold a watchdog in partition manager will log information about partition shutdown not making progress\n","stream":"stderr","time":"2024-12-19T22:25:32.322292803Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.pp_sr_smp_max_non_local_requests:{nullopt}\u0009- Maximum number of x-core requests pending in Panda Proxy and Schema Registry seastar::smp group. (for more details look at `seastar::smp_service_group` documentation)\n","stream":"stderr","time":"2024-12-19T22:25:32.322294265Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.quota_manager_gc_sec:30000\u0009- Quota manager GC frequency in milliseconds\n","stream":"stderr","time":"2024-12-19T22:25:32.322295728Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.raft_enable_longest_log_detection:1\u0009- Enables additional step in leader election where candidate is allowed to wait for all the replies from node it requested votes from. This may introduce a small delay when recovering from failure but will prevent truncation if any of the replicas has more data than the majority.\n","stream":"stderr","time":"2024-12-19T22:25:32.322297121Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.raft_enable_lw_heartbeat:1\u0009- enables raft optimization of heartbeats\n","stream":"stderr","time":"2024-12-19T22:25:32.322300527Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.raft_flush_timer_interval_ms:100\u0009- Interval of checking partition against the `raft_replica_max_pending_flush_bytes`, deprecated started 24.1, use raft_replica_max_flush_delay_ms instead \n","stream":"stderr","time":"2024-12-19T22:25:32.32230189Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.raft_heartbeat_disconnect_failures:3\u0009- After how many failed heartbeats to forcibly close an unresponsive TCP connection. Set to 0 to disable force disconnection.\n","stream":"stderr","time":"2024-12-19T22:25:32.322303342Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.raft_heartbeat_interval_ms:150\u0009- Milliseconds for raft leader heartbeats\n","stream":"stderr","time":"2024-12-19T22:25:32.322304745Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.raft_heartbeat_timeout_ms:3000\u0009- raft heartbeat RPC timeout\n","stream":"stderr","time":"2024-12-19T22:25:32.322306057Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.raft_io_timeout_ms:10000\u0009- Raft I/O timeout\n","stream":"stderr","time":"2024-12-19T22:25:32.32230735Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.raft_learner_recovery_rate:104857600\u0009- Raft learner recovery rate limit in bytes per sec\n","stream":"stderr","time":"2024-12-19T22:25:32.322308642Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.raft_max_concurrent_append_requests_per_follower:16\u0009- Maximum number of concurrent append entries requests sent by leader to one follower\n","stream":"stderr","time":"2024-12-19T22:25:32.322309975Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.raft_max_recovery_memory:{nullopt}\u0009- Max memory that can be used for reads in raft recovery process by default 15% of total memory\n","stream":"stderr","time":"2024-12-19T22:25:32.322311337Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.raft_recovery_concurrency_per_shard:64\u0009- How many partitions may simultaneously recover data to a particular shard. This is limited to avoid overwhelming nodes when they come back online after an outage.\n","stream":"stderr","time":"2024-12-19T22:25:32.32231285Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.raft_recovery_default_read_size:524288\u0009- default size of read issued during raft follower recovery\n","stream":"stderr","time":"2024-12-19T22:25:32.322314403Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.raft_recovery_throttle_disable_dynamic_mode:0\u0009- Disables dynamic rate allocation in recovery throttle (advanced).\n","stream":"stderr","time":"2024-12-19T22:25:32.322315746Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.raft_replica_max_flush_delay_ms:100\u0009- Maximum delay (in ms) between two subsequent flushes. After this delay, the log will be automatically force flushed.\n","stream":"stderr","time":"2024-12-19T22:25:32.322317098Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.raft_replica_max_pending_flush_bytes:{262144}\u0009- Max not flushed bytes per partition. If configured threshold is reached log will automatically be flushed even though it wasn't explicitly requested\n","stream":"stderr","time":"2024-12-19T22:25:32.322318481Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.raft_replicate_batch_window_size:1048576\u0009- Max size of requests cached for replication\n","stream":"stderr","time":"2024-12-19T22:25:32.322321837Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.raft_smp_max_non_local_requests:{nullopt}\u0009- Maximum number of x-core requests pending in Raft seastar::smp group. (for more details look at `seastar::smp_service_group` documentation)\n","stream":"stderr","time":"2024-12-19T22:25:32.32232323Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.raft_timeout_now_timeout_ms:1000\u0009- Timeout for a timeout now request\n","stream":"stderr","time":"2024-12-19T22:25:32.322324672Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.raft_transfer_leader_recovery_timeout_ms:10000\u0009- Timeout waiting for follower recovery when transferring leadership\n","stream":"stderr","time":"2024-12-19T22:25:32.322326005Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.readers_cache_eviction_timeout_ms:30000\u0009- Duration after which inactive readers will be evicted from cache\n","stream":"stderr","time":"2024-12-19T22:25:32.322327488Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.readers_cache_target_max_size:200\u0009- Maximum desired number of readers cached per ntp. This a soft limit, a number of readers in cache may temporary increase as cleanup is done in background\n","stream":"stderr","time":"2024-12-19T22:25:32.32232883Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.reclaim_batch_cache_min_free:67108864\u0009- Free memory limit that will be kept by batch cache background reclaimer\n","stream":"stderr","time":"2024-12-19T22:25:32.322330233Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.reclaim_growth_window:3000\u0009- Length of time in which reclaim sizes grow\n","stream":"stderr","time":"2024-12-19T22:25:32.322331595Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.reclaim_max_size:4194304\u0009- Maximum batch cache reclaim size\n","stream":"stderr","time":"2024-12-19T22:25:32.322332938Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.reclaim_min_size:131072\u0009- Minimum batch cache reclaim size\n","stream":"stderr","time":"2024-12-19T22:25:32.32233424Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.reclaim_stable_window:10000\u0009- Length of time above which growth is reset\n","stream":"stderr","time":"2024-12-19T22:25:32.322335493Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.recovery_append_timeout_ms:5000\u0009- Timeout for append entries requests issued while updating stale follower\n","stream":"stderr","time":"2024-12-19T22:25:32.322336785Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.release_cache_on_segment_roll:0\u0009- Free cache when segments roll\n","stream":"stderr","time":"2024-12-19T22:25:32.322338118Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.replicate_append_timeout_ms:3000\u0009- Timeout for append entries requests issued while replicating entries\n","stream":"stderr","time":"2024-12-19T22:25:32.3223394Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.retention_bytes:{nullopt}\u0009- Default max bytes per partition on disk before triggering a compaction\n","stream":"stderr","time":"2024-12-19T22:25:32.322340712Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.retention_local_strict:0\u0009- Trim log data when a cloud topic reaches its local retention limit. When this option is disabled Redpanda will allow partitions to grow past the local retention limit, and will be trimmed automatically as storage reaches the configured target size.\n","stream":"stderr","time":"2024-12-19T22:25:32.322342155Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.retention_local_strict_override:1\u0009- Trim log data when a cloud topic reaches its local retention limit. When this option is disabled Redpanda will allow partitions to grow past the local retention limit, and will be trimmed automatically as storage reaches the configured target size.\n","stream":"stderr","time":"2024-12-19T22:25:32.322345421Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.retention_local_target_bytes_default:{nullopt}\u0009- Local retention size target for partitions of topics with cloud storage write enabled\n","stream":"stderr","time":"2024-12-19T22:25:32.322346994Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.retention_local_target_capacity_bytes:{nullopt}\u0009- The target capacity in bytes that log storage will try to use before additional retention rules will take over to trim data in order to meet the target. When no target is specified storage usage is unbounded.\n","stream":"stderr","time":"2024-12-19T22:25:32.322348407Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.retention_local_target_capacity_percent:{80}\u0009- The target capacity in percent of unreserved space (see disk_reservation_percent) that log storage will try to use before additional retention rules will take over to trim data in order to meet the target. When no target is specified storage usage is unbounded.\n","stream":"stderr","time":"2024-12-19T22:25:32.32234995Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.retention_local_target_ms_default:86400000\u0009- Local retention time target for partitions of topics with cloud storage write enabled\n","stream":"stderr","time":"2024-12-19T22:25:32.322354989Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.retention_local_trim_interval:30000\u0009- The maximum amount of time before log storage will examine usage to determine of the target capacity has been exceeded and additional data trimming is required.\n","stream":"stderr","time":"2024-12-19T22:25:32.322356422Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.retention_local_trim_overage_coeff:2\u0009- The space management control loop will reclaim the overage multiplied by this this coefficient in order to compensate for data that is written during the idle period between control loop invocations.\n","stream":"stderr","time":"2024-12-19T22:25:32.322357945Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.rm_sync_timeout_ms:10000\u0009- Time to wait state catch up before rejecting a request\n","stream":"stderr","time":"2024-12-19T22:25:32.322359418Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.rm_violation_recovery_policy:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.322362724Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.rpc_client_connections_per_peer:128\u0009- The maximum number of connections a broker will open to each of its peers\n","stream":"stderr","time":"2024-12-19T22:25:32.322364106Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.rpc_server_compress_replies:0\u0009- Enable compression for internal rpc server replies\n","stream":"stderr","time":"2024-12-19T22:25:32.322365469Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.rpc_server_listen_backlog:{nullopt}\u0009- TCP connection queue length for Kafka server and internal RPC server\n","stream":"stderr","time":"2024-12-19T22:25:32.322366771Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.rpc_server_tcp_recv_buf:{nullopt}\u0009- Internal RPC TCP receive buffer size in bytes.\n","stream":"stderr","time":"2024-12-19T22:25:32.322369927Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.rpc_server_tcp_send_buf:{nullopt}\u0009- Internal RPC TCP transmit buffer size in bytes.\n","stream":"stderr","time":"2024-12-19T22:25:32.32237131Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.rps_limit_acls_and_users_operations:1000\u0009- Rate limit for controller acls and users operations\n","stream":"stderr","time":"2024-12-19T22:25:32.322372652Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.rps_limit_configuration_operations:1000\u0009- Rate limit for controller configuration operations\n","stream":"stderr","time":"2024-12-19T22:25:32.322373985Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.rps_limit_move_operations:1000\u0009- Rate limit for controller move operations\n","stream":"stderr","time":"2024-12-19T22:25:32.322375287Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.rps_limit_node_management_operations:1000\u0009- Rate limit for controller node management operations\n","stream":"stderr","time":"2024-12-19T22:25:32.32237658Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.rps_limit_topic_operations:1000\u0009- Rate limit for controller topic operations\n","stream":"stderr","time":"2024-12-19T22:25:32.322378032Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.sasl_kerberos_config:/etc/krb5.conf\u0009- The location of the Kerberos krb5.conf file for Redpanda\n","stream":"stderr","time":"2024-12-19T22:25:32.322379335Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.sasl_kerberos_keytab:/var/lib/redpanda/redpanda.keytab\u0009- The location of the Kerberos keytab file for Redpanda\n","stream":"stderr","time":"2024-12-19T22:25:32.322380677Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.sasl_kerberos_principal:redpanda\u0009- The primary of the Kerberos Service Principal Name (SPN) for Redpanda\n","stream":"stderr","time":"2024-12-19T22:25:32.32238202Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.sasl_kerberos_principal_mapping:{DEFAULT}\u0009- Rules for mapping Kerberos Principal Names to Redpanda User Principals\n","stream":"stderr","time":"2024-12-19T22:25:32.322383352Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.sasl_mechanisms:{SCRAM}\u0009- A list of supported SASL mechanisms. `SCRAM`, `GSSAPI`, and `OAUTHBEARER` are allowed.\n","stream":"stderr","time":"2024-12-19T22:25:32.322384695Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.schema_registry_normalize_on_startup:0\u0009- Normalize schemas as they are read from the topic on startup.\n","stream":"stderr","time":"2024-12-19T22:25:32.322386027Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.seed_server_meta_topic_partitions:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.32238739Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.segment_appender_flush_timeout_ms:1000\u0009- Maximum delay until buffered data is written\n","stream":"stderr","time":"2024-12-19T22:25:32.322388813Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.segment_fallocation_step:33554432\u0009- Size for segments fallocation\n","stream":"stderr","time":"2024-12-19T22:25:32.322390125Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.seq_table_min_size:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.322393131Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.space_management_enable:1\u0009- Enable automatic space management.\n","stream":"stderr","time":"2024-12-19T22:25:32.322394463Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.space_management_enable_override:0\u0009- Enable automatic space management. This option is ignored and deprecated in versions \u003e= v23.3.\n","stream":"stderr","time":"2024-12-19T22:25:32.322395886Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.space_management_max_log_concurrency:20\u0009- Maximum parallel logs inspected during space management process.\n","stream":"stderr","time":"2024-12-19T22:25:32.322397309Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.space_management_max_segment_concurrency:10\u0009- Maximum parallel segments inspected during space management process.\n","stream":"stderr","time":"2024-12-19T22:25:32.322398641Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.storage_compaction_index_memory:134217728\u0009- Maximum number of bytes that may be used on each shard by compactionindex writers\n","stream":"stderr","time":"2024-12-19T22:25:32.322399964Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.storage_compaction_key_map_memory:134217728\u0009- Maximum number of bytes that may be used on each shard by compaction key-offset maps. Only respected when `log_compaction_use_sliding_window` is true.\n","stream":"stderr","time":"2024-12-19T22:25:32.322401326Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.storage_compaction_key_map_memory_limit_percent:12\u0009- Limit on `storage_compaction_key_map_memory`, expressed as a percentage of memory per shard, that bounds the amount of memory used by compaction key-offset maps. NOTE: Memory per shard is computed after `wasm_per_core_memory_reservation`. Only respected when `log_compaction_use_sliding_window` is true.\n","stream":"stderr","time":"2024-12-19T22:25:32.322402749Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.storage_ignore_cstore_hints:0\u0009- if set, cstore hints will be ignored and will not be used for data access (but will otherwise be generated)\n","stream":"stderr","time":"2024-12-19T22:25:32.322404342Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.storage_ignore_timestamps_in_future_sec:{nullopt}\u0009- If set, timestamps more than this many seconds in the future relative tothe server's clock will be ignored for data retention purposes, and retention will act based on another timestamp in the same segment, or the mtime of the segment file if no valid timestamp is available\n","stream":"stderr","time":"2024-12-19T22:25:32.322405855Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.storage_max_concurrent_replay:1024\u0009- Maximum number of partitions' logs that will be replayed concurrently at startup, or flushed concurrently on shutdown.\n","stream":"stderr","time":"2024-12-19T22:25:32.322407448Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.storage_min_free_bytes:10485760\u0009- Threshold of minimum bytes free space before rejecting producers.\n","stream":"stderr","time":"2024-12-19T22:25:32.32240885Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.storage_read_buffer_size:131072\u0009- Size of each read buffer (one per in-flight read, per log segment)\n","stream":"stderr","time":"2024-12-19T22:25:32.322410183Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.storage_read_readahead_count:10\u0009- How many additional reads to issue ahead of current read location\n","stream":"stderr","time":"2024-12-19T22:25:32.322411505Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.storage_reserve_min_segments:2\u0009- The number of segments per partition that the system will attempt to reserve disk capacity for. For example, if the maximum segment size is configured to be 100 MB, and the value of this option is 2, then in a system with 10 partitions Redpanda will attempt to reserve at least 2 GB of disk space.\n","stream":"stderr","time":"2024-12-19T22:25:32.322414781Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.storage_space_alert_free_threshold_bytes:0\u0009- Threshold of minimum bytes free space before setting storage space alert\n","stream":"stderr","time":"2024-12-19T22:25:32.322416414Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.storage_space_alert_free_threshold_percent:5\u0009- Threshold of minimum percent free space before setting storage space alert\n","stream":"stderr","time":"2024-12-19T22:25:32.322417777Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.storage_strict_data_init:0\u0009- Requires that an empty file named `.redpanda_data_dir` be present in the data directory. Redpanda will refuse to start if it is not found.\n","stream":"stderr","time":"2024-12-19T22:25:32.322419129Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.storage_target_replay_bytes:10737418240\u0009- Target bytes to replay from disk on startup after clean shutdown: controls frequency of snapshots and checkpoints\n","stream":"stderr","time":"2024-12-19T22:25:32.322420642Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.superusers:{}\u0009- List of superuser usernames\n","stream":"stderr","time":"2024-12-19T22:25:32.322422045Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.target_fetch_quota_byte_rate:{nullopt}\u0009- Target fetch size quota byte rate (bytes per second) - disabled default\n","stream":"stderr","time":"2024-12-19T22:25:32.322423317Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.target_quota_byte_rate:0\u0009- Target request size quota byte rate (bytes per second)\n","stream":"stderr","time":"2024-12-19T22:25:32.32242466Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.tls_min_version:v1.2\u0009- The minimum TLS version that Redpanda supports.\n","stream":"stderr","time":"2024-12-19T22:25:32.322426042Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.tm_sync_timeout_ms:10000\u0009- Time to wait state catch up before rejecting a request\n","stream":"stderr","time":"2024-12-19T22:25:32.322427325Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.tm_violation_recovery_policy:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.322428637Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.topic_fds_per_partition:{5}\u0009- Required file handles per partition when creating topics\n","stream":"stderr","time":"2024-12-19T22:25:32.32242991Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.topic_memory_per_partition:{4194304}\u0009- Required memory per partition when creating topics\n","stream":"stderr","time":"2024-12-19T22:25:32.322431242Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.topic_partitions_per_shard:1000\u0009- Maximum number of partitions which may be allocated to one shard (CPU core)\n","stream":"stderr","time":"2024-12-19T22:25:32.322432565Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.topic_partitions_reserve_shard0:2\u0009- Reserved partition slots on shard (CPU core) 0 on each node. If this is \u003e= topic_partitions_per_core, no data partitions will be scheduled on shard 0\n","stream":"stderr","time":"2024-12-19T22:25:32.322433877Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.transaction_coordinator_cleanup_policy:delete\u0009- Cleanup policy for a transaction coordinator topic\n","stream":"stderr","time":"2024-12-19T22:25:32.322437003Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.transaction_coordinator_delete_retention_ms:604800000\u0009- delete segments older than this - default 1 week\n","stream":"stderr","time":"2024-12-19T22:25:32.322438526Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.transaction_coordinator_log_segment_size:1073741824\u0009- How large in bytes should each log segment be (default 1G)\n","stream":"stderr","time":"2024-12-19T22:25:32.322439918Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.transaction_coordinator_partitions:50\u0009- Amount of partitions for transactions coordinator\n","stream":"stderr","time":"2024-12-19T22:25:32.322441301Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.transaction_coordinator_replication:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.322442603Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.transactional_id_expiration_ms:604800000\u0009- Producer ids are expired once this time has elapsed after the last write with the given producer id.\n","stream":"stderr","time":"2024-12-19T22:25:32.322443896Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.tx_log_stats_interval_s:10000\u0009- How often to log per partition tx stats, works only with debug logging enabled.\n","stream":"stderr","time":"2024-12-19T22:25:32.322445268Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.tx_registry_log_capacity:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.322446611Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.tx_registry_sync_timeout_ms:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.322447863Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.tx_timeout_delay_ms:1000\u0009- Delay before scheduling next check for timed out transactions\n","stream":"stderr","time":"2024-12-19T22:25:32.322449136Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.unsafe_enable_consumer_offsets_delete_retention:0\u0009- Enables delete retention of consumer offsets topic. This is an internal-only configuration and should be enabled only after consulting with Redpanda Support or engineers.\n","stream":"stderr","time":"2024-12-19T22:25:32.322450468Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.usage_disk_persistance_interval_sec:300000\u0009- The interval in which all usage stats are written to disk\n","stream":"stderr","time":"2024-12-19T22:25:32.322451951Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.usage_num_windows:24\u0009- The number of windows to persist in memory and disk\n","stream":"stderr","time":"2024-12-19T22:25:32.322453384Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.usage_window_width_interval_sec:3600000\u0009- The width of a usage window, tracking cloud and kafka ingress/egress traffic each interval\n","stream":"stderr","time":"2024-12-19T22:25:32.322454706Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.use_fetch_scheduler_group:1\u0009- Use a separate scheduler group for fetch processing\n","stream":"stderr","time":"2024-12-19T22:25:32.322456079Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.use_scheduling_groups:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.322459405Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.virtual_cluster_min_producer_ids:18446744073709551615\u0009- Minimum number of active producers per virtual cluster\n","stream":"stderr","time":"2024-12-19T22:25:32.322460747Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.wait_for_leader_timeout_ms:5000\u0009- Timeout (ms) to wait for leadership in metadata cache\n","stream":"stderr","time":"2024-12-19T22:25:32.32246211Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.write_caching_default:true\u0009- Cache batches until the segment appender chunk is full instead of flushing for every acks=all write. This is the global default for all topics and can be overriden at a topic scope with property write.caching. 'disabled' mode takes precedence over topic overrides and disables the feature altogether for the entire cluster.\n","stream":"stderr","time":"2024-12-19T22:25:32.322463473Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.zstd_decompress_workspace_bytes:8388608\u0009- Size of the zstd decompression workspace\n","stream":"stderr","time":"2024-12-19T22:25:32.322465046Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:859 - Node configuration properties:\n","stream":"stderr","time":"2024-12-19T22:25:32.322466338Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:860 - (use `rpk redpanda config set \u003ccfg\u003e \u003cvalue\u003e` to change)\n","stream":"stderr","time":"2024-12-19T22:25:32.32246758Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.admin:{{:{host: 0.0.0.0, port: 9644}}}\u0009- Address and port of admin server\n","stream":"stderr","time":"2024-12-19T22:25:32.322468983Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.admin_api_doc_dir:/usr/share/redpanda/admin-api-doc\u0009- Admin API doc directory\n","stream":"stderr","time":"2024-12-19T22:25:32.322470416Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.admin_api_tls:{}\u0009- TLS configuration for admin HTTP server\n","stream":"stderr","time":"2024-12-19T22:25:32.322471748Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.advertised_kafka_api:{{:{host: 127.0.0.1, port: 9092}}}\u0009- Address of Kafka API published to the clients\n","stream":"stderr","time":"2024-12-19T22:25:32.322473041Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.advertised_rpc_api:{{host: 127.0.0.1, port: 33145}}\u0009- Address of RPC endpoint published to other cluster members\n","stream":"stderr","time":"2024-12-19T22:25:32.322474363Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.cloud_storage_cache_directory:{nullopt}\u0009- Directory for archival cache. Should be present when `cloud_storage_enabled` is present\n","stream":"stderr","time":"2024-12-19T22:25:32.322475746Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.coproc_supervisor_server:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.322477138Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.crash_loop_limit:{5}\u0009- Maximum consecutive crashes (unclean shutdowns) allowed after which operator intervention is needed to startup the broker. Limit is not enforced in developer mode.\n","stream":"stderr","time":"2024-12-19T22:25:32.322478421Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.dashboard_dir:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.322479853Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.data_directory:{data_directory=\"/var/lib/redpanda/data\"}\u0009- Place where redpanda will keep the data\n","stream":"stderr","time":"2024-12-19T22:25:32.322481106Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.developer_mode:1\u0009- Skips most of the checks performed at startup, not recomended for production use\n","stream":"stderr","time":"2024-12-19T22:25:32.322484232Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.emergency_disable_data_transforms:0\u0009- Override the cluster enablement setting and disable WebAssembly powered data transforms. Only used as an emergency shutoff button.\n","stream":"stderr","time":"2024-12-19T22:25:32.322485634Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.empty_seed_starts_cluster:1\u0009- If true, an empty seed_servers list will denote that this node should form a cluster. At most one node in the cluster should be configured configured with an empty seed_servers list. If no such configured node exists, or if configured to false, all nodes denoted by the seed_servers list must be identical among those nodes' configurations, and those nodes will form the initial cluster.\n","stream":"stderr","time":"2024-12-19T22:25:32.322487157Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.enable_central_config:\u0009- \n","stream":"stderr","time":"2024-12-19T22:25:32.32248881Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.fips_mode:disabled\u0009- Controls whether Redpanda starts in FIPS mode. This property allows for three values: 'disabled', 'enabled', and 'permissive'. With 'enabled', Redpanda first verifies that the operating system is enabled for FIPS by checking /proc/sys/crypto/fips_enabled. If the file does not exist or does not return '1', Redpanda immediately exits. With 'permissive', the same check is performed but a WARNING is logged and Redpanda continues to run. After the check is complete, Redpanda loads the OpenSSL FIPS provider into the OpenSSL library. After this is complete, Redpanda is operating in FIPS mode, which means that the TLS cipher suites available to users are limited to TLSv1.2 and TLSv1.3, and of those, only the ones that use NIST-approved cryptographic methods. For more information about FIPS, refer to Redpanda documentation.\n","stream":"stderr","time":"2024-12-19T22:25:32.322490143Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_api:{{:{host: 0.0.0.0, port: 9092}:{nullopt}}}\u0009- Address and port of an interface to listen for Kafka API requests\n","stream":"stderr","time":"2024-12-19T22:25:32.322492166Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.kafka_api_tls:{}\u0009- TLS configuration for Kafka API endpoint\n","stream":"stderr","time":"2024-12-19T22:25:32.32249422Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.memory_allocation_warning_threshold:{131073}\u0009- Enables log messages for allocations greater than the given size.\n","stream":"stderr","time":"2024-12-19T22:25:32.322495994Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.node_id:{nullopt}\u0009- Unique id identifying a node in the cluster. If missing, a unique id will be assigned for this node when it joins the cluster\n","stream":"stderr","time":"2024-12-19T22:25:32.322498087Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.node_id_overrides:{}\u0009- List of node ID and UUID overrides to be applied at broker startup. Each entry includes the current UUID and desired ID and UUID. Each entry applies to a given node if and only if 'current' matches that node's current UUID.\n","stream":"stderr","time":"2024-12-19T22:25:32.322500121Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.openssl_config_file:{nullopt}\u0009- Path to the configuration file used by OpenSSL to propertly load the FIPS-compliant module.\n","stream":"stderr","time":"2024-12-19T22:25:32.322502085Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.openssl_module_directory:{nullopt}\u0009- Path to the directory that contains the OpenSSL FIPS-compliant module.\n","stream":"stderr","time":"2024-12-19T22:25:32.322507736Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.rack:{nullopt}\u0009- Rack identifier\n","stream":"stderr","time":"2024-12-19T22:25:32.322509569Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.recovery_mode_enabled:0\u0009- If true, start redpanda in \"metadata only\" mode, skipping loading user partitions and allowing only metadata operations.\n","stream":"stderr","time":"2024-12-19T22:25:32.322511242Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.rpc_server:{host: 0.0.0.0, port: 33145}\u0009- IpAddress and port for RPC server\n","stream":"stderr","time":"2024-12-19T22:25:32.322513076Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.rpc_server_tls:{ enabled: 0 key/cert files: {nullopt} ca file: {nullopt} crl file: {nullopt} client_auth_required: 0 }\u0009- TLS configuration for RPC server\n","stream":"stderr","time":"2024-12-19T22:25:32.322514779Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.seed_servers:{}\u0009- List of the seed servers used to join current cluster. If the seed_server list is empty the node will be a cluster root and it will form a new cluster\n","stream":"stderr","time":"2024-12-19T22:25:32.322516542Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.storage_failure_injection_config_path:{nullopt}\u0009- Path to the configuration file used for low level storage failure injection\n","stream":"stderr","time":"2024-12-19T22:25:32.322518396Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.storage_failure_injection_enabled:0\u0009- If true, inject low level storage failures on the write path. **Not** for production usage.\n","stream":"stderr","time":"2024-12-19T22:25:32.322520359Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.upgrade_override_checks:0\u0009- Whether to violate safety checks when starting a redpanda version newer than the cluster's consensus version\n","stream":"stderr","time":"2024-12-19T22:25:32.322523044Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - redpanda.verbose_logging_timeout_sec_max:{nullopt}\u0009- Maximum duration in seconds for verbose (i.e. TRACE or DEBUG) logging. Values configured above this will be clamped. If null (the default) there is no limit. Can be overridded in the Admin API on a per-request basis.\n","stream":"stderr","time":"2024-12-19T22:25:32.322525098Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy.advertised_pandaproxy_api:{}\u0009- Rest API address and port to publish to client\n","stream":"stderr","time":"2024-12-19T22:25:32.322527402Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy.api_doc_dir:/usr/share/redpanda/proxy-api-doc\u0009- API doc directory\n","stream":"stderr","time":"2024-12-19T22:25:32.322529386Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy.client_cache_max_size:10\u0009- The maximum number of kafka clients in the LRU cache\n","stream":"stderr","time":"2024-12-19T22:25:32.32253128Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy.client_keep_alive:300000\u0009- Time in milliseconds that an idle connection may remain open\n","stream":"stderr","time":"2024-12-19T22:25:32.322532873Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy.consumer_instance_timeout_ms:300000\u0009- How long to wait for an idle consumer before removing it\n","stream":"stderr","time":"2024-12-19T22:25:32.32253651Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy.pandaproxy_api:{{:{host: 0.0.0.0, port: 8082}:\u003cnullopt\u003e}}\u0009- Rest API listen address and port\n","stream":"stderr","time":"2024-12-19T22:25:32.322538253Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy.pandaproxy_api_tls:{}\u0009- TLS configuration for Pandaproxy api\n","stream":"stderr","time":"2024-12-19T22:25:32.322540146Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy_client.broker_tls:{ enabled: 0 key/cert files: {nullopt} ca file: {nullopt} crl file: {nullopt} client_auth_required: 0 }\u0009- TLS configuration for the brokers\n","stream":"stderr","time":"2024-12-19T22:25:32.322541809Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy_client.brokers:{{host: 0.0.0.0, port: 9092}}\u0009- List of address and port of the brokers\n","stream":"stderr","time":"2024-12-19T22:25:32.322543743Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy_client.client_identifier:{pandaproxy_client}\u0009- Identifier to use within the kafka request header\n","stream":"stderr","time":"2024-12-19T22:25:32.322545727Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy_client.consumer_heartbeat_interval_ms:500\u0009- Interval (in milliseconds) for consumer heartbeats\n","stream":"stderr","time":"2024-12-19T22:25:32.32254749Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy_client.consumer_rebalance_timeout_ms:2000\u0009- Timeout (in milliseconds) for consumer rebalance\n","stream":"stderr","time":"2024-12-19T22:25:32.322549123Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy_client.consumer_request_max_bytes:1048576\u0009- Max bytes to fetch per request\n","stream":"stderr","time":"2024-12-19T22:25:32.322551077Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy_client.consumer_request_min_bytes:1\u0009- Min bytes to fetch per request\n","stream":"stderr","time":"2024-12-19T22:25:32.32255297Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy_client.consumer_request_timeout_ms:100\u0009- Interval (in milliseconds) for consumer request timeout\n","stream":"stderr","time":"2024-12-19T22:25:32.322554563Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy_client.consumer_session_timeout_ms:300000\u0009- Timeout (in milliseconds) for consumer session\n","stream":"stderr","time":"2024-12-19T22:25:32.322556136Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy_client.produce_ack_level:-1\u0009- Number of acknowledgments the producer requires the leader to have received before considering a request complete, choices are 0, 1 and -1\n","stream":"stderr","time":"2024-12-19T22:25:32.322557819Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy_client.produce_batch_delay_ms:100\u0009- Delay (in milliseconds) to wait before sending batch\n","stream":"stderr","time":"2024-12-19T22:25:32.322559513Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy_client.produce_batch_record_count:1000\u0009- Number of records to batch before sending to broker\n","stream":"stderr","time":"2024-12-19T22:25:32.322561436Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy_client.produce_batch_size_bytes:1048576\u0009- Number of bytes to batch before sending to broker\n","stream":"stderr","time":"2024-12-19T22:25:32.32256336Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy_client.produce_compression_type:none\u0009- Enable or disable compression by the kafka client. Specify 'none' to disable compression or one of the supported types [gzip, snappy, lz4, zstd]\n","stream":"stderr","time":"2024-12-19T22:25:32.322567147Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy_client.produce_shutdown_delay_ms:0\u0009- Delay (in milliseconds) to allow for final flush of buffers before shutting down\n","stream":"stderr","time":"2024-12-19T22:25:32.322569231Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy_client.retries:5\u0009- Number of times to retry a request to a broker\n","stream":"stderr","time":"2024-12-19T22:25:32.322570944Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy_client.retry_base_backoff_ms:100\u0009- Delay (in milliseconds) for initial retry backoff\n","stream":"stderr","time":"2024-12-19T22:25:32.322572487Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy_client.sasl_mechanism:\u0009- The SASL mechanism to use when connecting\n","stream":"stderr","time":"2024-12-19T22:25:32.32257405Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy_client.scram_password:\u0009- Password to use for SCRAM authentication mechanisms\n","stream":"stderr","time":"2024-12-19T22:25:32.322575773Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - pandaproxy_client.scram_username:\u0009- Username to use for SCRAM authentication mechanisms\n","stream":"stderr","time":"2024-12-19T22:25:32.322577356Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry.api_doc_dir:/usr/share/redpanda/proxy-api-doc\u0009- API doc directory\n","stream":"stderr","time":"2024-12-19T22:25:32.322579149Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry.mode_mutability:1\u0009- Allow modifying mode\n","stream":"stderr","time":"2024-12-19T22:25:32.322580943Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry.schema_registry_api:{{:{host: 0.0.0.0, port: 8081}:\u003cnullopt\u003e}}\u0009- Schema Registry API listen address and port\n","stream":"stderr","time":"2024-12-19T22:25:32.322582826Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry.schema_registry_api_tls:{}\u0009- TLS configuration for Schema Registry API\n","stream":"stderr","time":"2024-12-19T22:25:32.32258501Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry.schema_registry_replication_factor:{nullopt}\u0009- Replication factor for internal _schemas topic. If unset, defaults to `default_topic_replication`\n","stream":"stderr","time":"2024-12-19T22:25:32.322586804Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry_client.broker_tls:{ enabled: 0 key/cert files: {nullopt} ca file: {nullopt} crl file: {nullopt} client_auth_required: 0 }\u0009- TLS configuration for the brokers\n","stream":"stderr","time":"2024-12-19T22:25:32.322588818Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry_client.brokers:{{host: 0.0.0.0, port: 9092}}\u0009- List of address and port of the brokers\n","stream":"stderr","time":"2024-12-19T22:25:32.322590891Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry_client.client_identifier:{schema_registry_client}\u0009- Identifier to use within the kafka request header\n","stream":"stderr","time":"2024-12-19T22:25:32.322592735Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry_client.consumer_heartbeat_interval_ms:500\u0009- Interval (in milliseconds) for consumer heartbeats\n","stream":"stderr","time":"2024-12-19T22:25:32.322594378Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry_client.consumer_rebalance_timeout_ms:2000\u0009- Timeout (in milliseconds) for consumer rebalance\n","stream":"stderr","time":"2024-12-19T22:25:32.322598376Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry_client.consumer_request_max_bytes:1048576\u0009- Max bytes to fetch per request\n","stream":"stderr","time":"2024-12-19T22:25:32.322600029Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry_client.consumer_request_min_bytes:1\u0009- Min bytes to fetch per request\n","stream":"stderr","time":"2024-12-19T22:25:32.322601962Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry_client.consumer_request_timeout_ms:100\u0009- Interval (in milliseconds) for consumer request timeout\n","stream":"stderr","time":"2024-12-19T22:25:32.322603706Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry_client.consumer_session_timeout_ms:10000\u0009- Timeout (in milliseconds) for consumer session\n","stream":"stderr","time":"2024-12-19T22:25:32.322605489Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry_client.produce_ack_level:-1\u0009- Number of acknowledgments the producer requires the leader to have received before considering a request complete, choices are 0, 1 and -1\n","stream":"stderr","time":"2024-12-19T22:25:32.322607322Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry_client.produce_batch_delay_ms:0\u0009- Delay (in milliseconds) to wait before sending batch\n","stream":"stderr","time":"2024-12-19T22:25:32.322609106Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry_client.produce_batch_record_count:0\u0009- Number of records to batch before sending to broker\n","stream":"stderr","time":"2024-12-19T22:25:32.322611039Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry_client.produce_batch_size_bytes:0\u0009- Number of bytes to batch before sending to broker\n","stream":"stderr","time":"2024-12-19T22:25:32.322612903Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry_client.produce_compression_type:none\u0009- Enable or disable compression by the kafka client. Specify 'none' to disable compression or one of the supported types [gzip, snappy, lz4, zstd]\n","stream":"stderr","time":"2024-12-19T22:25:32.322614476Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry_client.produce_shutdown_delay_ms:0\u0009- Delay (in milliseconds) to allow for final flush of buffers before shutting down\n","stream":"stderr","time":"2024-12-19T22:25:32.322616329Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry_client.retries:5\u0009- Number of times to retry a request to a broker\n","stream":"stderr","time":"2024-12-19T22:25:32.322618283Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry_client.retry_base_backoff_ms:100\u0009- Delay (in milliseconds) for initial retry backoff\n","stream":"stderr","time":"2024-12-19T22:25:32.322619946Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry_client.sasl_mechanism:\u0009- The SASL mechanism to use when connecting\n","stream":"stderr","time":"2024-12-19T22:25:32.322621619Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry_client.scram_password:\u0009- Password to use for SCRAM authentication mechanisms\n","stream":"stderr","time":"2024-12-19T22:25:32.322623192Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - schema_registry_client.scram_username:\u0009- Username to use for SCRAM authentication mechanisms\n","stream":"stderr","time":"2024-12-19T22:25:32.322625066Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - audit_log_client.broker_tls:{ enabled: 0 key/cert files: {nullopt} ca file: {nullopt} crl file: {nullopt} client_auth_required: 0 }\u0009- TLS configuration for the brokers\n","stream":"stderr","time":"2024-12-19T22:25:32.322629093Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - audit_log_client.brokers:{{host: 0.0.0.0, port: 9092}}\u0009- List of address and port of the brokers\n","stream":"stderr","time":"2024-12-19T22:25:32.322631037Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - audit_log_client.client_identifier:{audit_log_client}\u0009- Identifier to use within the kafka request header\n","stream":"stderr","time":"2024-12-19T22:25:32.32263292Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - audit_log_client.consumer_heartbeat_interval_ms:500\u0009- Interval (in milliseconds) for consumer heartbeats\n","stream":"stderr","time":"2024-12-19T22:25:32.322634583Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - audit_log_client.consumer_rebalance_timeout_ms:2000\u0009- Timeout (in milliseconds) for consumer rebalance\n","stream":"stderr","time":"2024-12-19T22:25:32.322636277Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - audit_log_client.consumer_request_max_bytes:1048576\u0009- Max bytes to fetch per request\n","stream":"stderr","time":"2024-12-19T22:25:32.32263822Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - audit_log_client.consumer_request_min_bytes:1\u0009- Min bytes to fetch per request\n","stream":"stderr","time":"2024-12-19T22:25:32.322639863Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - audit_log_client.consumer_request_timeout_ms:100\u0009- Interval (in milliseconds) for consumer request timeout\n","stream":"stderr","time":"2024-12-19T22:25:32.322641647Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - audit_log_client.consumer_session_timeout_ms:10000\u0009- Timeout (in milliseconds) for consumer session\n","stream":"stderr","time":"2024-12-19T22:25:32.32264351Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - audit_log_client.produce_ack_level:1\u0009- Number of acknowledgments the producer requires the leader to have received before considering a request complete, choices are 0, 1 and -1\n","stream":"stderr","time":"2024-12-19T22:25:32.322645193Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - audit_log_client.produce_batch_delay_ms:0\u0009- Delay (in milliseconds) to wait before sending batch\n","stream":"stderr","time":"2024-12-19T22:25:32.322646987Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - audit_log_client.produce_batch_record_count:0\u0009- Number of records to batch before sending to broker\n","stream":"stderr","time":"2024-12-19T22:25:32.322649441Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - audit_log_client.produce_batch_size_bytes:0\u0009- Number of bytes to batch before sending to broker\n","stream":"stderr","time":"2024-12-19T22:25:32.322651175Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - audit_log_client.produce_compression_type:zstd\u0009- Enable or disable compression by the kafka client. Specify 'none' to disable compression or one of the supported types [gzip, snappy, lz4, zstd]\n","stream":"stderr","time":"2024-12-19T22:25:32.322652798Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - audit_log_client.produce_shutdown_delay_ms:3000\u0009- Delay (in milliseconds) to allow for final flush of buffers before shutting down\n","stream":"stderr","time":"2024-12-19T22:25:32.322654681Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - audit_log_client.retries:5\u0009- Number of times to retry a request to a broker\n","stream":"stderr","time":"2024-12-19T22:25:32.322658849Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - audit_log_client.retry_base_backoff_ms:100\u0009- Delay (in milliseconds) for initial retry backoff\n","stream":"stderr","time":"2024-12-19T22:25:32.322660512Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - audit_log_client.sasl_mechanism:\u0009- The SASL mechanism to use when connecting\n","stream":"stderr","time":"2024-12-19T22:25:32.322662285Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - audit_log_client.scram_password:\u0009- Password to use for SCRAM authentication mechanisms\n","stream":"stderr","time":"2024-12-19T22:25:32.322664049Z"} | |
{"log":"INFO 2024-12-19 22:25:32,322 [shard 0:main] main - application.cc:808 - audit_log_client.scram_username:\u0009- Username to use for SCRAM authentication mechanisms\n","stream":"stderr","time":"2024-12-19T22:25:32.322665652Z"} | |
{"log":"INFO 2024-12-19 22:25:32,323 [shard 0:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.324808189Z"} | |
{"log":"INFO 2024-12-19 22:25:32,323 [shard 16:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.324827865Z"} | |
{"log":"INFO 2024-12-19 22:25:32,323 [shard 1:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.32483044Z"} | |
{"log":"INFO 2024-12-19 22:25:32,323 [shard 20:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.324832654Z"} | |
{"log":"INFO 2024-12-19 22:25:32,323 [shard 6:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.324834959Z"} | |
{"log":"INFO 2024-12-19 22:25:32,323 [shard 3:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.324837003Z"} | |
{"log":"INFO 2024-12-19 22:25:32,323 [shard 8:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.324838846Z"} | |
{"log":"INFO 2024-12-19 22:25:32,324 [shard 18:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.32484086Z"} | |
{"log":"INFO 2024-12-19 22:25:32,324 [shard 23:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.324842934Z"} | |
{"log":"INFO 2024-12-19 22:25:32,324 [shard 21:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.324890803Z"} | |
{"log":"INFO 2024-12-19 22:25:32,324 [shard 9:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.32489404Z"} | |
{"log":"INFO 2024-12-19 22:25:32,324 [shard 15:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.324896023Z"} | |
{"log":"INFO 2024-12-19 22:25:32,324 [shard 17:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.324897967Z"} | |
{"log":"INFO 2024-12-19 22:25:32,324 [shard 19:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.324900051Z"} | |
{"log":"INFO 2024-12-19 22:25:32,324 [shard 7:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.324901864Z"} | |
{"log":"INFO 2024-12-19 22:25:32,324 [shard 11:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.324903818Z"} | |
{"log":"INFO 2024-12-19 22:25:32,324 [shard 2:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.32491565Z"} | |
{"log":"INFO 2024-12-19 22:25:32,324 [shard 13:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.324919117Z"} | |
{"log":"INFO 2024-12-19 22:25:32,324 [shard 4:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.32492103Z"} | |
{"log":"INFO 2024-12-19 22:25:32,323 [shard 10:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.324923064Z"} | |
{"log":"INFO 2024-12-19 22:25:32,325 [shard 5:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.32505955Z"} | |
{"log":"INFO 2024-12-19 22:25:32,325 [shard 12:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.325122578Z"} | |
{"log":"INFO 2024-12-19 22:25:32,325 [shard 22:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.325222045Z"} | |
{"log":"INFO 2024-12-19 22:25:32,325 [shard 14:main] seastar - Enabling heap profiler - using 3000037 bytes sampling rate\n","stream":"stderr","time":"2024-12-19T22:25:32.325266818Z"} | |
{"log":"INFO 2024-12-19 22:25:32,325 [shard 0:main] main - application.cc:523 - Setting abort_on_allocation_failure (abort on OOM): true\n","stream":"stderr","time":"2024-12-19T22:25:32.325301063Z"} | |
{"log":"INFO 2024-12-19 22:25:32,326 [shard 0:main] syschecks - Writing pid file \"/var/lib/redpanda/data/pid.lock\"\n","stream":"stderr","time":"2024-12-19T22:25:32.326518355Z"} | |
{"log":"ERROR 2024-12-19 22:25:32,330 [shard 0:main] syschecks - Memory: '174063616' below recommended: '1073741824'\n","stream":"stderr","time":"2024-12-19T22:25:32.330636706Z"} | |
{"log":"INFO 2024-12-19 22:25:32,330 [shard 0:main] storage - directories.h:33 - Checking `/var/lib/redpanda/data` for supported filesystems\n","stream":"stderr","time":"2024-12-19T22:25:32.330874442Z"} | |
{"log":"INFO 2024-12-19 22:25:32,331 [shard 0:main] syschecks - Detected file system type is ext2\n","stream":"stderr","time":"2024-12-19T22:25:32.331037397Z"} | |
{"log":"WARN 2024-12-19 22:25:32,331 [shard 0:main] syschecks - Path: `/var/lib/redpanda/data' is on ext4, not XFS. This will probably work, but Redpanda is only tested on XFS and XFS is recommended for best performance.\n","stream":"stderr","time":"2024-12-19T22:25:32.331039952Z"} | |
{"log":"INFO 2024-12-19 22:25:32,331 [shard 0:main] cloud_storage - cache_service.cc:1939 - Creating cache directory \"/var/lib/redpanda/data/cloud_storage_cache\"\n","stream":"stderr","time":"2024-12-19T22:25:32.331087822Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 0:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.332893147Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 3:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.332945615Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 5:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.332951426Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 1:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.332960303Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 2:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.332961645Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 6:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.332962998Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 4:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.332968608Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 7:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.332969971Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 8:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.332971213Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 9:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.332972526Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 10:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.332973808Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 12:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.333011148Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 14:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.333014043Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 16:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.333015616Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 15:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.333016989Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 18:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.333018261Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 19:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.333019534Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 17:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.333037077Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 22:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.33303914Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 21:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.333040513Z"} | |
{"log":"INFO 2024-12-19 22:25:32,332 [shard 11:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.333041745Z"} | |
{"log":"INFO 2024-12-19 22:25:32,333 [shard 13:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.333043008Z"} | |
{"log":"INFO 2024-12-19 22:25:32,333 [shard 20:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.33304424Z"} | |
{"log":"INFO 2024-12-19 22:25:32,333 [shard 23:main] ossl-library-context-service - ossl_context_service.cc:254 - OpenSSL Context loaded and ready\n","stream":"stderr","time":"2024-12-19T22:25:32.333045613Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 0:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523625173Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 2:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523643207Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 5:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523645992Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 4:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523648136Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 9:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.52365012Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 10:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523652133Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 3:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523654127Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 1:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523665689Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 14:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523670538Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 16:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523672993Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 15:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523675076Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 8:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.52367714Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 19:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523684224Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 18:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523686879Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 11:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523688722Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 20:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523690555Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 7:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523692199Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 13:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523694102Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 12:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523695976Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 17:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.52369806Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 21:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523699923Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 22:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523704692Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 6:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523861135Z"} | |
{"log":"INFO 2024-12-19 22:25:32,523 [shard 23:main] rpc - server.cc:41 - Creating net::server for internal_rpc with config {{://0.0.0.0:33145:PLAINTEXT}, max_service_memory_per_core: 30635196, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.523863991Z"} | |
{"log":"INFO 2024-12-19 22:25:32,526 [shard 0:main] features - feature_table.cc:434 - Activating features from bootstrap version 7\n","stream":"stderr","time":"2024-12-19T22:25:32.52642933Z"} | |
{"log":"INFO 2024-12-19 22:25:32,526 [shard 0:main] main - application.cc:2460 - Generated new UUID for node: 43671e8f-c60b-419b-b10e-e66740b3e36e\n","stream":"stderr","time":"2024-12-19T22:25:32.526644894Z"} | |
{"log":"INFO 2024-12-19 22:25:32,536 [shard 0:main] storage - segment.cc:811 - Creating new segment /var/lib/redpanda/data/redpanda/kvstore/0_0/0-0-v1.log\n","stream":"stderr","time":"2024-12-19T22:25:32.536623914Z"} | |
{"log":"INFO 2024-12-19 22:25:32,538 [shard 0:main] main - application.cc:2509 - Started RPC server listening at {host: 0.0.0.0, port: 33145}\n","stream":"stderr","time":"2024-12-19T22:25:32.538278897Z"} | |
{"log":"INFO 2024-12-19 22:25:32,538 [shard 0:main] main - application.cc:2615 - Starting Redpanda with node_id 0, cluster UUID {nullopt}\n","stream":"stderr","time":"2024-12-19T22:25:32.538486836Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 0:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539526275Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 1:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539532777Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 2:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539534921Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 3:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539536785Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 8:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539538087Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 5:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539540221Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 7:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539541584Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 12:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539542846Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 4:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539557464Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 17:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539561772Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 9:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539563635Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 20:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539570238Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 19:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.53957174Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 21:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539573013Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 13:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.53967807Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 14:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539680685Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 11:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539682809Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 18:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539684432Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 15:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539685955Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 10:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539687467Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 16:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539689231Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 22:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539690744Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 6:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539692377Z"} | |
{"log":"INFO 2024-12-19 22:25:32,539 [shard 23:main] raft - coordinated_recovery_throttle.cc:126 - Starting recovery throttle, rate: 104857600\n","stream":"stderr","time":"2024-12-19T22:25:32.539721662Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 0:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540191993Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 1:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540202733Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 2:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540205168Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 8:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540207192Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 13:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540209196Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 7:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540211129Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 15:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540214285Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 6:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540216209Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 10:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540224003Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 5:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540226057Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 22:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540228321Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 12:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540238751Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 20:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540241676Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 21:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540243259Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 14:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540245013Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 16:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540255903Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 11:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540377Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 18:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540383282Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 19:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540385316Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 9:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540386969Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 3:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540388372Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 17:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540389834Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 4:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.540391247Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 23:main] cluster - producer_state_manager.cc:45 - Started producer state manager\n","stream":"stderr","time":"2024-12-19T22:25:32.54039265Z"} | |
{"log":"INFO 2024-12-19 22:25:32,540 [shard 0:main] main - application.cc:1583 - Partition manager started\n","stream":"stderr","time":"2024-12-19T22:25:32.540850939Z"} | |
{"log":"INFO 2024-12-19 22:25:32,543 [shard 0:main] main - application.cc:1671 - Archiver service setup, cloud_storage_enabled: false, legacy_upload_mode_enabled: true\n","stream":"stderr","time":"2024-12-19T22:25:32.54336938Z"} | |
{"log":"INFO 2024-12-19 22:25:32,544 [shard 0:main] resource_mgmt - storage.cc:182 - Setting new target log data size 526.229GiB. Disk size 877.049GiB reservation percent 25 target percent {80} bytes {nullopt}\n","stream":"stderr","time":"2024-12-19T22:25:32.544418808Z"} | |
{"log":"INFO 2024-12-19 22:25:32,620 [shard 0:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.620885944Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 3:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.621667669Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 2:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.621675885Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 5:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.621678449Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 11:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.621680674Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 8:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.621682888Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 4:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.621684922Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 15:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.621687216Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 6:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.62168922Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 10:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.621691374Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 9:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.621693357Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 12:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.621695401Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 19:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.621701723Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 18:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.621720769Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 1:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.621722863Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 20:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.621724325Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 23:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.621725738Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 14:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.621727281Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 22:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.621728704Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 13:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.621730267Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 7:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.621923589Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 21:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.62274021Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 16:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.622747604Z"} | |
{"log":"INFO 2024-12-19 22:25:32,621 [shard 17:main] kafka - server.cc:41 - Creating net::server for kafka_rpc with config {{://0.0.0.0:9092:PLAINTEXT}, max_service_memory_per_core: 45952794, metrics_enabled:true, listen_backlog:{nullopt}, tcp_recv_buf:{nullopt}, tcp_send_buf:{nullopt}, stream_recv_buf:{nullopt}}\n","stream":"stderr","time":"2024-12-19T22:25:32.622755679Z"} | |
{"log":"INFO 2024-12-19 22:25:32,636 [shard 0:main] cluster - controller.cc:1154 - persisted initial configuration invariants: { version: 0, node_id: 0, core_count: 24 }\n","stream":"stderr","time":"2024-12-19T22:25:32.636224401Z"} | |
{"log":"INFO 2024-12-19 22:25:32,636 [shard 0:main] cluster - raft0_utils.h:30 - Current node is a cluster founder\n","stream":"stderr","time":"2024-12-19T22:25:32.636230903Z"} | |
{"log":"INFO 2024-12-19 22:25:32,657 [shard 0:main] raft - [group_id:0, {redpanda/controller/0}] consensus.cc:1402 - Starting with voted_for {id: -2147483648, revision: -9223372036854775808} term 0 initial_state true\n","stream":"stderr","time":"2024-12-19T22:25:32.657273615Z"} | |
{"log":"INFO 2024-12-19 22:25:32,677 [shard 0:main] raft - [group_id:0, {redpanda/controller/0}] consensus.cc:1446 - Current log offsets: {start_offset:-9223372036854775808, committed_offset:-9223372036854775808, committed_offset_term:-9223372036854775808, dirty_offset:-9223372036854775808, dirty_offset_term:-9223372036854775808}, read bootstrap state: data_seen 0 config_seen 0 eol false commit 0 term 0 prev_idx 0 prev_term 0 config_tracker -9223372036854775808 commit_base_tracker -9223372036854775808 configurations []\n","stream":"stderr","time":"2024-12-19T22:25:32.677491992Z"} | |
{"log":"INFO 2024-12-19 22:25:32,677 [shard 0:main] raft - [group_id:0, {redpanda/controller/0}] consensus.cc:1473 - Truncating configurations at -9223372036854775808\n","stream":"stderr","time":"2024-12-19T22:25:32.677497933Z"} | |
{"log":"INFO 2024-12-19 22:25:32,697 [shard 0:main] raft - [group_id:0, {redpanda/controller/0}] consensus.cc:935 - starting pre-vote leader election, current term: 0, leadership transfer: false\n","stream":"stderr","time":"2024-12-19T22:25:32.69768443Z"} | |
{"log":"INFO 2024-12-19 22:25:32,707 [shard 0:main] raft - [group_id:0, {redpanda/controller/0}] consensus.cc:1584 - started raft, log offsets: {start_offset:-9223372036854775808, committed_offset:-9223372036854775808, committed_offset_term:-9223372036854775808, dirty_offset:-9223372036854775808, dirty_offset_term:-9223372036854775808}, term: 0, configuration: {current: {voters: {{id: 0, revision: 0}}, learners: {}}, old:{nullopt}, revision: 0, update: {nullopt}, version: 4}, brokers: {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.707757876Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 0:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.71116265Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 1:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711176716Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 2:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711181465Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 5:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.71118391Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 7:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711186745Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 4:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711188749Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 8:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711196483Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 10:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711203065Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 11:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711205811Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 15:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711207764Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 12:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711209457Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 16:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711211141Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 6:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711212874Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 18:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711214617Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 14:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711216641Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 13:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711222231Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 9:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711224876Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 19:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.71122679Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 23:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711228724Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 20:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711230597Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 3:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711239604Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 21:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711244814Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 17:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711246948Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 22:main] cluster - drain_manager.cc:21 - Drain manager starting\n","stream":"stderr","time":"2024-12-19T22:25:32.711257578Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 0:main] cluster - members_manager.cc:98 - starting members manager with founding brokers: {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.711383644Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 0:main] cluster - controller.cc:550 - Controller log replay starting (to offset -9223372036854775808)\n","stream":"stderr","time":"2024-12-19T22:25:32.711389575Z"} | |
{"log":"INFO 2024-12-19 22:25:32,711 [shard 0:main] cluster - controller.cc:561 - Controller log replay complete.\n","stream":"stderr","time":"2024-12-19T22:25:32.711391709Z"} | |
{"log":"INFO 2024-12-19 22:25:32,717 [shard 0:main] raft - [group_id:0, {redpanda/controller/0}] vote_stm.cc:421 - becoming the leader term:1\n","stream":"stderr","time":"2024-12-19T22:25:32.718004647Z"} | |
{"log":"INFO 2024-12-19 22:25:32,718 [shard 0:main] storage - segment.cc:811 - Creating new segment /var/lib/redpanda/data/redpanda/controller/0_0/0-1-v1.log\n","stream":"stderr","time":"2024-12-19T22:25:32.718072254Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 0:main] cluster - members_manager.cc:208 - processing raft-0 configuration at offset: 0 with brokers: [{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}]\n","stream":"stderr","time":"2024-12-19T22:25:32.738677035Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 0:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738684189Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 0:main] raft - [group_id:0, {redpanda/controller/0}] vote_stm.cc:436 - became the leader term: 1\n","stream":"stderr","time":"2024-12-19T22:25:32.738691923Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 5:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738728392Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 9:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738731638Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 6:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738734002Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 2:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738736116Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 4:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738741486Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 13:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738744652Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 12:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738747207Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 7:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738784537Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 19:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738797261Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 10:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738800276Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 8:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738802631Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 23:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738805025Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 14:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738807209Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 16:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738809784Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 15:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738812199Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 1:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738814263Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 3:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738816296Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 11:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.7388184Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 22:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.73882357Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 20:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738843457Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 17:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738847655Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 21:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738886247Z"} | |
{"log":"INFO 2024-12-19 22:25:32,738 [shard 18:main] cluster - members_table.cc:82 - applying node add command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 0, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.738966408Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 0:main] cluster - controller.cc:950 - Creating cluster UUID 441852a4-ee0a-4991-8660-8da6773ec75e\n","stream":"stderr","time":"2024-12-19T22:25:32.811456637Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 0:main] cluster - bootstrap_backend.cc:92 - Applying update to bootstrap_manager\n","stream":"stderr","time":"2024-12-19T22:25:32.811584337Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 0:main] cluster - members_manager.cc:847 - Initial node UUID map: {43671e8f-c60b-419b-b10e-e66740b3e36e: 0}\n","stream":"stderr","time":"2024-12-19T22:25:32.811625614Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 0:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811633138Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 1:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811880782Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 5:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811883858Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 4:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811886072Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 3:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811892655Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 7:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811894719Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 8:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811896752Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 9:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811898456Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 12:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811899898Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 10:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811901301Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 6:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811902754Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 2:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811904186Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 11:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811905579Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 17:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811907082Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 16:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811910658Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 19:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811912121Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 13:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811913534Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 15:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811915027Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 18:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811916409Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 21:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811917952Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 20:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811919375Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 22:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811920767Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 23:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.81192219Z"} | |
{"log":"INFO 2024-12-19 22:25:32,811 [shard 14:main] cluster - members_table.cc:98 - setting initial nodes {{id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.811923643Z"} | |
{"log":"INFO 2024-12-19 22:25:32,821 [shard 0:main] features - feature_table.cc:434 - Activating features from bootstrap version 13\n","stream":"stderr","time":"2024-12-19T22:25:32.821934672Z"} | |
{"log":"INFO 2024-12-19 22:25:32,822 [shard 0:main] cluster - feature_backend.cc:149 - Saving feature_table_snapshot at version 13...\n","stream":"stderr","time":"2024-12-19T22:25:32.822396929Z"} | |
{"log":"INFO 2024-12-19 22:25:32,842 [shard 0:main] cluster - controller.cc:958 - Cluster UUID created 441852a4-ee0a-4991-8660-8da6773ec75e\n","stream":"stderr","time":"2024-12-19T22:25:32.842680688Z"} | |
{"log":"INFO 2024-12-19 22:25:32,842 [shard 0:main] cluster - controller_backend.cc:813 - Cleaning up orphan topic files. bootstrap_revision: -9223372036854775808\n","stream":"stderr","time":"2024-12-19T22:25:32.842947648Z"} | |
{"log":"INFO 2024-12-19 22:25:32,844 [shard 0:main] cluster - feature_manager.cc:92 - Starting...\n","stream":"stderr","time":"2024-12-19T22:25:32.844052269Z"} | |
{"log":"INFO 2024-12-19 22:25:32,844 [shard 0:main] cluster - feature_manager.cc:531 - Activating features after upgrade...\n","stream":"stderr","time":"2024-12-19T22:25:32.844058701Z"} | |
{"log":"INFO 2024-12-19 22:25:32,844 [shard 0:main] cluster - feature_manager.cc:540 - Activating feature broker_time_based_retention (logical version 13)\n","stream":"stderr","time":"2024-12-19T22:25:32.844060555Z"} | |
{"log":"INFO 2024-12-19 22:25:32,844 [shard 0:main] cluster - metrics_reporter.cc:335 - Waiting to initialize cluster metrics ID...\n","stream":"stderr","time":"2024-12-19T22:25:32.844103576Z"} | |
{"log":"INFO 2024-12-19 22:25:32,844 [shard 0:main] cluster - partition_balancer_backend.cc:101 - partition balancer started\n","stream":"stderr","time":"2024-12-19T22:25:32.844112262Z"} | |
{"log":"INFO 2024-12-19 22:25:32,844 [shard 0:main] data-migrate - data_migration_backend.cc:104 - backend starting\n","stream":"stderr","time":"2024-12-19T22:25:32.844336613Z"} | |
{"log":"INFO 2024-12-19 22:25:32,844 [shard 0:main] data-migrate - data_migration_backend.cc:165 - backend not started as cloud_storage_api is not available\n","stream":"stderr","time":"2024-12-19T22:25:32.844338867Z"} | |
{"log":"INFO 2024-12-19 22:25:32,844 [shard 0:main] cluster - leader_balancer.cc:104 - Leader balancer: controller leadership detected. Starting rebalancer in 30 seconds\n","stream":"stderr","time":"2024-12-19T22:25:32.84467653Z"} | |
{"log":"INFO 2024-12-19 22:25:32,844 [shard 0:main] cluster - config_manager.cc:164 - Importing property auto_create_topics_enabled:1\n","stream":"stderr","time":"2024-12-19T22:25:32.844704442Z"} | |
{"log":"INFO 2024-12-19 22:25:32,844 [shard 0:main] cluster - config_manager.cc:164 - Importing property storage_min_free_bytes:10485760\n","stream":"stderr","time":"2024-12-19T22:25:32.844714731Z"} | |
{"log":"INFO 2024-12-19 22:25:32,844 [shard 0:main] cluster - config_manager.cc:164 - Importing property group_initial_rebalance_delay:0\n","stream":"stderr","time":"2024-12-19T22:25:32.844716114Z"} | |
{"log":"INFO 2024-12-19 22:25:32,844 [shard 0:main] cluster - config_manager.cc:164 - Importing property write_caching_default:true\n","stream":"stderr","time":"2024-12-19T22:25:32.844721003Z"} | |
{"log":"INFO 2024-12-19 22:25:32,844 [shard 0:main] cluster - config_manager.cc:164 - Importing property fetch_reads_debounce_timeout:10\n","stream":"stderr","time":"2024-12-19T22:25:32.844722326Z"} | |
{"log":"INFO 2024-12-19 22:25:32,844 [shard 0:main] cluster - config_manager.cc:164 - Importing property log_segment_size_min:{1}\n","stream":"stderr","time":"2024-12-19T22:25:32.844744648Z"} | |
{"log":"INFO 2024-12-19 22:25:32,844 [shard 0:main] cluster - config_manager.cc:164 - Importing property group_topic_partitions:3\n","stream":"stderr","time":"2024-12-19T22:25:32.844755798Z"} | |
{"log":"INFO 2024-12-19 22:25:32,846 [shard 0:main] cluster - feature_backend.cc:149 - Saving feature_table_snapshot at version 13...\n","stream":"stderr","time":"2024-12-19T22:25:32.846032642Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 0:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847626861Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 0:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.847634455Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 3:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847726308Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 6:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847730245Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 3:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.8477326Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 6:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.847734613Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 2:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847736567Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 2:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.84773842Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 9:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847740164Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 1:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847742388Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 10:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847753489Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 10:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.847793694Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 13:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847797802Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 17:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847800006Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 11:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847801929Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 9:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.847803833Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 11:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.847805687Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 19:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.84780762Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 15:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847814844Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 16:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847831104Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 20:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.84783972Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 8:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847842445Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 20:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.84784488Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 8:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.847847094Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 23:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847849198Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 18:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847851202Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 5:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847853216Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 22:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847855229Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 12:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847857273Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 12:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.847888452Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 22:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.847895114Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 21:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847897459Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 1:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.847899753Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 4:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847902127Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 19:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.847904331Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 14:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847906646Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 18:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.847912737Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 14:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.847914831Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 4:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.847916685Z"} | |
{"log":"WARN 2024-12-19 22:25:32,847 [shard 7:main] admin_api_server - server.cc:539 - Insecure Admin API listener on 0.0.0.0:9644, consider enabling `admin_api_require_auth`\n","stream":"stderr","time":"2024-12-19T22:25:32.847918528Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 15:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.847922105Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 7:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.847923598Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 21:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.8479249Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 17:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.847926323Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 23:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.847927645Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 13:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.847929038Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 5:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.847948184Z"} | |
{"log":"INFO 2024-12-19 22:25:32,847 [shard 16:main] admin_api_server - server.cc:349 - Started HTTP admin service listening at {{:{host: 0.0.0.0, port: 9644}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.847991695Z"} | |
{"log":"INFO 2024-12-19 22:25:32,851 [shard 0:main] resource_mgmt - storage.cc:73 - Starting disk space manager service (enabled)\n","stream":"stderr","time":"2024-12-19T22:25:32.851115362Z"} | |
{"log":"INFO 2024-12-19 22:25:32,853 [shard 0:main] cluster - metrics_reporter.cc:379 - Generated cluster metrics ID 5083e540-3f03-4bf5-b6bd-041da4c098f1\n","stream":"stderr","time":"2024-12-19T22:25:32.853297964Z"} | |
{"log":"INFO 2024-12-19 22:25:32,854 [shard 0:main] cluster - config_manager.cc:136 - Completed bootstrap as leader\n","stream":"stderr","time":"2024-12-19T22:25:32.854525295Z"} | |
{"log":"INFO 2024-12-19 22:25:32,854 [shard 0:main] cluster - config_manager.cc:124 - Bootstrap complete (version 1)\n","stream":"stderr","time":"2024-12-19T22:25:32.854530815Z"} | |
{"log":"INFO 2024-12-19 22:25:32,854 [shard 0:main] main - application.cc:2656 - Started Pandaproxy listening at {{:{host: 0.0.0.0, port: 8082}:\u003cnullopt\u003e}}\n","stream":"stderr","time":"2024-12-19T22:25:32.854542417Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 0:main] cluster - members_manager.cc:441 - applying node update command - broker: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}, offset: 6\n","stream":"stderr","time":"2024-12-19T22:25:32.856067476Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 0:main] cluster - members_manager.cc:542 - processing node update command - broker: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}, offset: 6\n","stream":"stderr","time":"2024-12-19T22:25:32.856082074Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 0:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856084749Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 0:main] cluster - metrics_reporter.cc:409 - Initialized cluster_id to 5083e540-3f03-4bf5-b6bd-041da4c098f1\n","stream":"stderr","time":"2024-12-19T22:25:32.856087935Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 1:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856124213Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 18:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.85615495Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 3:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856158196Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 12:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856160501Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 16:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856162595Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 6:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856164739Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 4:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856167093Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 20:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856177523Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 21:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856179887Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 23:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856182151Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 13:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856184355Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 10:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856186539Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 11:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856191218Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 15:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856193533Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 5:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856195687Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 17:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856197901Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 2:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856200726Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 22:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856205816Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 19:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.85620808Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 8:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856215714Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 7:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856235181Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 9:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856239379Z"} | |
{"log":"INFO 2024-12-19 22:25:32,856 [shard 14:main] cluster - members_table.cc:111 - applying update node config command for: {id: 0, kafka_advertised_listeners: {{:{host: 127.0.0.1, port: 9092}}}, rpc_address: {host: 127.0.0.1, port: 33145}, rack: {nullopt}, properties: {cores 24, mem_available 4177526784, disk_available 877, in_fips_mode disabled}}\n","stream":"stderr","time":"2024-12-19T22:25:32.856345758Z"} | |
{"log":"INFO 2024-12-19 22:25:32,863 [shard 0:main] cluster - members_manager.cc:164 - Node configuration updated successfully\n","stream":"stderr","time":"2024-12-19T22:25:32.863030501Z"} | |
{"log":"INFO 2024-12-19 22:25:32,919 [shard 0:main] main - application.cc:2664 - Started Schema Registry listening at {{:{host: 0.0.0.0, port: 8081}:\u003cnullopt\u003e}}\n","stream":"stderr","time":"2024-12-19T22:25:32.919950678Z"} | |
{"log":"INFO 2024-12-19 22:25:32,920 [shard 0:main] main - application.cc:2999 - Waiting for cluster membership\n","stream":"stderr","time":"2024-12-19T22:25:32.920064652Z"} | |
{"log":"INFO 2024-12-19 22:25:32,920 [shard 0:main] main - application.cc:3028 - Started Kafka API server listening at {{:{host: 0.0.0.0, port: 9092}:{nullopt}}}\n","stream":"stderr","time":"2024-12-19T22:25:32.920550072Z"} | |
{"log":"INFO 2024-12-19 22:25:32,920 [shard 0:main] main - application.cc:2707 - Successfully started Redpanda!\n","stream":"stderr","time":"2024-12-19T22:25:32.920899036Z"} | |
{"log":"INFO 2024-12-19 22:25:35,845 [shard 0:main] cluster - partition_balancer_planner.cc:1972 - counts rebalancing objective in domain 0: 1 -\u003e 1\n","stream":"stderr","time":"2024-12-19T22:25:35.845417393Z"} | |
{"log":"INFO 2024-12-19 22:25:35,845 [shard 0:main] cluster - partition_balancer_planner.cc:1972 - counts rebalancing objective in domain -1: 1 -\u003e 1\n","stream":"stderr","time":"2024-12-19T22:25:35.845428885Z"} | |
{"log":"INFO 2024-12-19 22:25:35,845 [shard 0:main] cluster - partition_balancer_backend.cc:423 - last status: in_progress; violations: unavailable nodes: 0, full nodes: 0; nodes to rebalance count: 1; on demand rebalance requested: false; updates in progress: 0; action counts: reassignments: 0, cancellations: 0, failed: 0; counts rebalancing finished: true, force refresh health report: false\n","stream":"stderr","time":"2024-12-19T22:25:35.845435798Z"} | |
{"log":"INFO 2024-12-19 22:25:35,845 [shard 0:main] cluster - members_manager.cc:375 - applying finish_reallocations_cmd, offset: 8, node id: 0\n","stream":"stderr","time":"2024-12-19T22:25:35.845604384Z"} | |
{"log":"INFO 2024-12-19 22:26:32,322 [shard 0:main] main - memory_sampling.cc:143 - 34812723 bytes of available memory left - Top-N alloc sites: size: 33554432 count: 2 at: 0xa66e23a 0xa2aadd6 0xa2b43d6 0xa2a6bce 0x96b50f7 0x96b5e46 0x96b49d3 0x32ff1a3 0x335f9a2 0xa49ffeb|size: 18000222 count: 6 at: 0xa66e23a 0xa2cb4a0 0xa2b3b57 0xa2a6bce 0x3656fe1 0x365780c 0xa39f22f 0xa3a2951 0xa39fb16 0xa27de00 0xa27c1f8 0x32fc17d 0xa8ea329 /opt/redpanda/lib/libc.so.6+0x2a087 /opt/redpanda/lib/libc.so.6+0x2a14a 0x32f4624|size: 16777216 count: 1 at: 0xa66e23a 0xa2aadd6 0xa2b43d6 0xa2a6bce 0x96afa9c 0x32ff194 0x335f9a2 0xa49ffeb|size: 9000111 count: 3 at: 0xa66e23a 0xa2aadd6 0xa2b34cf 0xa2bf9c8 0x969843f 0x344c8da 0xa39f22f 0xa3a2951 0xa39fb16 0xa27de00 0xa27c1f8 0x32fc17d 0xa8ea329 /opt/redpanda/lib/libc.so.6+0x2a087 /opt/redpanda/lib/libc.so.6+0x2a14a 0x32f4624|size: 8388608 count: 1 at: 0xa66e23a 0xa2aadd6 0xa2b43d6 0xa2a6bce 0x96c469d 0x96c4b7f 0x32ff1bc 0x335f9a2 0xa49ffeb\n","stream":"stderr","time":"2024-12-19T22:26:32.322952489Z"} | |
{"log":"INFO 2024-12-19 22:26:32,842 [shard 0:main] cluster - controller_stm.cc:136 - created snapshot at offset 8 in 0 ms\n","stream":"stderr","time":"2024-12-19T22:26:32.842758035Z"} | |
{"log":"INFO 2024-12-19 22:26:32,853 [shard 0:main] storage - disk_log_impl.cc:2486 - Removing \"/var/lib/redpanda/data/redpanda/controller/0_0/0-1-v1.log\" (remove_prefix_full_segments, {offset_tracker:{term:1, base_offset:0, committed_offset:8, dirty_offset:8}, compacted_segment=0, finished_self_compaction=0, finished_windowed_compaction=0, generation=18, reader={/var/lib/redpanda/data/redpanda/controller/0_0/0-1-v1.log, (1545 bytes)}, writer={no_of_chunks:64, closed:0, fallocation_offset:33554432, stable_offset:1545, flushed_offset:1545, committed_offset:1545, inflight:0, dispatched:0, merged:0, bytes_flush_pending:0}, cache={cache_size=9, dirty tracker: {min: -9223372036854775808, max: -9223372036854775808}}, compaction_index:nullopt, closed=0, tombstone=0, index={file:/var/lib/redpanda/data/redpanda/controller/0_0/0-1-v1.base_index, offsets:0, index:{header_bitflags:0, base_offset:0, max_offset:8, base_timestamp:{timestamp: 1734647132717}, max_timestamp:{timestamp: 1734647135845}, batch_timestamps_are_monotonic:1, with_offset:false, non_data_timestamps:0, broker_timestamp:{{timestamp: 1734647135845}}, num_compactible_records_appended:{8}, index(1,1,1)}, step:32768, needs_persistence:1}})\n","stream":"stderr","time":"2024-12-19T22:26:32.853323945Z"} | |
{"log":"INFO 2024-12-19 22:26:36,192 [shard 1:main] cluster - topics_frontend.cc:210 - Create topics [{configuration: { topic: {kafka/atr}, partition_count: 150, replication_factor: 3, is_migrated: false, properties: {compression: {nullopt}, cleanup_policy_bitflags: {nullopt}, compaction_strategy: {nullopt}, retention_bytes: {}, retention_duration_ms: {}, segment_size: {nullopt}, timestamp_type: {nullopt}, recovery_enabled: {nullopt}, shadow_indexing: {disabled}, read_replica: {nullopt}, read_replica_bucket: {nullopt}, remote_topic_namespace_override: {nullopt}, remote_topic_properties: {nullopt}, batch_max_bytes: {nullopt}, retention_local_target_bytes: {}, retention_local_target_ms: {}, remote_delete: true, segment_ms: {}, record_key_schema_id_validation: {nullopt}, record_key_schema_id_validation_compat: {nullopt}, record_key_subject_name_strategy: {nullopt}, record_key_subject_name_strategy_compat: {nullopt}, record_value_schema_id_validation: {nullopt}, record_value_schema_id_validation_compat: {nullopt}, record_value_subject_name_strategy: {nullopt}, record_value_subject_name_strategy_compat: {nullopt}, initial_retention_local_target_bytes: {}, initial_retention_local_target_ms: {}, mpx_virtual_cluster_id: {nullopt}, write_caching: {nullopt}, flush_ms: {nullopt}, flush_bytes: {nullopt}, remote_label: {nullopt}}}, custom_assignments: {}},]\n","stream":"stderr","time":"2024-12-19T22:26:36.192474594Z"} | |
{"log":"WARN 2024-12-19 22:26:36,192 [shard 1:main] kafka - create_topics.cc:159 - Failed to create topic(s) {{kafka/atr}} error_code observed: cluster::errc::topic_invalid_replication_factor\n","stream":"stderr","time":"2024-12-19T22:26:36.192546299Z"} | |
{"log":"INFO 2024-12-19 22:26:43,377 [shard 1:main] cluster - topics_frontend.cc:210 - Create topics [{configuration: { topic: {kafka/atr}, partition_count: 1, replication_factor: 3, is_migrated: false, properties: {compression: {nullopt}, cleanup_policy_bitflags: {nullopt}, compaction_strategy: {nullopt}, retention_bytes: {}, retention_duration_ms: {}, segment_size: {nullopt}, timestamp_type: {nullopt}, recovery_enabled: {nullopt}, shadow_indexing: {disabled}, read_replica: {nullopt}, read_replica_bucket: {nullopt}, remote_topic_namespace_override: {nullopt}, remote_topic_properties: {nullopt}, batch_max_bytes: {nullopt}, retention_local_target_bytes: {}, retention_local_target_ms: {}, remote_delete: true, segment_ms: {}, record_key_schema_id_validation: {nullopt}, record_key_schema_id_validation_compat: {nullopt}, record_key_subject_name_strategy: {nullopt}, record_key_subject_name_strategy_compat: {nullopt}, record_value_schema_id_validation: {nullopt}, record_value_schema_id_validation_compat: {nullopt}, record_value_subject_name_strategy: {nullopt}, record_value_subject_name_strategy_compat: {nullopt}, initial_retention_local_target_bytes: {}, initial_retention_local_target_ms: {}, mpx_virtual_cluster_id: {nullopt}, write_caching: {nullopt}, flush_ms: {nullopt}, flush_bytes: {nullopt}, remote_label: {nullopt}}}, custom_assignments: {}},]\n","stream":"stderr","time":"2024-12-19T22:26:43.37738497Z"} | |
{"log":"WARN 2024-12-19 22:26:43,377 [shard 1:main] kafka - create_topics.cc:159 - Failed to create topic(s) {{kafka/atr}} error_code observed: cluster::errc::topic_invalid_replication_factor\n","stream":"stderr","time":"2024-12-19T22:26:43.377415417Z"} | |
{"log":"INFO 2024-12-19 22:26:50,994 [shard 1:main] cluster - topics_frontend.cc:210 - Create topics [{configuration: { topic: {kafka/atr}, partition_count: 3, replication_factor: 3, is_migrated: false, properties: {compression: {nullopt}, cleanup_policy_bitflags: {nullopt}, compaction_strategy: {nullopt}, retention_bytes: {}, retention_duration_ms: {}, segment_size: {nullopt}, timestamp_type: {nullopt}, recovery_enabled: {nullopt}, shadow_indexing: {disabled}, read_replica: {nullopt}, read_replica_bucket: {nullopt}, remote_topic_namespace_override: {nullopt}, remote_topic_properties: {nullopt}, batch_max_bytes: {nullopt}, retention_local_target_bytes: {}, retention_local_target_ms: {}, remote_delete: true, segment_ms: {}, record_key_schema_id_validation: {nullopt}, record_key_schema_id_validation_compat: {nullopt}, record_key_subject_name_strategy: {nullopt}, record_key_subject_name_strategy_compat: {nullopt}, record_value_schema_id_validation: {nullopt}, record_value_schema_id_validation_compat: {nullopt}, record_value_subject_name_strategy: {nullopt}, record_value_subject_name_strategy_compat: {nullopt}, initial_retention_local_target_bytes: {}, initial_retention_local_target_ms: {}, mpx_virtual_cluster_id: {nullopt}, write_caching: {nullopt}, flush_ms: {nullopt}, flush_bytes: {nullopt}, remote_label: {nullopt}}}, custom_assignments: {}},]\n","stream":"stderr","time":"2024-12-19T22:26:50.99467825Z"} | |
{"log":"WARN 2024-12-19 22:26:50,994 [shard 1:main] kafka - create_topics.cc:159 - Failed to create topic(s) {{kafka/atr}} error_code observed: cluster::errc::topic_invalid_replication_factor\n","stream":"stderr","time":"2024-12-19T22:26:50.994745817Z"} | |
{"log":"INFO 2024-12-19 22:27:03,188 [shard 1:main] cluster - topics_frontend.cc:210 - Create topics [{configuration: { topic: {kafka/atr}, partition_count: 2, replication_factor: 3, is_migrated: false, properties: {compression: {nullopt}, cleanup_policy_bitflags: {nullopt}, compaction_strategy: {nullopt}, retention_bytes: {}, retention_duration_ms: {}, segment_size: {nullopt}, timestamp_type: {nullopt}, recovery_enabled: {nullopt}, shadow_indexing: {disabled}, read_replica: {nullopt}, read_replica_bucket: {nullopt}, remote_topic_namespace_override: {nullopt}, remote_topic_properties: {nullopt}, batch_max_bytes: {nullopt}, retention_local_target_bytes: {}, retention_local_target_ms: {}, remote_delete: true, segment_ms: {}, record_key_schema_id_validation: {nullopt}, record_key_schema_id_validation_compat: {nullopt}, record_key_subject_name_strategy: {nullopt}, record_key_subject_name_strategy_compat: {nullopt}, record_value_schema_id_validation: {nullopt}, record_value_schema_id_validation_compat: {nullopt}, record_value_subject_name_strategy: {nullopt}, record_value_subject_name_strategy_compat: {nullopt}, initial_retention_local_target_bytes: {}, initial_retention_local_target_ms: {}, mpx_virtual_cluster_id: {nullopt}, write_caching: {nullopt}, flush_ms: {nullopt}, flush_bytes: {nullopt}, remote_label: {nullopt}}}, custom_assignments: {}},]\n","stream":"stderr","time":"2024-12-19T22:27:03.188162697Z"} | |
{"log":"WARN 2024-12-19 22:27:03,188 [shard 1:main] kafka - create_topics.cc:159 - Failed to create topic(s) {{kafka/atr}} error_code observed: cluster::errc::topic_invalid_replication_factor\n","stream":"stderr","time":"2024-12-19T22:27:03.188218762Z"} | |
{"log":"INFO 2024-12-19 22:35:32,625 [shard 0:main] kafka - group_manager.cc:234 - Group offset retention is now enabled (prev {nullopt}). Legacy enabled false retention_sec {604800000} original version 13.\n","stream":"stderr","time":"2024-12-19T22:35:32.62559986Z"} |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment