Repository Metadata Analyzer
Cache¶
quarkus.cache.caffeine."metaAnalyzer".expire-after-write¶
Defines the time-to-live of cache entries.
Required | true |
---|---|
Type | duration |
Default | PT2H |
ENV | QUARKUS_CACHE_CAFFEINE__METAANALYZER__EXPIRE_AFTER_WRITE |
quarkus.cache.caffeine."metaAnalyzer".initial-capacity¶
Defines the initial capacity of the cache.
Required | true |
---|---|
Type | integer |
Default | 5 |
ENV | QUARKUS_CACHE_CAFFEINE__METAANALYZER__INITIAL_CAPACITY |
quarkus.cache.enabled¶
Defines whether caching of analysis results shall be enabled.
Required | true |
---|---|
Type | boolean |
Default | true |
ENV | QUARKUS_CACHE_ENABLED |
Database¶
quarkus.datasource.jdbc.url¶
Specifies the JDBC URL to use when connecting to the database.
Required | true |
---|---|
Type | string |
Default | null |
ENV | QUARKUS_DATASOURCE_JDBC_URL |
quarkus.datasource.password¶
Specifies the password to use when authenticating to the database.
Required | true |
---|---|
Type | string |
Default | null |
ENV | QUARKUS_DATASOURCE_PASSWORD |
quarkus.datasource.username¶
Specifies the username to use when authenticating to the database.
Required | true |
---|---|
Type | string |
Default | null |
ENV | QUARKUS_DATASOURCE_USERNAME |
General¶
secret.key.path¶
Defines the path to the secret key to be used for data encryption and decryption.
Required | false |
---|---|
Type | string |
Default | ~/.dependency-track/keys/secret.key |
ENV | SECRET_KEY_PATH |
HTTP¶
quarkus.http.port¶
HTTP port to listen on. Application metrics will be available via this port.
Required | false |
---|---|
Type | integer |
Default | 8091 |
ENV | QUARKUS_HTTP_PORT |
Kafka¶
dt.kafka.topic.prefix¶
Defines an optional prefix to assume for all Kafka topics the application consumes from, or produces to. The prefix will also be prepended to the application's consumer group ID.
Required | false |
---|---|
Type | string |
Default | null |
Example | acme- |
ENV | DT_KAFKA_TOPIC_PREFIX |
kafka-streams.auto.offset.reset¶
Refer to https://kafka.apache.org/documentation/#consumerconfigs_auto.offset.reset for details.
Required | false |
---|---|
Type | enum |
Valid Values | [earliest, latest, none] |
Default | earliest |
ENV | KAFKA_STREAMS_AUTO_OFFSET_RESET |
kafka-streams.commit.interval.ms¶
Defines the interval in milliseconds at which consumer offsets are committed to the Kafka brokers. The Kafka default of 30s
has been modified to 5s
.
Refer to https://kafka.apache.org/documentation/#streamsconfigs_commit.interval.ms for details.
Required | false |
---|---|
Type | integer |
Default | 5000 |
ENV | KAFKA_STREAMS_COMMIT_INTERVAL_MS |
kafka-streams.exception.thresholds.deserialization.count¶
Defines the threshold for records failing to be deserialized within kafka-streams.exception.thresholds.deserialization.interval
. Deserialization failures within the threshold will be logged, failures exceeding the threshold cause the application to stop processing further records, and shutting down.
Required | true |
---|---|
Type | integer |
Default | 5 |
ENV | KAFKA_STREAMS_EXCEPTION_THRESHOLDS_DESERIALIZATION_COUNT |
kafka-streams.exception.thresholds.deserialization.interval¶
Defines the interval within which up to kafka-streams.exception.thresholds.deserialization.count
records are allowed to fail deserialization. Deserialization failures within the threshold will be logged, failures exceeding the threshold cause the application to stop processing further records, and shutting down.
Required | true |
---|---|
Type | duration |
Default | PT30M |
ENV | KAFKA_STREAMS_EXCEPTION_THRESHOLDS_DESERIALIZATION_INTERVAL |
kafka-streams.exception.thresholds.processing.count¶
Defines the threshold for records failing to be processed within kafka-streams.exception.thresholds.processing.interval
. Processing failures within the threshold will be logged, failures exceeding the threshold cause the application to stop processing further records, and shutting down.
Required | true |
---|---|
Type | integer |
Default | 50 |
ENV | KAFKA_STREAMS_EXCEPTION_THRESHOLDS_PROCESSING_COUNT |
kafka-streams.exception.thresholds.processing.interval¶
Defines the interval within which up to kafka-streams.exception.thresholds.processing.count
records are allowed to fail processing. Processing failures within the threshold will be logged, failures exceeding the threshold cause the application to stop processing further records, and shutting down.
Required | true |
---|---|
Type | duration |
Default | PT30M |
ENV | KAFKA_STREAMS_EXCEPTION_THRESHOLDS_PROCESSING_INTERVAL |
kafka-streams.exception.thresholds.production.count¶
Defines the threshold for records failing to be produced within kafka-streams.exception.thresholds.production.interval
. Production failures within the threshold will be logged, failures exceeding the threshold cause the application to stop processing further records, and shutting down.
Required | true |
---|---|
Type | integer |
Default | 5 |
ENV | KAFKA_STREAMS_EXCEPTION_THRESHOLDS_PRODUCTION_COUNT |
kafka-streams.exception.thresholds.production.interval¶
Defines the interval within which up to kafka-streams.exception.thresholds.production.count
records are allowed to fail producing. Production failures within the threshold will be logged, failures exceeding the threshold cause the application to stop processing further records, and shutting down.
Required | true |
---|---|
Type | duration |
Default | PT30M |
ENV | KAFKA_STREAMS_EXCEPTION_THRESHOLDS_PRODUCTION_INTERVAL |
kafka-streams.metrics.recording.level¶
Refer to https://kafka.apache.org/documentation/#adminclientconfigs_metrics.recording.level for details.
Required | false |
---|---|
Type | enum |
Valid Values | [INFO, DEBUG, TRACE] |
Default | DEBUG |
ENV | KAFKA_STREAMS_METRICS_RECORDING_LEVEL |
kafka-streams.num.stream.threads¶
Refer to https://kafka.apache.org/documentation/#streamsconfigs_num.stream.threads for details.
Required | true |
---|---|
Type | integer |
Default | 3 |
ENV | KAFKA_STREAMS_NUM_STREAM_THREADS |
kafka.bootstrap.servers¶
Comma-separated list of brokers to use for establishing the initial connection to the Kafka cluster.
Refer to https://kafka.apache.org/documentation/#consumerconfigs_bootstrap.servers for details.
Required | true |
---|---|
Type | string |
Default | null |
Example | broker-01.acme.com:9092,broker-02.acme.com:9092 |
ENV | KAFKA_BOOTSTRAP_SERVERS |
quarkus.kafka-streams.application-id¶
Defines the ID to uniquely identify this application in the Kafka cluster.
Refer to https://kafka.apache.org/documentation/#streamsconfigs_application.id for details.
Required | false |
---|---|
Type | string |
Default | ${dt.kafka.topic.prefix}hyades-repository-meta-analyzer |
ENV | QUARKUS_KAFKA_STREAMS_APPLICATION_ID |
Observability¶
quarkus.log.console.json¶
Defines whether logs should be written in JSON format.
Required | false |
---|---|
Type | boolean |
Default | false |
ENV | QUARKUS_LOG_CONSOLE_JSON |