Skip to content

Commit 506b6d4

Browse files
authored
Merge branch 'develop' into u/ch/store_BigQuery
2 parents 4a1c6ed + d1e5a6c commit 506b6d4

10 files changed

+85
-24
lines changed

broker/consumer/admin.properties

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
# Authentication information.
1+
# Kafka Admin client configs
22
# This file is part of a workflow that creates an authenticated connection to the Kafka broker.
33
# In cases where we can connect without authentication (e.g., ZTF), this file is not used.
44
# For config options, see https://kafka.apache.org/documentation/#adminclientconfigs
Lines changed: 5 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,11 +1,13 @@
1-
# Authentication information.
1+
# Kafka Admin client configs
22
# This file is part of a workflow that creates an authenticated connection to the Kafka broker.
3-
# For config options, see https://gcn.nasa.gov/docs/client#java
3+
# In cases where we can connect without authentication (e.g., ZTF), this file is not used.
4+
# For config options, see https://kafka.apache.org/documentation/#adminclientconfigs
5+
# For LVK-specific options, see https://gcn.nasa.gov/docs/client#java
46

57
security.protocol=SASL_SSL
68
sasl.mechanism=OAUTHBEARER
79
sasl.login.callback.handler.class=org.apache.kafka.common.security.oauthbearer.OAuthBearerLoginCallbackHandler
810
sasl.oauthbearer.token.endpoint.url=https://auth.gcn.nasa.gov/oauth2/token
911
sasl.jaas.config=org.apache.kafka.common.security.oauthbearer.OAuthBearerLoginModule required \
1012
clientId="CLIENT_ID" \
11-
clientSecret="CLIENT_SECRET";
13+
clientSecret="CLIENT_SECRET";

broker/consumer/lvk/ps-connector.properties

Lines changed: 10 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,13 @@
1-
# Adapted from:
1+
# Kafka Connect sink connector configs
2+
# For config options, see https://docs.confluent.io/platform/current/installation/configuration/connect/sink-connect-configs.html
3+
# For additional Pub/Sub-specific options, see https://github.com/googleapis/java-pubsub-group-kafka-connector?tab=readme-ov-file#sink-connector
4+
#
5+
# --------------------------------------------------------------------------
6+
# This file is adapted from:
27
# https://github.com/googleapis/java-pubsub-group-kafka-connector/blob/main/config/cps-sink-connector.properties
3-
4-
# Copyright 2024 Google LLC
8+
# The original copyright and license are reproduced below.
9+
#
10+
# Copyright 2022 Google LLC
511
#
612
# Licensed under the Apache License, Version 2.0 (the "License");
713
# you may not use this file except in compliance with the License.
@@ -14,6 +20,7 @@
1420
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
1521
# See the License for the specific language governing permissions and
1622
# limitations under the License.
23+
# --------------------------------------------------------------------------
1724

1825
# Unique name for the Pub/Sub sink connector.
1926
name=CPSSinkConnector

broker/consumer/lvk/psconnect-worker-authenticated.properties

Lines changed: 5 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,7 @@
1-
# Kafka connect worker configuration
2-
# See: https://docs.confluent.io/platform/current/connect/references/allconfigs.html
1+
# Kafka Connect worker configuration
2+
# This file is part of a workflow that creates an authenticated connection to the Kafka broker.
3+
# For config options, see https://docs.confluent.io/platform/current/connect/references/allconfigs.html#worker-configuration-properties
4+
# See also: https://kafka.apache.org/documentation/#adminclientconfigs
35

46
bootstrap.servers=kafka.gcn.nasa.gov:9092
57
plugin.path=/usr/local/share/kafka/plugins
@@ -27,7 +29,7 @@ sasl.jaas.config=org.apache.kafka.common.security.oauthbearer.OAuthBearerLoginMo
2729
clientId="CLIENT_ID" \
2830
clientSecret="CLIENT_SECRET";
2931

30-
# settings with `consumer` prefixes are passed through to the Kafka consumer
32+
# settings with `consumer.` prefixes are passed through to the Kafka consumer
3133
consumer.group.id=pitt-google-broker-test
3234
consumer.auto.offset.reset=earliest
3335
consumer.sasl.mechanism=OAUTHBEARER

broker/consumer/ps-connector.properties

Lines changed: 24 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,27 @@
1+
# Kafka Connect sink connector configs
2+
# For config options, see https://docs.confluent.io/platform/current/installation/configuration/connect/sink-connect-configs.html
3+
# For additional Pub/Sub-specific options, see https://github.com/googleapis/java-pubsub-group-kafka-connector?tab=readme-ov-file#sink-connector
4+
#
5+
# --------------------------------------------------------------------------
6+
# This file is adapted from:
7+
# https://github.com/googleapis/java-pubsub-group-kafka-connector/blob/main/config/cps-sink-connector.properties
8+
# The original copyright and license are reproduced below.
9+
#
10+
# Copyright 2022 Google LLC
11+
#
12+
# Licensed under the Apache License, Version 2.0 (the "License");
13+
# you may not use this file except in compliance with the License.
14+
# You may obtain a copy of the License at
15+
#
16+
# http://www.apache.org/licenses/LICENSE-2.0
17+
#
18+
# Unless required by applicable law or agreed to in writing, software
19+
# distributed under the License is distributed on an "AS IS" BASIS,
20+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
21+
# See the License for the specific language governing permissions and
22+
# limitations under the License.
23+
# --------------------------------------------------------------------------
24+
125
name=ps-sink-connector
226
connector.class=com.google.pubsub.kafka.sink.CloudPubSubSinkConnector
327
tasks.max=10

broker/consumer/psconnect-worker-authenticated.properties

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
1-
# Kafka connect worker configuration for the ZTF consumer
1+
# Kafka Connect worker configuration
22
# This file is part of a workflow that creates an authenticated connection to the Kafka broker.
3-
# reference: https://docs.confluent.io/platform/current/connect/references/allconfigs.html#worker-configuration-properties
4-
# see also: https://kafka.apache.org/documentation/#adminclientconfigs
3+
# For config options, see https://docs.confluent.io/platform/current/connect/references/allconfigs.html#worker-configuration-properties
4+
# See also: https://kafka.apache.org/documentation/#adminclientconfigs
55

66
bootstrap.servers=public2.alerts.ztf.uw.edu:9094
77
plugin.path=/usr/local/share/kafka/plugins

broker/consumer/psconnect-worker-unauthenticated.properties

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
1-
# Kafka connect worker configuration for the ZTF consumer
1+
# Kafka Connect worker configuration
22
# This file is part of a workflow that creates an unauthenticated connection to the Kafka broker.
3-
# reference: https://docs.confluent.io/platform/current/connect/references/allconfigs.html#worker-configuration-properties
4-
# see also: https://kafka.apache.org/documentation/#adminclientconfigs
3+
# For config options, see https://docs.confluent.io/platform/current/connect/references/allconfigs.html#worker-configuration-properties
4+
# See also: https://kafka.apache.org/documentation/#adminclientconfigs
55

66
bootstrap.servers=public.alerts.ztf.uw.edu:9092
77
plugin.path=/usr/local/share/kafka/plugins

broker/consumer/rubin/admin.properties

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,7 @@
1-
# see https://kafka.apache.org/documentation/#adminclientconfigs
1+
# Kafka Admin client configs
2+
# This file is part of a workflow that creates an authenticated connection to the Kafka broker.
3+
# In cases where we can connect without authentication (e.g., ZTF), this file is not used.
4+
# For config options, see https://kafka.apache.org/documentation/#adminclientconfigs
25

36
bootstrap.servers=usdf-alert-stream-dev.lsst.cloud:9094
47
sasl.mechanism=SCRAM-SHA-512

broker/consumer/rubin/ps-connector.properties

Lines changed: 24 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,30 @@
1-
# Adapted from:
2-
# https://github.com/GoogleCloudPlatform/pubsub/blob/master/kafka-connector/config/cps-sink-connector.properties
1+
# Kafka Connect sink connector configs
2+
# For config options, see https://docs.confluent.io/platform/current/installation/configuration/connect/sink-connect-configs.html
3+
# For additional Pub/Sub-specific options, see https://github.com/googleapis/java-pubsub-group-kafka-connector?tab=readme-ov-file#sink-connector
4+
#
5+
# --------------------------------------------------------------------------
6+
# This file is adapted from:
7+
# https://github.com/googleapis/java-pubsub-group-kafka-connector/blob/main/config/cps-sink-connector.properties
8+
# The original copyright and license are reproduced below.
9+
#
10+
# Copyright 2022 Google LLC
11+
#
12+
# Licensed under the Apache License, Version 2.0 (the "License");
13+
# you may not use this file except in compliance with the License.
14+
# You may obtain a copy of the License at
15+
#
16+
# http://www.apache.org/licenses/LICENSE-2.0
17+
#
18+
# Unless required by applicable law or agreed to in writing, software
19+
# distributed under the License is distributed on an "AS IS" BASIS,
20+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
21+
# See the License for the specific language governing permissions and
22+
# limitations under the License.
23+
# --------------------------------------------------------------------------
324

425
name=ps-sink-connector
526
connector.class=com.google.pubsub.kafka.sink.CloudPubSubSinkConnector
6-
# Tasks do the actual data copying and determine parallelism of the consumer.
27+
# Tasks do the actual data copying and determine parallelism of the consumer.
728
# When I (Troy Raen) used 10 max tasks (as we do with ZTF),
829
# the consumer VM (size e2-standard-2) ran out of memory.
930
# Using 1 task for now.

broker/consumer/rubin/psconnect-worker.properties

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,7 @@
1-
# Kafka connect worker configuration
2-
# See: https://docs.confluent.io/platform/current/connect/references/allconfigs.html
1+
# Kafka Connect worker configuration
2+
# This file is part of a workflow that creates an authenticated connection to the Kafka broker.
3+
# For config options, see https://docs.confluent.io/platform/current/connect/references/allconfigs.html#worker-configuration-properties
4+
# See also: https://kafka.apache.org/documentation/#adminclientconfigs
35

46
bootstrap.servers=usdf-alert-stream-dev.lsst.cloud:9094
57
plugin.path=/usr/local/share/kafka/plugins
@@ -32,8 +34,8 @@ sasl.jaas.config=org.apache.kafka.common.security.scram.ScramLoginModule require
3234
username="pittgoogle-idfint"\
3335
password="KAFKA_PASSWORD";
3436

35-
# settings with `consumer` prefixes are passed through to the Kafka consumer
36-
# consumer.group.id must begin with our username: pittgoogle-idfint
37+
# settings with `consumer.` prefixes are passed through to the Kafka consumer
38+
# for Rubin, consumer.group.id must begin with our username: pittgoogle-idfint
3739
consumer.group.id=pittgoogle-idfint-kafka-pubsub-connector
3840
consumer.auto.offset.reset=earliest
3941
consumer.sasl.mechanism=SCRAM-SHA-512

0 commit comments

Comments
 (0)