I hereby claim:
- I am ninadingole on github.
- I am iamninad (https://keybase.io/iamninad) on keybase.
- I have a public key ASDQ_lKHv4aZIwHqiD5wWJOCHV4euhaPoJGdYPa3Anq_hwo
To claim this, I am signing this object:
name: "My CodeQL config" | |
disable-default-queries: true | |
languages: java | |
queries: | |
- name: Extended Security | |
uses: security-extended | |
- name: Security and Quality |
name: "CodeQL" | |
on: | |
push: | |
branches: [master, ] | |
pull_request: | |
# The branches below must be a subset of the branches above | |
branches: [master] | |
# schedule: | |
# - cron: '0 3 * * 4' |
I hereby claim:
To claim this, I am signing this object:
{"lastUpload":"2020-05-06T12:15:46.434Z","extensionVersion":"v3.4.3"} |
package internal | |
import ( | |
"context" | |
"encoding/json" | |
"errors" | |
"fmt" | |
"log" | |
"net/http" | |
"os" |
######################################### | |
#### Inside project/E2E.class object #### | |
######################################### | |
import sbt.{Configuration, Defaults, Test, inConfig} | |
import sbt._ | |
import sbt.Keys._ | |
object E2E { | |
final val E2ETest = Configuration.of("EndToEndTest", "e2e") extend (Test) |
REPLICAT rkafka | |
TARGETDB LIBFILE libggjava.so SET property=dirprm/rkafka_handler.props | |
REPORTCOUNT EVERY 1 MINUTES, RATE | |
GROUPTRANSOPS 1000 | |
MAP *.*.*, TARGET *.*.*; |
gg.handlerlist=kafkaconnect | |
#The handler properties | |
gg.handler.kafkaconnect.type=kafkaconnect | |
gg.handler.kafkaconnect.kafkaProducerConfigFile=kafkaconnect.properties | |
# See http://docs.oracle.com/goldengate/bd123110/gg-bd/GADBD/using-kafka-connect-handler.htm#GUID-A87CAFFA-DACF-43A0-8C6C-5C64B578D606 | |
gg.handler.kafkaconnect.topicMappingTemplate=ora-ogg-${schemaName}-${tableName}-avro | |
gg.handler.kafkaconnect.keyMappingTemplate=${primaryKeys} |
bootstrap.servers=localhost:9092 | |
key.converter=io.confluent.connect.avro.AvroConverter | |
value.converter=io.confluent.connect.avro.AvroConverter | |
key.converter.schema.registry.url=http://localhost:8081 | |
value.converter.schema.registry.url=http://localhost:8081 | |
# For Json | |
#value.converter=org.apache.kafka.connect.json.JsonConverter | |
#key.converter=org.apache.kafka.connect.json.JsonConverter |
def buildEventStream = { | |
import AppSerdes.movieBEventSerde.eventConsumed | |
builder.stream[Int, BusinessEvent]("events") | |
} | |
private val eventStreams: KStreamS[Int, BusinessEvent] = buildEventStream | |
def filterEventsByType(eventType: String): KStreamS[Int, BusinessEvent] = { | |
eventStreams.filter((_: Int, event: BusinessEvent) => event.eventType.equalsIgnoreCase(eventType)) | |
} |