I hereby claim:
- I am retroryan on github.
- I am binaryknight (https://keybase.io/binaryknight) on keybase.
- I have a public key ASD4qP7q58L0A88jrHOloskDr5ixa3CT77MWpbiEtxO5Cwo
To claim this, I am signing this object:
| import zhttp.service.{ChannelFactory, Client, EventLoopGroup} | |
| import zio.{Tag, ZEnvironment, ZIO, ZLayer} | |
| trait NumService: | |
| val get: ZIO[Any, Throwable, Int] | |
| final case class NumServiceLive(c: ChannelFactory, e: EventLoopGroup) extends NumService: | |
| val url = "https://random-num-x5ht4amjia-uc.a.run.app/" | |
| val get: ZIO[Any, Throwable, Int] = | |
| val request = |
| #create class | |
| CreateClass({ name: "ledger" }) | |
| #create index | |
| CreateIndex( | |
| { | |
| name: "UNIQUE_ENTRY_CONSTRAINT", | |
| source: Class("ledger"), | |
| terms: [{ field: ["data", "clientId"] }], | |
| values: [{ field: ["data", "counter"] }], |
| #Create a Class | |
| CreateClass({ name: "orders" }) | |
| #Create an Index | |
| CreateIndex( | |
| { | |
| name: "orders_index", | |
| source: Class("orders"), | |
| values: [{ field: ["data", "orderId"], reverse:true }, { field: ["ref"] }], | |
| serialized: true |
| 15kX9REWJS9v3NmT1JeeUN6NnovQk9WnWz |
I hereby claim:
To claim this, I am signing this object:
| override val supervisorStrategy: SupervisorStrategy = { | |
| val decider: SupervisorStrategy.Decider = { | |
| case Guest.CaffeineException => | |
| SupervisorStrategy.Stop | |
| case Waiter.FrustratedException(coffee, guest) => | |
| barista.tell(Barista.PrepareCoffee(coffee, guest), sender()) | |
| SupervisorStrategy.Restart | |
| } | |
| OneForOneStrategy()(decider orElse super.supervisorStrategy.decider) | |
| } |
| import org.apache.spark.streaming.kafka.KafkaUtils | |
| import org.apache.spark.streaming.Time | |
| import kafka.serializer.StringDecoder | |
| import org.joda.time.DateTime | |
| import org.apache.spark.sql.SaveMode | |
| import sqlContext.implicits._ | |
| val ratingsStream = KafkaUtils.createDirectStream[String, String, StringDecoder, StringDecoder](ssc, kafkaParams, topics) | |
| val msgs = ratingsStream.transform { |
| import scala.annotation.tailrec | |
| def atoi(chList:List[String]):Int = { | |
| @tailrec | |
| def atoiAccumulator(chList: List[String], accumulator: Int): Int = chList match { | |
| case Nil => accumulator | |
| case head :: tail => | |
| val tensMult = scala.math.pow(10, tail.length).toInt | |
| val nxtAccumulator = (head.toInt * tensMult) + accumulator |
| class DseSparkContextFactory extends SparkContextFactory { | |
| import SparkJobUtils._ | |
| type C = SparkContext with ContextLike | |
| def makeContext(config: Config, contextConfig: Config, contextName: String): C = { | |
| val conf = configToSparkConf(config, contextConfig, contextName) | |
| val sparkMaster = "dsetool sparkmaster".!!.trim | |
| conf.set("spark.master",sparkMaster) |
| ################ TEST RUN 1 Java 8 Async Receiver ########################### | |
| java -jar target/cqltestclient.jar --host=127.0.0.1 --activity=WriteTelemetryAsync:7000000:150:1500 --activity=ReadTelemetryAsync:700000:50:500 | |
| 5/4/15 9:16:55 PM ============================================================== | |
| -- Counters -------------------------------------------------------------------- | |
| ActivityExecutorService.activities | |
| count = 2 | |
| ReadTelemetryAsyncActivity.async-pending |