diff --git a/examples/src/main/scala/org/apache/spark/streaming/examples/ActorWordCount.scala b/examples/src/main/scala/org/apache/spark/streaming/examples/ActorWordCount.scala index a5888811cc5eaa1d8e834ae55a57b87cd1cc8ed9..bc0d1633f1c1ef1095a61dacf2d5d30e6754fbc3 100644 --- a/examples/src/main/scala/org/apache/spark/streaming/examples/ActorWordCount.scala +++ b/examples/src/main/scala/org/apache/spark/streaming/examples/ActorWordCount.scala @@ -88,7 +88,7 @@ extends Actor with Receiver { override def preStart = remotePublisher ! SubscribeReceiver(context.self) def receive = { - case msg ⇒ pushBlock(msg.asInstanceOf[T]) + case msg => pushBlock(msg.asInstanceOf[T]) } override def postStop() = remotePublisher ! UnsubscribeReceiver(context.self) diff --git a/external/twitter/src/main/scala/org/apache/spark/streaming/twitter/TwitterInputDStream.scala b/external/twitter/src/main/scala/org/apache/spark/streaming/twitter/TwitterInputDStream.scala index 5cc721d7f90ea9c7478a162c0d65f48c04363179..3316b6dc39d6baaf79877dbe20cf809f29f56233 100644 --- a/external/twitter/src/main/scala/org/apache/spark/streaming/twitter/TwitterInputDStream.scala +++ b/external/twitter/src/main/scala/org/apache/spark/streaming/twitter/TwitterInputDStream.scala @@ -17,14 +17,11 @@ package org.apache.spark.streaming.twitter -import java.util.prefs.Preferences import twitter4j._ import twitter4j.auth.Authorization import twitter4j.conf.ConfigurationBuilder -import twitter4j.conf.PropertyConfiguration import twitter4j.auth.OAuthAuthorization -import twitter4j.auth.AccessToken -import org.apache.spark._ + import org.apache.spark.streaming._ import org.apache.spark.streaming.dstream._ import org.apache.spark.storage.StorageLevel diff --git a/external/zeromq/src/main/scala/org/apache/spark/streaming/zeromq/ZeroMQReceiver.scala b/external/zeromq/src/main/scala/org/apache/spark/streaming/zeromq/ZeroMQReceiver.scala index 769761e3b860c54d7f94eb83c82f6e8f238216bf..960c6a389ef0f1a3902fed497dd3c45096883b0f 100644 --- a/external/zeromq/src/main/scala/org/apache/spark/streaming/zeromq/ZeroMQReceiver.scala +++ b/external/zeromq/src/main/scala/org/apache/spark/streaming/zeromq/ZeroMQReceiver.scala @@ -31,7 +31,7 @@ import org.apache.spark.streaming.receivers._ */ private[streaming] class ZeroMQReceiver[T: ClassTag](publisherUrl: String, subscribe: Subscribe, - bytesToObjects: Seq[ByteString] ⇒ Iterator[T]) + bytesToObjects: Seq[ByteString] => Iterator[T]) extends Actor with Receiver with Logging { override def preStart() = ZeroMQExtension(context.system).newSocket(SocketType.Sub, Listener(self), @@ -39,16 +39,16 @@ private[streaming] class ZeroMQReceiver[T: ClassTag](publisherUrl: String, def receive: Receive = { - case Connecting ⇒ logInfo("connecting ...") + case Connecting => logInfo("connecting ...") - case m: ZMQMessage ⇒ + case m: ZMQMessage => logDebug("Received message for:" + m.frame(0)) //We ignore first frame for processing as it is the topic val bytes = m.frames.tail pushBlock(bytesToObjects(bytes)) - case Closed ⇒ logInfo("received closed ") + case Closed => logInfo("received closed ") } } diff --git a/external/zeromq/src/main/scala/org/apache/spark/streaming/zeromq/ZeroMQUtils.scala b/external/zeromq/src/main/scala/org/apache/spark/streaming/zeromq/ZeroMQUtils.scala index 7a14b3d2bf27859c346f4490cac61acd83e68a69..b47d786986a917e55bfbde287a7fe0a6554f1d1f 100644 --- a/external/zeromq/src/main/scala/org/apache/spark/streaming/zeromq/ZeroMQUtils.scala +++ b/external/zeromq/src/main/scala/org/apache/spark/streaming/zeromq/ZeroMQUtils.scala @@ -46,7 +46,7 @@ object ZeroMQUtils { ssc: StreamingContext, publisherUrl: String, subscribe: Subscribe, - bytesToObjects: Seq[ByteString] ⇒ Iterator[T], + bytesToObjects: Seq[ByteString] => Iterator[T], storageLevel: StorageLevel = StorageLevel.MEMORY_AND_DISK_SER_2, supervisorStrategy: SupervisorStrategy = ReceiverSupervisorStrategy.defaultStrategy ): DStream[T] = { diff --git a/streaming/src/main/scala/org/apache/spark/streaming/receivers/ActorReceiver.scala b/streaming/src/main/scala/org/apache/spark/streaming/receivers/ActorReceiver.scala index 79ed696814f07388e8911e6c1655f949d8a5c0d4..9c5b177c163c8399c373767effcfef27be1525d1 100644 --- a/streaming/src/main/scala/org/apache/spark/streaming/receivers/ActorReceiver.scala +++ b/streaming/src/main/scala/org/apache/spark/streaming/receivers/ActorReceiver.scala @@ -37,8 +37,8 @@ object ReceiverSupervisorStrategy { val defaultStrategy = OneForOneStrategy(maxNrOfRetries = 10, withinTimeRange = 15 millis) { - case _: RuntimeException ⇒ Restart - case _: Exception ⇒ Escalate + case _: RuntimeException => Restart + case _: Exception => Escalate } } @@ -66,7 +66,7 @@ object ReceiverSupervisorStrategy { */ trait Receiver { - self: Actor ⇒ // to ensure that this can be added to Actor classes only + self: Actor => // to ensure that this can be added to Actor classes only /** * Push an iterator received data into Spark Streaming for processing @@ -139,25 +139,25 @@ private[streaming] class ActorReceiver[T: ClassTag]( def receive = { - case Data(iter: Iterator[_]) ⇒ pushBlock(iter.asInstanceOf[Iterator[T]]) + case Data(iter: Iterator[_]) => pushBlock(iter.asInstanceOf[Iterator[T]]) - case Data(msg) ⇒ + case Data(msg) => blocksGenerator += msg.asInstanceOf[T] n.incrementAndGet - case props: Props ⇒ + case props: Props => val worker = context.actorOf(props) logInfo("Started receiver worker at:" + worker.path) sender ! worker - case (props: Props, name: String) ⇒ + case (props: Props, name: String) => val worker = context.actorOf(props, name) logInfo("Started receiver worker at:" + worker.path) sender ! worker case _: PossiblyHarmful => hiccups.incrementAndGet() - case _: Statistics ⇒ + case _: Statistics => val workers = context.children sender ! Statistics(n.get, workers.size, hiccups.get, workers.mkString("\n"))