Closed akhld closed 9 years ago
Any idea on how to get these numbers up?
Hi, Sorry for replying late on this. I think this is an issue with Spark. If you see how I generate the Blocks and writes to BlockManager, I used following call..
_receiver.store(_dataBuffer.iterator());
You can see this in PartitionManager.java line # 216
I think if you write the Block this way, Spark not able to calculate the number of records properly.
You can probably raise a JIRA for Spark.
Dibyendu
Thanks for the reply Dibyendu. Will try to raise a JIRA with Spark.
Hi @akhld . I have fixed the UI Reporting issue in latest consumer. Please take a look at it and let me know how it looks ..
Dibyendu
Ohh nice. Will let you know. :) On 3 Jun 2015 17:33, "Dibyendu Bhattacharya" notifications@github.com wrote:
Hi @akhld https://github.com/akhld . I have fixed the UI Reporting issue in latest consumer. Please take a look at it and let me know how it looks ..
Dibyendu
— Reply to this email directly or view it on GitHub https://github.com/dibbhatt/kafka-spark-consumer/issues/6#issuecomment-108343350 .
I cloned the latest version and ran it once, but it doesn't show any metrices in the streaming tab. Previously it was -ve values now its 0.
Are you pushing messages to Kafka ? Initial few batches comes as Zero. Refresh the page and check if you see the stats
Yes, messages are pushed and its consuming and printing the count also.
I ran the job for a minute, it processed around 50 batches (batch interval being 1 second). And you can see from the image the "Processed Batches" is also broken as its always 1. I'm using spark 1.3.1 and recompiled lowlevel kafka with 1.3.1 version of spark.
You have 1.2 setup ? My team has successfully tested in 1.2. Also I am presently doing some work on Spark master ( 1.4) and that works there also . Not tested with 1.3.1 though.
I have it just tested on 1.2.2, still the same. I'm trying in local mode.
Closing this as latest changes solved this problem
Hi @akhld
I have created a JIRA to track the progress of contributing back this project to Apache Spark.
https://issues.apache.org/jira/browse/SPARK-11045
This project is now presently in spark-packages and I believe this is the correct time to contribute it to Apache Spark Project and give better options to larger community around Kafka Connectivity for Spark Streaming.
kindly Vote for this JIRA.
Great, voted already. Hope people will start contributing to it.
On Sat, Oct 10, 2015 at 11:59 AM, Dibyendu Bhattacharya < notifications@github.com> wrote:
Hi @akhld https://github.com/akhld
I have created a JIRA to track the progress of contributing back this project to Apache Spark.
https://issues.apache.org/jira/browse/SPARK-11045
This project is now presently in spark-packages and I believe this is the correct time to contribute it to Apache Spark Project and give better options to larger community around Kafka Connectivity for Spark Streaming.
kindly Vote for this JIRA.
— Reply to this email directly or view it on GitHub https://github.com/dibbhatt/kafka-spark-consumer/issues/6#issuecomment-147045584 .
Cheers!
Thanks a lot Akhil. Much appreciated.
I have a jobListener extending StreamingListener interface to perform some tasks onBatchCompleted, it works pretty well with native SparkStreaming and KafkaUtil. But its giving wrong values when i used with this lowlevel consumer.
Here's the Listener class:
You can attach it to your ssc as:
Let me know if there's some other way to pull batch info