Any hints for debuging errors like "java.io.exception: read 95 bytes, should read 159" ?

classic Classic list List threaded Threaded
3 messages Options
Reply | Threaded
Open this post in threaded view
|

Any hints for debuging errors like "java.io.exception: read 95 bytes, should read 159" ?

wuqi-2
Hi,
 
I am trying to modify the Fetcher code in Nutch.81 , but always get the exceptions below in the hadoop.log.

java.lang.RuntimeException: java.io.IOException: Version: 4
Status: 5 (fetch_success)
Fetch time: Wed Mar 14 22:15:58 CST 2007
Modified time: Thu Jan 01 08:00:00 CST 1970
Retries since fetch: 0
Fetch interval: 0.5 days
Score: 1.0
Signature: f4b0490606f416542246fb2bf625d0cb
Metadata: null
 read 95 bytes, should read 159
 at org.apache.hadoop.mapred.ReduceTask$ValuesIterator.next(ReduceTask.java:129)
 at org.apache.nutch.crawl.CrawlDbReducer.reduce(CrawlDbReducer.java:56)
 at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:247)
 at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:112)

Why "159" bytes should be read here? Any hints for me to find the answer?

Thanks
-Qi
 
Reply | Threaded
Open this post in threaded view
|

Re: Any hints for debuging errors like "java.io.exception: read 95 bytes, should read 159" ?

Dennis Kubes
Did you change the Parse metadata writable object.  This error looks
like ones that show up when a change to a writeable source code has been
made to write out something (i.e. like adding meta tags to the parsedata
metadata) but the code to read it in has not been made in the writeable
source.

The errors is basically stating that you wrote something out but haven't
read it back in.

Dennis Kubes

qi wu wrote:

> Hi,
>  
> I am trying to modify the Fetcher code in Nutch.81 , but always get the exceptions below in the hadoop.log.
>
> java.lang.RuntimeException: java.io.IOException: Version: 4
> Status: 5 (fetch_success)
> Fetch time: Wed Mar 14 22:15:58 CST 2007
> Modified time: Thu Jan 01 08:00:00 CST 1970
> Retries since fetch: 0
> Fetch interval: 0.5 days
> Score: 1.0
> Signature: f4b0490606f416542246fb2bf625d0cb
> Metadata: null
>  read 95 bytes, should read 159
>  at org.apache.hadoop.mapred.ReduceTask$ValuesIterator.next(ReduceTask.java:129)
>  at org.apache.nutch.crawl.CrawlDbReducer.reduce(CrawlDbReducer.java:56)
>  at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:247)
>  at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:112)
>
> Why "159" bytes should be read here? Any hints for me to find the answer?
>
> Thanks
> -Qi
>  
Reply | Threaded
Open this post in threaded view
|

Re: Any hints for debuging errors like "java.io.exception: read 95 bytes, should read 159" ?

wuqi-2
Thank you Dennis!
Yes, the modified Fetcher.java of mine added new  values to metadata,but were not retrieved out during the subsequent process!

----- Original Message -----
From: "Dennis Kubes" <[hidden email]>
To: <[hidden email]>
Sent: Wednesday, March 14, 2007 11:40 PM
Subject: Re: Any hints for debuging errors like "java.io.exception: read 95 bytes, should read 159" ?


> Did you change the Parse metadata writable object.  This error looks
> like ones that show up when a change to a writeable source code has been
> made to write out something (i.e. like adding meta tags to the parsedata
> metadata) but the code to read it in has not been made in the writeable
> source.
>
> The errors is basically stating that you wrote something out but haven't
> read it back in.
>
> Dennis Kubes
>
> qi wu wrote:
>> Hi,
>>  
>> I am trying to modify the Fetcher code in Nutch.81 , but always get the exceptions below in the hadoop.log.
>>
>> java.lang.RuntimeException: java.io.IOException: Version: 4
>> Status: 5 (fetch_success)
>> Fetch time: Wed Mar 14 22:15:58 CST 2007
>> Modified time: Thu Jan 01 08:00:00 CST 1970
>> Retries since fetch: 0
>> Fetch interval: 0.5 days
>> Score: 1.0
>> Signature: f4b0490606f416542246fb2bf625d0cb
>> Metadata: null
>>  read 95 bytes, should read 159
>>  at org.apache.hadoop.mapred.ReduceTask$ValuesIterator.next(ReduceTask.java:129)
>>  at org.apache.nutch.crawl.CrawlDbReducer.reduce(CrawlDbReducer.java:56)
>>  at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:247)
>>  at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:112)
>>
>> Why "159" bytes should be read here? Any hints for me to find the answer?
>>
>> Thanks
>> -Qi
>>  
>