Closed Bug 1281770 Opened 8 years ago Closed 8 years ago

Longitudinal dataset tutorial notebook errors out

Categories

(Cloud Services Graveyard :: Metrics: Pipeline, defect, P1)

defect

Tracking

(Not tracked)

RESOLVED FIXED

People

(Reporter: rvitillo, Assigned: rvitillo)

Details

Attachments

(1 file)

54 bytes, text/x-github-pull-request
mdoglio
: review+
Details | Review
The example notebook started failing two days ago. 

An error occurred while calling z:org.apache.spark.api.python.PythonRDD.runJob.: org.apache.spark.SparkException: Job aborted due to stage failure: Task 2 in stage 4.0 failed 4 times, most recent failure: Lost task 2.3 in stage 4.0 (TID 2494, ip-172-31-1-31.us-west-2.compute.internal): org.apache.spark.SparkException: Kryo serialization failed: Buffer overflow. Available: 0, required: 19593510. To avoid this, increase spark.kryoserializer.buffer.max value.
	at org.apache.spark.serializer.KryoSerializerInstance.serialize(KryoSerializer.scala:299)
	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:239)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:745)
Attached file PR
Attachment #8764593 - Flags: review?(mdoglio)
Assignee: nobody → rvitillo
Priority: -- → P1
Attachment #8764593 - Flags: review?(mdoglio) → review+
Status: NEW → RESOLVED
Closed: 8 years ago
Resolution: --- → FIXED
Product: Cloud Services → Cloud Services Graveyard
You need to log in before you can comment on or make changes to this bug.

Attachment

General

Created:
Updated:
Size: