Skip to content

tasks not able to parse for spark v 2.1.0 #1

Open
@rohitbhintade

Description

@rohitbhintade

Sample event of spark 2.1.0 :

{"Event":"SparkListenerTaskEnd","Stage ID":0,"Stage Attempt ID":0,"Task Type":"ShuffleMapTask","Task End Reason":{"Reason":"Success"},"Task Info":{"Task ID":0,"Index":0,"Attempt":0,"Launch Time":1511259408672,"Executor ID":"1","Host":"sandbox.hortonworks.com","Locality":"PROCESS_LOCAL","Speculative":false,"Getting Result Time":0,"Finish Time":1511259410923,"Failed":false,"Killed":false,"Accumulables":[{"ID":13,"Name":"internal.metrics.executorDeserializeTime","Update":1400,"Value":1400,"Internal":true,"Count Failed Values":true},{"ID":14,"Name":"internal.metrics.executorDeserializeCpuTime","Update":992315389,"Value":992315389,"Internal":true,"Count Failed Values":true},{"ID":15,"Name":"internal.metrics.executorRunTime","Update":677,"Value":677,"Internal":true,"Count Failed Values":true},{"ID":16,"Name":"internal.metrics.executorCpuTime","Update":606198111,"Value":606198111,"Internal":true,"Count Failed Values":true},{"ID":17,"Name":"internal.metrics.resultSize","Update":2705,"Value":2705,"Internal":true,"Count Failed Values":true},{"ID":18,"Name":"internal.metrics.jvmGCTime","Update":65,"Value":65,"Internal":true,"Count Failed Values":true},{"ID":19,"Name":"internal.metrics.resultSerializationTime","Update":1,"Value":1,"Internal":true,"Count Failed Values":true},{"ID":23,"Name":"internal.metrics.updatedBlockStatuses","Update":[{"Block ID":"broadcast_0_piece0","Status":{"Storage Level":{"Use Disk":false,"Use Memory":true,"Deserialized":false,"Replication":1},"Memory Size":4659,"Disk Size":0}},{"Block ID":"broadcast_0","Status":{"Storage Level":{"Use Disk":false,"Use Memory":true,"Deserialized":true,"Replication":1},"Memory Size":8952,"Disk Size":0}}],"Value":[{"Block ID":"broadcast_0_piece0","Status":{"Storage Level":{"Use Disk":false,"Use Memory":true,"Deserialized":false,"Replication":1},"Memory Size":4659,"Disk Size":0}},{"Block ID":"broadcast_0","Status":{"Storage Level":{"Use Disk":false,"Use Memory":true,"Deserialized":true,"Replication":1},"Memory Size":8952,"Disk Size":0}}],"Internal":true,"Count Failed Values":true},{"ID":30,"Name":"internal.metrics.shuffle.write.bytesWritten","Update":59,"Value":59,"Internal":true,"Count Failed Values":true},{"ID":31,"Name":"internal.metrics.shuffle.write.recordsWritten","Update":1,"Value":1,"Internal":true,"Count Failed Values":true},{"ID":32,"Name":"internal.metrics.shuffle.write.writeTime","Update":7044541,"Value":7044541,"Internal":true,"Count Failed Values":true},{"ID":12,"Name":"number of output rows","Update":"1","Value":"1","Internal":true,"Count Failed Values":true,"Metadata":"sql"},{"ID":7,"Name":"duration total (min, med, max)","Update":"37","Value":"36","Internal":true,"Count Failed Values":true,"Metadata":"sql"},{"ID":8,"Name":"number of output rows","Update":"1","Value":"1","Internal":true,"Count Failed Values":true,"Metadata":"sql"},{"ID":1,"Name":"data size total (min, med, max)","Update":"15","Value":"14","Internal":true,"Count Failed Values":true,"Metadata":"sql"}]},"Task Metrics":{"Executor Deserialize Time":1400,"Executor Deserialize CPU Time":992315389,"Executor Run Time":677,"Executor CPU Time":606198111,"Result Size":2705,"JVM GC Time":65,"Result Serialization Time":1,"Memory Bytes Spilled":0,"Disk Bytes Spilled":0,"Shuffle Read Metrics":{"Remote Blocks Fetched":0,"Local Blocks Fetched":0,"Fetch Wait Time":0,"Remote Bytes Read":0,"Local Bytes Read":0,"Total Records Read":0},"Shuffle Write Metrics":{"Shuffle Bytes Written":59,"Shuffle Write Time":7044541,"Shuffle Records Written":1},"Input Metrics":{"Bytes Read":0,"Records Read":0},"Output Metrics":{"Bytes Written":0,"Records Written":0},"Updated Blocks":[{"Block ID":"broadcast_0_piece0","Status":{"Storage Level":{"Use Disk":false,"Use Memory":true,"Deserialized":false,"Replication":1},"Memory Size":4659,"Disk Size":0}},{"Block ID":"broadcast_0","Status":{"Storage Level":{"Use Disk":false,"Use Memory":true,"Deserialized":true,"Replication":1},"Memory Size":8952,"Disk Size":0}}]}}

The event has varying data type for update field thats why getting error:

Caused by: org.json4s.package$MappingException: Expected collection but got JInt(16959) for root JInt(16959) and mapping Seq[SparkAccumulableUpdate]
No usable value for Task Info
No usable value for Accumulables
No usable value for Update
Expected collection but got JInt(16959) for root JInt(16959) and mapping Seq[SparkAccumulableUpdate]

Can you please check it against spark 2.1.0, any help would be appreciated.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions