Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

PicklingError: Could not serialize object: IndexError: tuple index out of range #5084

Open
Srirangan02 opened this issue Feb 2, 2025 · 4 comments

Comments

@Srirangan02
Copy link

Srirangan02 commented Feb 2, 2025

When i try to run a Pyspark code getting the error while creating data frame.this same code runs in intellij and also i used colab for three months without any issue with the same versions of spark and python.

This should run without any issues

I'm using Chrome

https://colab.research.google.com/drive/1-s_pjQqKOWxWzzFjkypePfZ76jQU2f7M#scrollTo=cZLFdCJahMUl&line=21&uniqifier=1

Image

@Srirangan02 Srirangan02 added the bug label Feb 2, 2025
@cperry-goog
Copy link

can you grant open access to the notebook?

@Srirangan02
Copy link
Author

can you grant open access to the notebook?

Could you please specify what you need to check? I’ve already attached the screenshot and issue details.

@epsi10nvn
Copy link

try to use spark-3.5.0

!sudo apt update
!apt-get install openjdk-8-jdk-headless -qq > /dev/null
!wget -q http://archive.apache.org/dist/spark/spark-3.5.0/spark-3.5.0-bin-hadoop3.tgz
!tar xf spark-3.5.0-bin-hadoop3.tgz
!pip install -q findspark
`
````
import os
import findspark

os.environ["JAVA_HOME"] = "/usr/lib/jvm/java-8-openjdk-amd64"
os.environ["SPARK_HOME"] = "/content/spark-3.5.0-bin-hadoop3"

findspark.init()
`

@Srirangan02
Copy link
Author

try to use spark-3.5.0

!sudo apt update
!apt-get install openjdk-8-jdk-headless -qq > /dev/null
!wget -q http://archive.apache.org/dist/spark/spark-3.5.0/spark-3.5.0-bin-hadoop3.tgz
!tar xf spark-3.5.0-bin-hadoop3.tgz
!pip install -q findspark
`

import os
import findspark

os.environ["JAVA_HOME"] = "/usr/lib/jvm/java-8-openjdk-amd64"
os.environ["SPARK_HOME"] = "/content/spark-3.5.0-bin-hadoop3"

findspark.init()
`

Thank you so much its working why its not working in spark 3.1?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

3 participants