-
Notifications
You must be signed in to change notification settings - Fork 15
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
PicklingError: Could not serialize object: TypeError: can't pickle _abc_data objects #5
Comments
@ijoseph @kevinwang @variablenix @prasad-kamat please help |
Wow, we really should have pinned (and |
Alright, @saichaitanyamolabanti can you please try to pull this #7 PR, then |
Hey @ijoseph, I've noticed to install few libraries as per your comments and began installing them, mainly the install and import of cloudpickle. Here are my observations, I can still find some errors, please help !! Scenario-1 then row = dataset.filter(dataset.xxxx == '5').rdd.first() is working fine Scenario-2: |
then tried to pull those import of cloudpicklet and spark.serializers down below the investigation row like: but, still able to see error like - cloudpickle doesn't have the method 'print_exec' |
@ijoseph Or you can consider this scenario: |
@ijoseph @kevinwang @variablenix @prasad-kamat any help ? |
Isn't it this issue? It looks like it's solved in pyspark 3.0.0 (PR). So maybe it would be enough to set the lower bound for pyspark dependency in REQUIRED_PACKAGES = [
…,
'pyspark>=3.0.0',
] |
I wanted to try out this package, because this implements pyspark version of shapley value generations.
So, I just copy pasted "simple.ipynb" file into my environment to just observe everything basic is working alright or not, but able to see code is breaking at input cell [32]. Attached are the screenshots, could anyone please look into them?
The text was updated successfully, but these errors were encountered: