Spark Release 2.4.0 can not work on win7 because of ImportError: No module named 'resource'












1














I attempt to install Spark Release 2.4.0 on my pc, which system is win7_x64.



However when I try to run simple code to check whether spark is ready to work:
code:



import os
from pyspark import SparkConf, SparkContext
conf = SparkConf().setMaster('local[*]').setAppName('word_count')
sc = SparkContext(conf=conf)
d = ['a b c d', 'b c d e', 'c d e f']
d_rdd = sc.parallelize(d)
rdd_res = d_rdd.flatMap(lambda x: x.split(' ')).map(lambda word: (word, 1)).reduceByKey(lambda a, b: a+b)
print(rdd_res)
print(rdd_res.collect())


I get this error:



error1



I open the worker.py file to check the code.



I find that, in version 2.4.0, the code is :



worker.py v2.4.0



However, in version 2.3.2, the code is:



worker.py v2.3.2



Then I reinstall spark-2.3.2-bin-hadoop2.7 , the code works fine.



Also, I find this question:
ImportError: No module named 'resource'



So, I think spark-2.4.0-bin-hadoop2.7 can not work in win7 because of importing
resource module in worker.py, which is a Unix specific package.



I hope someone could fix this problem in spark.










share|improve this question
























  • Welcome to Stack Overflow Jason. Please don't post exceptions as images. Could edit your question and include these as text? Thanks.
    – user6910411
    Nov 19 '18 at 16:19
















1














I attempt to install Spark Release 2.4.0 on my pc, which system is win7_x64.



However when I try to run simple code to check whether spark is ready to work:
code:



import os
from pyspark import SparkConf, SparkContext
conf = SparkConf().setMaster('local[*]').setAppName('word_count')
sc = SparkContext(conf=conf)
d = ['a b c d', 'b c d e', 'c d e f']
d_rdd = sc.parallelize(d)
rdd_res = d_rdd.flatMap(lambda x: x.split(' ')).map(lambda word: (word, 1)).reduceByKey(lambda a, b: a+b)
print(rdd_res)
print(rdd_res.collect())


I get this error:



error1



I open the worker.py file to check the code.



I find that, in version 2.4.0, the code is :



worker.py v2.4.0



However, in version 2.3.2, the code is:



worker.py v2.3.2



Then I reinstall spark-2.3.2-bin-hadoop2.7 , the code works fine.



Also, I find this question:
ImportError: No module named 'resource'



So, I think spark-2.4.0-bin-hadoop2.7 can not work in win7 because of importing
resource module in worker.py, which is a Unix specific package.



I hope someone could fix this problem in spark.










share|improve this question
























  • Welcome to Stack Overflow Jason. Please don't post exceptions as images. Could edit your question and include these as text? Thanks.
    – user6910411
    Nov 19 '18 at 16:19














1












1








1


1





I attempt to install Spark Release 2.4.0 on my pc, which system is win7_x64.



However when I try to run simple code to check whether spark is ready to work:
code:



import os
from pyspark import SparkConf, SparkContext
conf = SparkConf().setMaster('local[*]').setAppName('word_count')
sc = SparkContext(conf=conf)
d = ['a b c d', 'b c d e', 'c d e f']
d_rdd = sc.parallelize(d)
rdd_res = d_rdd.flatMap(lambda x: x.split(' ')).map(lambda word: (word, 1)).reduceByKey(lambda a, b: a+b)
print(rdd_res)
print(rdd_res.collect())


I get this error:



error1



I open the worker.py file to check the code.



I find that, in version 2.4.0, the code is :



worker.py v2.4.0



However, in version 2.3.2, the code is:



worker.py v2.3.2



Then I reinstall spark-2.3.2-bin-hadoop2.7 , the code works fine.



Also, I find this question:
ImportError: No module named 'resource'



So, I think spark-2.4.0-bin-hadoop2.7 can not work in win7 because of importing
resource module in worker.py, which is a Unix specific package.



I hope someone could fix this problem in spark.










share|improve this question















I attempt to install Spark Release 2.4.0 on my pc, which system is win7_x64.



However when I try to run simple code to check whether spark is ready to work:
code:



import os
from pyspark import SparkConf, SparkContext
conf = SparkConf().setMaster('local[*]').setAppName('word_count')
sc = SparkContext(conf=conf)
d = ['a b c d', 'b c d e', 'c d e f']
d_rdd = sc.parallelize(d)
rdd_res = d_rdd.flatMap(lambda x: x.split(' ')).map(lambda word: (word, 1)).reduceByKey(lambda a, b: a+b)
print(rdd_res)
print(rdd_res.collect())


I get this error:



error1



I open the worker.py file to check the code.



I find that, in version 2.4.0, the code is :



worker.py v2.4.0



However, in version 2.3.2, the code is:



worker.py v2.3.2



Then I reinstall spark-2.3.2-bin-hadoop2.7 , the code works fine.



Also, I find this question:
ImportError: No module named 'resource'



So, I think spark-2.4.0-bin-hadoop2.7 can not work in win7 because of importing
resource module in worker.py, which is a Unix specific package.



I hope someone could fix this problem in spark.







python apache-spark pyspark






share|improve this question















share|improve this question













share|improve this question




share|improve this question








edited Nov 19 '18 at 13:02

























asked Nov 19 '18 at 12:55









Jason

63




63












  • Welcome to Stack Overflow Jason. Please don't post exceptions as images. Could edit your question and include these as text? Thanks.
    – user6910411
    Nov 19 '18 at 16:19


















  • Welcome to Stack Overflow Jason. Please don't post exceptions as images. Could edit your question and include these as text? Thanks.
    – user6910411
    Nov 19 '18 at 16:19
















Welcome to Stack Overflow Jason. Please don't post exceptions as images. Could edit your question and include these as text? Thanks.
– user6910411
Nov 19 '18 at 16:19




Welcome to Stack Overflow Jason. Please don't post exceptions as images. Could edit your question and include these as text? Thanks.
– user6910411
Nov 19 '18 at 16:19

















active

oldest

votes











Your Answer






StackExchange.ifUsing("editor", function () {
StackExchange.using("externalEditor", function () {
StackExchange.using("snippets", function () {
StackExchange.snippets.init();
});
});
}, "code-snippets");

StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "1"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);

StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});

function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});


}
});














draft saved

draft discarded


















StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53375128%2fspark-release-2-4-0-can-not-work-on-win7-because-of-importerror-no-module-named%23new-answer', 'question_page');
}
);

Post as a guest















Required, but never shown






























active

oldest

votes













active

oldest

votes









active

oldest

votes






active

oldest

votes
















draft saved

draft discarded




















































Thanks for contributing an answer to Stack Overflow!


  • Please be sure to answer the question. Provide details and share your research!

But avoid



  • Asking for help, clarification, or responding to other answers.

  • Making statements based on opinion; back them up with references or personal experience.


To learn more, see our tips on writing great answers.





Some of your past answers have not been well-received, and you're in danger of being blocked from answering.


Please pay close attention to the following guidance:


  • Please be sure to answer the question. Provide details and share your research!

But avoid



  • Asking for help, clarification, or responding to other answers.

  • Making statements based on opinion; back them up with references or personal experience.


To learn more, see our tips on writing great answers.




draft saved


draft discarded














StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53375128%2fspark-release-2-4-0-can-not-work-on-win7-because-of-importerror-no-module-named%23new-answer', 'question_page');
}
);

Post as a guest















Required, but never shown





















































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown

































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown







Popular posts from this blog

android studio warns about leanback feature tag usage required on manifest while using Unity exported app?

SQL update select statement

'app-layout' is not a known element: how to share Component with different Modules