Getting value in a dataframe in PySpark
.everyoneloves__top-leaderboard:empty,.everyoneloves__mid-leaderboard:empty,.everyoneloves__bot-mid-leaderboard:empty{ height:90px;width:728px;box-sizing:border-box;
}
I have the below dataframe and I'm trying to get the value 3097 as a int, e.g. storing it in a python variable to manipulate it, multiply it by another int etc.
I've managed to get the row, but I don't even now if it's a good way to do it and I still can't have the value as a int.
data.groupBy("card_bank", "failed").count().filter(data["failed"] == "true").collect()
python apache-spark pyspark
add a comment |
I have the below dataframe and I'm trying to get the value 3097 as a int, e.g. storing it in a python variable to manipulate it, multiply it by another int etc.
I've managed to get the row, but I don't even now if it's a good way to do it and I still can't have the value as a int.
data.groupBy("card_bank", "failed").count().filter(data["failed"] == "true").collect()
python apache-spark pyspark
add a comment |
I have the below dataframe and I'm trying to get the value 3097 as a int, e.g. storing it in a python variable to manipulate it, multiply it by another int etc.
I've managed to get the row, but I don't even now if it's a good way to do it and I still can't have the value as a int.
data.groupBy("card_bank", "failed").count().filter(data["failed"] == "true").collect()
python apache-spark pyspark
I have the below dataframe and I'm trying to get the value 3097 as a int, e.g. storing it in a python variable to manipulate it, multiply it by another int etc.
I've managed to get the row, but I don't even now if it's a good way to do it and I still can't have the value as a int.
data.groupBy("card_bank", "failed").count().filter(data["failed"] == "true").collect()
python apache-spark pyspark
python apache-spark pyspark
asked Jan 3 at 9:57
LaSulLaSul
615321
615321
add a comment |
add a comment |
3 Answers
3
active
oldest
votes
Get the first record from the Row object using index 0 and get the value using the index "count"
from pyspark.sql.functions import col
data.groupby("card_bank", "failed").count().filter(col("failed") == "true").collect()[0]["count"]
1
Exactly what I was looking for :)
– LaSul
Jan 3 at 10:12
add a comment |
You need to get a row
from the sequence (wither for loop or map function) and then row.getInt(2)
according to https://spark.apache.org/docs/1.4.0/api/java/org/apache/spark/sql/Row.html.
add a comment |
Try selecting the value from spark dataframe :-
df =data.groupBy("card_bank", "failed").count().filter(data["failed"] == "true").collect()
value = df.select("count").as[int].collect()
here, value will be a list.
add a comment |
Your Answer
StackExchange.ifUsing("editor", function () {
StackExchange.using("externalEditor", function () {
StackExchange.using("snippets", function () {
StackExchange.snippets.init();
});
});
}, "code-snippets");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "1"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f54019931%2fgetting-value-in-a-dataframe-in-pyspark%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
3 Answers
3
active
oldest
votes
3 Answers
3
active
oldest
votes
active
oldest
votes
active
oldest
votes
Get the first record from the Row object using index 0 and get the value using the index "count"
from pyspark.sql.functions import col
data.groupby("card_bank", "failed").count().filter(col("failed") == "true").collect()[0]["count"]
1
Exactly what I was looking for :)
– LaSul
Jan 3 at 10:12
add a comment |
Get the first record from the Row object using index 0 and get the value using the index "count"
from pyspark.sql.functions import col
data.groupby("card_bank", "failed").count().filter(col("failed") == "true").collect()[0]["count"]
1
Exactly what I was looking for :)
– LaSul
Jan 3 at 10:12
add a comment |
Get the first record from the Row object using index 0 and get the value using the index "count"
from pyspark.sql.functions import col
data.groupby("card_bank", "failed").count().filter(col("failed") == "true").collect()[0]["count"]
Get the first record from the Row object using index 0 and get the value using the index "count"
from pyspark.sql.functions import col
data.groupby("card_bank", "failed").count().filter(col("failed") == "true").collect()[0]["count"]
answered Jan 3 at 10:10
Ernest S KirubakaranErnest S Kirubakaran
911510
911510
1
Exactly what I was looking for :)
– LaSul
Jan 3 at 10:12
add a comment |
1
Exactly what I was looking for :)
– LaSul
Jan 3 at 10:12
1
1
Exactly what I was looking for :)
– LaSul
Jan 3 at 10:12
Exactly what I was looking for :)
– LaSul
Jan 3 at 10:12
add a comment |
You need to get a row
from the sequence (wither for loop or map function) and then row.getInt(2)
according to https://spark.apache.org/docs/1.4.0/api/java/org/apache/spark/sql/Row.html.
add a comment |
You need to get a row
from the sequence (wither for loop or map function) and then row.getInt(2)
according to https://spark.apache.org/docs/1.4.0/api/java/org/apache/spark/sql/Row.html.
add a comment |
You need to get a row
from the sequence (wither for loop or map function) and then row.getInt(2)
according to https://spark.apache.org/docs/1.4.0/api/java/org/apache/spark/sql/Row.html.
You need to get a row
from the sequence (wither for loop or map function) and then row.getInt(2)
according to https://spark.apache.org/docs/1.4.0/api/java/org/apache/spark/sql/Row.html.
answered Jan 3 at 10:05


windwind
360215
360215
add a comment |
add a comment |
Try selecting the value from spark dataframe :-
df =data.groupBy("card_bank", "failed").count().filter(data["failed"] == "true").collect()
value = df.select("count").as[int].collect()
here, value will be a list.
add a comment |
Try selecting the value from spark dataframe :-
df =data.groupBy("card_bank", "failed").count().filter(data["failed"] == "true").collect()
value = df.select("count").as[int].collect()
here, value will be a list.
add a comment |
Try selecting the value from spark dataframe :-
df =data.groupBy("card_bank", "failed").count().filter(data["failed"] == "true").collect()
value = df.select("count").as[int].collect()
here, value will be a list.
Try selecting the value from spark dataframe :-
df =data.groupBy("card_bank", "failed").count().filter(data["failed"] == "true").collect()
value = df.select("count").as[int].collect()
here, value will be a list.
answered Jan 3 at 10:07
SreejithSreejith
463
463
add a comment |
add a comment |
Thanks for contributing an answer to Stack Overflow!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f54019931%2fgetting-value-in-a-dataframe-in-pyspark%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown