-
Notifications
You must be signed in to change notification settings - Fork 28.7k
[SPARK-4695][SQL] Get result using executeCollect #3547
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Test build #24023 has started for PR 3547 at commit
|
Test build #24023 has finished for PR 3547 at commit
|
Test FAILed. |
@liancheng can you take a look at this (the test failure)? Thanks. |
I am looking this. Here are some case we not cover in HiveContext toHiveString method |
I will fix this |
This reverts commit 0db7ce8.
Test build #24043 has started for PR 3547 at commit
|
Test build #24043 has finished for PR 3547 at commit
|
Test PASSed. |
@@ -416,6 +416,8 @@ object HiveContext { | |||
case (bin: Array[Byte], BinaryType) => new String(bin, "UTF-8") | |||
case (decimal: Decimal, DecimalType()) => // Hive strips trailing zeros so use its toString | |||
HiveShim.createDecimal(decimal.toBigDecimal.underlying()).toString |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I believe this case can be deleted now. Internal decimal types should never be returned to users.
Thanks for tracking this down! I'm going to go ahead and merge now so we can include this in 1.2. Can you please open a follow up PR to remove the unnecessary condition? |
Using ```executeCollect``` to collect the result, because executeCollect is a custom implementation of collect in spark sql which better than rdd's collect Author: wangfei <[email protected]> Closes #3547 from scwf/executeCollect and squashes the following commits: a5ab68e [wangfei] Revert "adding debug info" a60d680 [wangfei] fix test failure 0db7ce8 [wangfei] adding debug info 184c594 [wangfei] using executeCollect instead collect (cherry picked from commit 3ae0cda) Signed-off-by: Michael Armbrust <[email protected]>
Sure, opened #3563 to delete it. |
a follow up of #3547 /cc marmbrus Author: scwf <[email protected]> Closes #3563 from scwf/rnc and squashes the following commits: 9395661 [scwf] remove unnecessary condition
Using
executeCollect
to collect the result, because executeCollect is a custom implementation of collect in spark sql which better than rdd's collect