Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

REL-368 #25

Merged
merged 45 commits into from
Nov 17, 2014
Merged

REL-368 #25

merged 45 commits into from
Nov 17, 2014

Conversation

markhamstra
Copy link

Adding datanucleus jars to debian package, as well as catching up to current Apache 1.1.1 release candidate.

markhamstra and others added 30 commits November 4, 2014 10:22
For me the core tests failed because there are two locale dependent parts in the code.
Look at the Jira ticket for details.

Why is it necessary to check the exception message in isBindCollision in
https://github.com/apache/spark/blob/master/core/src/main/scala/org/apache/spark/util/Utils.scala#L1686
?

Author: Niklas Wilcke <1wilcke@informatik.uni-hamburg.de>

Closes apache#3036 from numbnut/core-test-fix and squashes the following commits:

1fb0d04 [Niklas Wilcke] Fixing locale dependend code and tests

(cherry picked from commit ed8bf1e)
Signed-off-by: Andrew Or <andrew@databricks.com>
Here's my attempt to re-port `RecoverableNetworkWordCount` to Java, following the example of its Scala and Java siblings. I fixed a few minor doc/formatting issues along the way I believe.

Author: Sean Owen <sowen@cloudera.com>

Closes apache#2564 from srowen/SPARK-2548 and squashes the following commits:

0d0bf29 [Sean Owen] Update checkpoint call as in apache#2735
35f23e3 [Sean Owen] Remove old comment about running in standalone mode
179b3c2 [Sean Owen] Re-port RecoverableNetworkWordCount to Java example, and touch up doc / formatting in related examples

(cherry picked from commit 3a02d41)
Signed-off-by: Tathagata Das <tathagata.das1565@gmail.com>
Trying this example, I missed the moment when the checkpoint was iniciated

Author: comcmipi <pitonak@fns.uniba.sk>

Closes apache#2735 from comcmipi/patch-1 and squashes the following commits:

b6d8001 [comcmipi] Update RecoverableNetworkWordCount.scala
96fe274 [comcmipi] Update RecoverableNetworkWordCount.scala

(cherry picked from commit 0340c56)
Signed-off-by: Tathagata Das <tathagata.das1565@gmail.com>
…ons in branch 1.1

Author: Tathagata Das <tathagata.das1565@gmail.com>

Closes apache#3188 from tdas/branch-1.1 and squashes the following commits:

f1996d3 [Tathagata Das] [SPARK-2548][HOTFIX] Removed use of o.a.s.streaming.Durations
PR apache#3175 is for master branch only and can't be backported to branch 1.1 directly because Hive 0.13.1 support.

Author: Cheng Lian <lian@databricks.com>

Closes apache#3176 from liancheng/fix-op-state-for-1.1 and squashes the following commits:

8791d87 [Cheng Lian] This is a follow up of apache#3175 for branch 1.1
This PR backports apache#2843 to branch-1.1. The key difference is that this one doesn't support Hive 0.13.1 and thus always returns `0.12.0` when `spark.sql.hive.version` is queried.

6 other commits on which apache#2843 depends were also backported, they are:

- apache#2887 for `SessionState` lifecycle control
- apache#2675, apache#2823 & apache#3060 for major test suite refactoring and bug fixes
- apache#2164, for Parquet test suites updates
- apache#2493, for reading `spark.sql.*` configurations

Author: Cheng Lian <lian@databricks.com>
Author: Cheng Lian <lian.cs.zju@gmail.com>
Author: Michael Armbrust <michael@databricks.com>

Closes apache#3113 from liancheng/get-info-for-1.1 and squashes the following commits:

d354161 [Cheng Lian] Provides Spark and Hive version in HiveThriftServer2 for branch-1.1
0c2a244 [Michael Armbrust] [SPARK-3646][SQL] Copy SQL configuration from SparkConf when a SQLContext is created.
3202a36 [Michael Armbrust] [SQL] Decrease partitions when testing
7f395b7 [Cheng Lian] [SQL] Fixes race condition in CliSuite
0dd28ec [Cheng Lian] [SQL] Fixes the race condition that may cause test failure
5928b39 [Cheng Lian] [SPARK-3809][SQL] Fixes test suites in hive-thriftserver
faeca62 [Cheng Lian] [SPARK-4037][SQL] Removes the SessionState instance created in HiveThriftServer2
about convert files to RDDS there are 3 loops with files sequence in spark source.
loops files sequence:
1.files.map(...)
2.files.zip(fileRDDs)
3.files-size.foreach
It's will very time consuming when lots of files.So I do the following correction:
3 loops with files sequence => only one loop

Author: surq <surq@asiainfo.com>

Closes apache#2811 from surq/SPARK-3954 and squashes the following commits:

321bbe8 [surq]  updated the code style.The style from [for...yield]to [files.map(file=>{})]
88a2c20 [surq] Merge branch 'master' of https://github.com/apache/spark into SPARK-3954
178066f [surq] modify code's style. [Exceeds 100 columns]
626ef97 [surq] remove redundant import(ArrayBuffer)
739341f [surq] promote the speed of convert files to RDDS

(cherry picked from commit ce6ed2a)
Signed-off-by: Tathagata Das <tathagata.das1565@gmail.com>
…master to branch 1.1

The original PR was apache#2366

This backport was non-trivial because Spark 1.1 uses ConnectionManager instead of NioBlockTransferService, which required slight modification to unit tests. Other than that the code is exactly same as in the original PR. Please refer to discussion in the original PR if you have any thoughts.

Author: Tathagata Das <tathagata.das1565@gmail.com>

Closes apache#3191 from tdas/replication-fix-branch-1.1-backport and squashes the following commits:

593214a [Tathagata Das] Merge remote-tracking branch 'apache-github/branch-1.1' into branch-1.1
2ed927f [Tathagata Das] Fixed error in unit test.
de4ff73 [Tathagata Das] [SPARK-3495] Block replication fails continuously when the replication target node is dead AND [SPARK-3496] Block replication by mistake chooses driver as target
Author: Michael Armbrust <michael@databricks.com>
Author: ravipesala <ravindra.pesala@huawei.com>
Author: scwf <wangfei1@huawei.com>

Closes apache#3199 from marmbrus/backport1.1 and squashes the following commits:

019a0dd [Michael Armbrust] Drop incorrectly ported test cases
4c9f3e6 [ravipesala] [SPARK-3708][SQL] Backticks aren't handled correctly is aliases
064750d [scwf] [SPARK-3704][SQL] Fix ColumnValue type for Short values in thrift server
f4e17cd [ravipesala] [SPARK-3834][SQL] Backticks not correctly handled in subquery aliases
In running-on-yarn.md, a link to YARN overview is here.
But the URL is to YARN alpha's.
It should be stable's.

Author: Kousuke Saruta <sarutak@oss.nttdata.co.jp>

Closes apache#3196 from sarutak/SPARK-4330 and squashes the following commits:

30baa21 [Kousuke Saruta] Fixed running-on-yarn.md to point proper URL for YARN

(cherry picked from commit 3c07b8f)
Signed-off-by: Matei Zaharia <matei@databricks.com>
…rk back to Kryo

This reverts apache#2916 . We shouldn't change the default settings in a minor release. JoshRosen davies

Author: Xiangrui Meng <meng@databricks.com>

Closes apache#3187 from mengxr/SPARK-2652-1.1 and squashes the following commits:

372166b [Xiangrui Meng] change the default spark.serializer in pyspark back to Kryo
Because we switched back to Kryo in apache#3187 , we need to leave a note about the workaround.

Author: Xiangrui Meng <meng@databricks.com>

Closes apache#3190 from mengxr/SPARK-3990-1.1 and squashes the following commits:

d4818f3 [Xiangrui Meng] fix python style
53725b0 [Xiangrui Meng] add a note about SPARK-3990
56ad70e [Xiangrui Meng] add a note about SPARK-3990
Handle exception in SparkSinkSuite, please refer to [SPARK-4295]

Author: maji2014 <maji3@asiainfo.com>

Closes apache#3177 from maji2014/spark-4295 and squashes the following commits:

312620a [maji2014] change a new statement for spark-4295
24c3d21 [maji2014] add log4j.properties for SparkSinkSuite and spark-4295
c807bf6 [maji2014] Fix exception in SparkSinkSuite

(cherry picked from commit f8811a5)
Signed-off-by: Tathagata Das <tathagata.das1565@gmail.com>
Conflicts:
	docs/_config.yml
	yarn/alpha/pom.xml
This reverts commit f3e62ff.

Conflicts:
	pom.xml
andrewor14 and others added 15 commits November 12, 2014 19:01
This involves a few main changes:
- Log all output message to the log file. Previously the log file
  was not useful because it did not indicate progress.
- Remove hive-site.xml in sbt_hive_app to avoid interference
- Add the appropriate repositories for new dependencies
Conflicts:
	assembly/pom.xml
	bagel/pom.xml
	core/pom.xml
	examples/pom.xml
	external/flume-sink/pom.xml
	external/flume/pom.xml
	external/kafka/pom.xml
	external/mqtt/pom.xml
	external/twitter/pom.xml
	external/zeromq/pom.xml
	extras/kinesis-asl/pom.xml
	extras/spark-ganglia-lgpl/pom.xml
	graphx/pom.xml
	mllib/pom.xml
	pom.xml
	repl/pom.xml
	sql/catalyst/pom.xml
	sql/core/pom.xml
	sql/hive-thriftserver/pom.xml
	sql/hive/pom.xml
	streaming/pom.xml
	tools/pom.xml
	yarn/pom.xml
	yarn/stable/pom.xml
andrewor14 This backports the bug fix in apache#3220 . It would be good if we can get it in 1.1.1. But this is minor.

Author: Xiangrui Meng <meng@databricks.com>

Closes apache#3251 from mengxr/SPARK-4355-1.1 and squashes the following commits:

33886b6 [Xiangrui Meng] Merge remote-tracking branch 'apache/branch-1.1' into SPARK-4355-1.1
91fe1a3 [Xiangrui Meng] fix OnlineSummarizer.merge when other.mean is zero
markhamstra added a commit that referenced this pull request Nov 17, 2014
@markhamstra markhamstra merged commit 38fde70 into alteryx:csd-1.1 Nov 17, 2014
markhamstra pushed a commit that referenced this pull request Apr 18, 2015
Invoking .size on arrays is valid, but requires an implicit conversion to SeqLike. This incurs a compile time overhead and more importantly a runtime overhead, as the Array must be wrapped before the method can be invoked. For example, the difference in generated byte code is:

  public int withSize();
    Code:
       0: getstatic     #23                 // Field scala/Predef$.MODULE$:Lscala/Predef$;
       3: aload_0
       4: invokevirtual #25                 // Method array:()[I
       7: invokevirtual #29                 // Method scala/Predef$.intArrayOps:([I)Lscala/collection/mutable/ArrayOps;
      10: invokeinterface #34,  1           // InterfaceMethod scala/collection/mutable/ArrayOps.size:()I
      15: ireturn

  public int withLength();
    Code:
       0: aload_0
       1: invokevirtual #25                 // Method array:()[I
       4: arraylength
       5: ireturn

Author: sksamuel <sam@sksamuel.com>

Closes apache#5376 from sksamuel/master and squashes the following commits:

77ec261 [sksamuel] Replace use of .size with .length for Arrays.
markhamstra pushed a commit to markhamstra/spark that referenced this pull request Nov 7, 2017
* Improved the example commands in running-on-k8s document.

* Fixed more example commands.

* Fixed typo.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

10 participants