Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Java 11 support #356

Closed
ruebot opened this issue Aug 27, 2019 · 5 comments
Closed

Java 11 support #356

ruebot opened this issue Aug 27, 2019 · 5 comments
Assignees
Labels

Comments

@ruebot
Copy link
Member

@ruebot ruebot commented Aug 27, 2019

From the Apache Spark mailing list:

Hi, All.

Thanks to your many many contributions,
Apache Spark master branch starts to pass on JDK11 as of today.
(with `hadoop-3.2` profile: Apache Hadoop 3.2 and Hive 2.3.6)

    https://amplab.cs.berkeley.edu/jenkins/view/Spark%20QA%20Test%20(Dashboard)/job/spark-master-test-maven-hadoop-3.2-jdk-11/326/
    (JDK11 is used for building and testing.)

We already verified all UTs (including PySpark/SparkR) before.

Please feel free to use JDK11 in order to build/test/run `master` branch and
share your experience including any issues. It will help Apache Spark 3.0.0 release.

For the follow-ups, please follow https://issues.apache.org/jira/browse/SPARK-24417 .
The next step is `how to support JDK8/JDK11 together in a single artifact`.

Bests,
Dongjoon.

We'll align with Apache Spark here on Java 11 support. Once we have a Spark release with Java 11, I'll pivot to getting aut stable with Java 11.

@ruebot ruebot added the Java label Aug 27, 2019
@ruebot ruebot self-assigned this Aug 27, 2019
ruebot added a commit that referenced this issue Aug 31, 2019
@ruebot
Copy link
Member Author

@ruebot ruebot commented Nov 7, 2019

Getting closer to Spark 3.0.0!

Hi all,

To enable wide-scale community testing of the upcoming Spark 3.0 release, the Apache Spark community has posted a preview release of Spark 3.0. This preview is not a stable release in terms of either API or functionality, but it is meant to give the community early access to try the code that will become Spark 3.0. If you would like to test the release, please download it, and send feedback using either the mailing lists or JIRA.

There are a lot of exciting new features added to Spark 3.0, including Dynamic Partition Pruning, Adaptive Query Execution, Accelerator-aware Scheduling, Data Source API with Catalog Supports, Vectorization in SparkR, support of Hadoop 3/JDK 11/Scala 2.12, and many more. For a full list of major features and changes in Spark 3.0.0-preview, please check the thread(http://apache-spark-developers-list.1001551.n3.nabble.com/Spark-3-0-preview-release-feature-list-and-major-changes-td28050.html).

We'd like to thank our contributors and users for their contributions and early feedback to this release. This release would not have been possible without you.

To download Spark 3.0.0-preview, head over to the download page: https://archive.apache.org/dist/spark/spark-3.0.0-preview

Thanks,

Xingbo
ruebot added a commit that referenced this issue Nov 9, 2019
- Some hacks to get a sucessful build
- Definitely need to loop back and clean-up a whole lot!
- Addresses #356
@ruebot
Copy link
Member Author

@ruebot ruebot commented Nov 9, 2019

Successful build:

[INFO] ------------------------------------------------------------------------
[INFO] BUILD SUCCESS
[INFO] ------------------------------------------------------------------------
[INFO] Total time:  01:58 min
[INFO] Finished at: 2019-11-09T18:19:56-05:00
[INFO] ------------------------------------------------------------------------

Successful Spark 3.0.0-preview load with ~/bin/spark-3.0.0-preview-bin-hadoop2.7/bin/spark-shell --master local\[10\] --driver-memory 35g --packages io.archivesunleashed:aut:0.18.1-SNAPSHOT:

Welcome to
      ____              __
     / __/__  ___ _____/ /__
    _\ \/ _ \/ _ `/ __/  '_/
   /___/ .__/\_,_/_/ /_/\_\   version 3.0.0-preview
      /_/
         
Using Scala version 2.12.10 (OpenJDK 64-Bit Server VM, Java 11.0.4)
Type in expressions to have them evaluated.
Type :help for more information.

scala> 

BUT! We're broken somewhere:

scala> :paste
// Entering paste mode (ctrl-D to finish)

import io.archivesunleashed._
import io.archivesunleashed.matchbox._

RecordLoader.loadArchives("/home/nruest/Projects/au/aut/src/test/resources/arc/example.arc.gz", sc).keepValidPages()
  .map(r => ExtractDomain(r.getUrl))
  .countItems()
  .take(10)

// Exiting paste mode, now interpreting.

java.lang.NoSuchMethodError: 'scala.collection.mutable.ArrayOps scala.Predef$.refArrayOps(java.lang.Object[])'
  at io.archivesunleashed.package$RecordLoader$.getFiles(package.scala:54)
  at io.archivesunleashed.package$RecordLoader$.loadArchives(package.scala:68)
  ... 47 elided

I'll have to dig in more later.

ruebot added a commit that referenced this issue Nov 10, 2019
@ruebot
Copy link
Member Author

@ruebot ruebot commented Nov 10, 2019

Successful run here: 06764f7!!

java11-spark-3 0 0-preview-jars-success

ruebot added a commit that referenced this issue Nov 10, 2019
…talled 🤦, and a bunch more pom cleanup.
ruebot added a commit that referenced this issue Nov 10, 2019
ruebot added a commit that referenced this issue Nov 10, 2019
ruebot added a commit that referenced this issue Nov 10, 2019
ruebot added a commit that referenced this issue Nov 10, 2019
ruebot added a commit that referenced this issue Nov 10, 2019
ruebot added a commit that referenced this issue Nov 10, 2019
ruebot added a commit that referenced this issue Nov 10, 2019
@ruebot ruebot added this to To Do in 1.0.0 Release of AUT Nov 14, 2019
@ruebot
Copy link
Member Author

@ruebot ruebot commented Dec 24, 2019

Hi all,

To enable wide-scale community testing of the upcoming Spark 3.0 release, the Apache Spark community has posted a new preview release of Spark 3.0. This preview is not a stable release in terms of either API or functionality, but it is meant to give the community early access to try the code that will become Spark 3.0. If you would like to test the release, please download it, and send feedback using either the mailing lists or JIRA.

There are a lot of exciting new features added to Spark 3.0, including Dynamic Partition Pruning, Adaptive Query Execution, Accelerator-aware Scheduling, Data Source API with Catalog Supports, Vectorization in SparkR, support of Hadoop 3/JDK 11/Scala 2.12, and many more. For a full list of major features and changes in Spark 3.0.0-preview2, please check the thread(http://apache-spark-developers-list.1001551.n3.nabble.com/Spark-3-0-preview-release-feature-list-and-major-changes-td28050.html and http://apache-spark-developers-list.1001551.n3.nabble.com/Spark-3-0-preview-release-2-td28491.html).

We'd like to thank our contributors and users for their contributions and early feedback to this release. This release would not have been possible without you.

To download Spark 3.0.0-preview2, head over to the download page: https://archive.apache.org/dist/spark/spark-3.0.0-preview2

Happy Holidays.

Yuming

I'll get the branch updated shortly.

@ruebot ruebot moved this from To Do to In Progress in 1.0.0 Release of AUT May 19, 2020
@ruebot
Copy link
Member Author

@ruebot ruebot commented Jun 23, 2020

Resolved with: 59b1d4e

@ruebot ruebot closed this Jun 23, 2020
1.0.0 Release of AUT automation moved this from In Progress to Done Jun 23, 2020
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
Linked pull requests

Successfully merging a pull request may close this issue.

None yet
1 participant
You can’t perform that action at this time.