ClassNotFoundException:com.datastax.spark.connector.rdd.partitioner.CassandraPartition

问题描述 投票:0回答:1

我正在使用

Spark版本2.2.1使用Scala 2.11.8版本OpenJDK 64位服务器VM,1.8.0_131

我使用代码添加了jar依赖项

JavaSparkContext sc = new JavaSparkContext(conf);
        sc.addJar("./target/CassandraSparkJava-1.0-SNAPSHOT-jar-with-dependencies.jar");

执行下面的代码,但是面对ClassNotFoundException:com.datastax.spark.connector.rdd.partitioner.CassandraPartition

Dataset<org.apache.spark.sql.Row> dataset = sparksession.read().format("org.apache.spark.sql.cassandra")
            .options(new HashMap<String, String>() {
                {
                    put("keyspace", "test_db");
                    put("table", "stats");
                }
            }).load().where("timestamp = '2018-01-09 00:00:00'");
    System.out.println(" Result " + dataset.collectAsList().size());

这是我的pom.xml文件

<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
    xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd">
    <modelVersion>4.0.0</modelVersion>
    <groupId>com.test.spark.poc</groupId>
    <artifactId>CassandraSparkJava</artifactId>
    <version>1.0-SNAPSHOT</version>
    <dependencies>
    <!-- https://mvnrepository.com/artifact/org.apache.felix/maven-bundle-plugin -->
 <dependency>
        <groupId>org.apache.spark</groupId>
        <artifactId>spark-core_2.11</artifactId>
        <version>2.2.1</version>
        <scope>provided</scope>
    </dependency>
    <dependency>
        <groupId>org.apache.spark</groupId>
        <artifactId>spark-sql_2.11</artifactId>
        <version>2.2.1</version>
        <scope>provided</scope>
    </dependency>
        <dependency>
            <groupId>com.datastax.spark</groupId>
            <artifactId>spark-cassandra-connector_2.11</artifactId>
            <version>2.0.0-M3</version>
        </dependency>
        <dependency>
            <groupId>com.datastax.cassandra</groupId>
            <artifactId>cassandra-driver-core</artifactId>
            <version>3.3.0</version>
        </dependency>
        <dependency>
      <groupId>com.google.guava</groupId>
      <artifactId>guava</artifactId>
      <version>18.0</version>
    </dependency>
    <dependency>
      <groupId>io.reactivex</groupId>
      <artifactId>rxjava</artifactId>
      <version>1.0.0-rc.8</version>
    </dependency>
    </dependencies>
    <build>
        <plugins>
        <!-- A plugin for compiling Java code -->
        <plugin>
            <groupId>org.apache.maven.plugins</groupId>
            <artifactId>maven-compiler-plugin</artifactId>
            <version>3.6.1</version>
            <configuration>
                <source>1.8</source>
                <target>1.8</target>
                <archive>
                        <manifest>
                            <mainClass>com.test.spark.poc.CassandraSparkMockStats</mainClass>
                        </manifest>
                    </archive>
            </configuration>
        </plugin>

        <!-- Or, a plugin for compiling Scala code -->
        <!-- Make sure you are not using "maven-scala-plugin", which is the older version -->
        <plugin>
            <groupId>net.alchim31.maven</groupId>
            <artifactId>scala-maven-plugin</artifactId>
            <version>3.2.2</version>
            <executions>
                <execution>
                    <goals>
                        <goal>compile</goal>
                    </goals>
                </execution>
            </executions>
        </plugin>             
            <plugin>
                <groupId>org.apache.maven.plugins</groupId>
                <artifactId>maven-assembly-plugin</artifactId>
                <version>2.4.1</version>
                <configuration>
                    <!-- get all project dependencies -->
                    <descriptorRefs>
                        <descriptorRef>jar-with-dependencies</descriptorRef>
                    </descriptorRefs>
                    <!-- MainClass in mainfest make a executable jar -->
                    <archive>
                      <manifest>
                        <mainClass>com.test.spark.poc.CassandraSparkMockStats</mainClass>
                      </manifest>
                    </archive>

                </configuration>
                <executions>
                  <execution>
                    <id>make-assembly</id>
                                        <!-- bind to the packaging phase -->
                    <phase>package</phase>
                    <goals>
                        <goal>single</goal>
                    </goals>
                  </execution>
                </executions>
            </plugin>
        </plugins>
    </build>
</project>

帮助我解决此问题

scala apache-spark datastax-java-driver spark-cassandra-connector
1个回答
0
投票

[尝试从PC上的Java(驱动程序)使用cassandra运行spark时遇到此异常。

在我的情况下,您必须将带有spark-cassandra-connector的jar添加到SparkContext中,如下例所示:

JavaSparkContext sc = new JavaSparkContext(conf);
    sc.addJar("./build/libs/spark-cassandra-connector_2.11-2.4.2.jar"); // location of driver could be different.
© www.soinside.com 2019 - 2024. All rights reserved.