I am attempting to configure Databricks with DSE BYOS. My Databricks Spark instances can see our Postgres. I'd like to connect them also to DSE Graph so that I can set up a data load pipeline from Postgres to DSE Graph.
So I'm wondering if I really need the DSEFS (hadoop) functionality. Of the various spark configuration items from the BYOS commands, what are the minimal ones required to support access to DSE Graph?
For example, my dse-byos.properties look's like the following. Does BYOS have to use DSEFS and DSE for the Hive metastore? Can we just connect our Databricks Spark to Cassandra & DSE Graph and let Databricks use its own resources for the rest, or is DseGraphFrames functionality dependent upon access to DSEFS?
spark.hadoop.cassandra.host 172.xxx.xxx.xxx spark.hadoop.cassandra.auth.kerberos.enabled false spark.cassandra.auth.conf.factory com.datastax.bdp.spark.DseByosAuthConfFactory spark.hadoop.fs.dsefs.impl com.datastax.bdp.fs.hadoop.DseFileSystem spark.sql.extensions com.datastax.spark.connector.DseSparkExtensions spark.hadoop.dse.advanced_replication.directory /path/to/cassandra/advrep spark.hadoop.com.datastax.bdp.fs.client.authentication.factory com.datastax.bdp.fs.hadoop.DseRestClientAuthProviderBuilderFactory spark.cassandra.connection.port 9042 spark.hadoop.cassandra.ssl.enabled false spark.hadoop.cassandra.auth.kerberos.defaultScheme false spark.cassandra.connection.host 172.xxx.xxx.xxx spark.hadoop.cassandra.ssl.optional false spark.hadoop.cassandra.connection.native.port 9042 spark.hadoop.dse.client.configuration.impl com.datastax.bdp.transport.client.HadoopBasedClientConfiguration spark.cassandra.connection.factory com.datastax.bdp.spark.DseCassandraConnectionFactory spark.hadoop.cassandra.config.loader com.datastax.bdp.config.DseConfigurationLoader spark.sql.hive.metastore.sharedPrefixes com.typesafe.scalalogging spark.hadoop.dse.system_memory_in_mb 15745 spark.cassandra.dev.customFromDriver com.datastax.spark.connector.types.DseTypeConverter spark.hadoop.cassandra.partitioner org.apache.cassandra.dht.Murmur3Partitioner spark.hadoop.cassandra.dsefs.port 5598