Can I use spark computer on CQL without hadoop cluster


Pawan Shriwas
 

Thanks team,

I will check it based on your response and let you know if anything is needed.


On Wed, Jan 5, 2022 at 8:24 PM <hadoopmarc@...> wrote:
It is also possible to run spark on kubernetes (in combination with distributed storage like S3 or minio):
https://spark.apache.org/docs/latest/running-on-kubernetes.html

It will require some time to get your head around this, but note that you can do this with or without the spark operator installed on your kubernetes cluster.

Marc



--
Thanks & Regard

PAWAN SHRIWAS


hadoopmarc@...
 

It is also possible to run spark on kubernetes (in combination with distributed storage like S3 or minio):
https://spark.apache.org/docs/latest/running-on-kubernetes.html

It will require some time to get your head around this, but note that you can do this with or without the spark operator installed on your kubernetes cluster.

Marc


Boxuan Li
 

Hi Pawan,

Do you want to run Spark traversal on a Spark standalone cluster rather than a Hadoop Yarn cluster? In that case, you could follow the JanusGraph documentation or check out this guide on Medium.

Best,
Boxuan


Pawan Shriwas
 

Hi All,

I am checking the possibility of using a graph computer using spark on CQL backend without hadoop installation.

Please let me know if we can do this and how can i achieve this. I don't want to introduce hadoop cluster just because of this use case. I will appreciate  if anyone can share some resources around it.
  
Thanks,
Pawan