Created on 06-06-2016 07:20 PM - edited 09-16-2022 03:23 AM
Hi,
We have started studies in order to implement Bigdata Continuous Delivery process. We'd like to know if someone has been implemented.
What we need is to know if there is any 'best practices' for:
Basically we develop: Hive, Python(spark), Shellscript, Flume, Sqoop. After all of above defined, we would like to provision these envs in containers to set a continuous integration deployment via:
Mesos + Jenkins + Marathon + Docker Containers to spin up dockers with Horton HDP 2.2.0. (same as production env)
Many thanks,
Fabricio
Created 06-12-2016 11:22 AM
Sure so basically regarding the cluster, you may find useful to
Created 09-28-2017 04:58 AM
Hello Everyone,
We are using scala with maven to build spark applications along with git as code repository and jenkins integrated with git to build the jar.
I am not sure how to use jenkins to deploy our apps on cluster.
Can anyone explain what could be the next step?
Is jenkins supporting deployment of spark apps like it does for other apps.
Tha ks
Created 12-12-2018 02:09 PM
Dear @Fabricio Carboni:
Can you please share some document on how we can implement CI/CD for pyspark based applications. Also, is it possbile to do it without using containers (like we do development in Java/Scala (first locally on windows and then build it on Linux dev/tst/Prod))
Thanks
Abhinav
Created 02-08-2019 02:58 PM
Hi, Were you able to find a solution to this? We have a similar setup and I can't seem to find any examples of that.