<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Bigdata Continuous Delivery in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123053#M85806</link>
    <description>&lt;P&gt;Hi, &lt;/P&gt;&lt;P&gt;We have started studies in order to implement Bigdata Continuous Delivery process. We'd like to know if someone has been implemented. &lt;/P&gt;&lt;P&gt;What we need is to know if there is any 'best practices' for:&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;Dev environment&lt;/LI&gt;&lt;LI&gt;Building process&lt;/LI&gt;&lt;LI&gt;Deploy on unit test env&lt;/LI&gt;&lt;LI&gt;Deploy on integration test env&lt;/LI&gt;&lt;LI&gt;Deploy on production &lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;Basically we develop: Hive, Python(spark), Shellscript, Flume, Sqoop.
After all of above defined, we would like to provision these envs in containers to set a 
continuous integration deployment via: &lt;/P&gt;&lt;P&gt;Mesos + Jenkins + Marathon + Docker Containers to spin up dockers with Horton HDP 2.2.0. (same as production env)&lt;/P&gt;&lt;P&gt;Many thanks, &lt;/P&gt;&lt;P&gt;Fabricio&lt;/P&gt;</description>
    <pubDate>Fri, 16 Sep 2022 10:23:43 GMT</pubDate>
    <dc:creator>fabricio_carbon</dc:creator>
    <dc:date>2022-09-16T10:23:43Z</dc:date>
    <item>
      <title>Bigdata Continuous Delivery</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123053#M85806</link>
      <description>&lt;P&gt;Hi, &lt;/P&gt;&lt;P&gt;We have started studies in order to implement Bigdata Continuous Delivery process. We'd like to know if someone has been implemented. &lt;/P&gt;&lt;P&gt;What we need is to know if there is any 'best practices' for:&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;Dev environment&lt;/LI&gt;&lt;LI&gt;Building process&lt;/LI&gt;&lt;LI&gt;Deploy on unit test env&lt;/LI&gt;&lt;LI&gt;Deploy on integration test env&lt;/LI&gt;&lt;LI&gt;Deploy on production &lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;Basically we develop: Hive, Python(spark), Shellscript, Flume, Sqoop.
After all of above defined, we would like to provision these envs in containers to set a 
continuous integration deployment via: &lt;/P&gt;&lt;P&gt;Mesos + Jenkins + Marathon + Docker Containers to spin up dockers with Horton HDP 2.2.0. (same as production env)&lt;/P&gt;&lt;P&gt;Many thanks, &lt;/P&gt;&lt;P&gt;Fabricio&lt;/P&gt;</description>
      <pubDate>Fri, 16 Sep 2022 10:23:43 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123053#M85806</guid>
      <dc:creator>fabricio_carbon</dc:creator>
      <dc:date>2022-09-16T10:23:43Z</dc:date>
    </item>
    <item>
      <title>Re: Bigdata Continuous Delivery</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123054#M85807</link>
      <description>&lt;P&gt;Dear Fabricio,&lt;/P&gt;&lt;P&gt;Yes we have several customers working on this topic. It is a interesting one. From what I have seen last time, the architecture was based on 2 real clusters, one PROD, one DR + TEST + INTEGRATION, with YARN queue and HDFS quota configured accordingly. Jenkins + SVN to take care of the versioning + build + test.&lt;/P&gt;&lt;P&gt;Some great team has build also their own project to validate the dev and follow the deployment across different environments.&lt;/P&gt;&lt;P&gt;I don't know too much about Docker, Mesos, Marathon so can't answer for this part.&lt;/P&gt;&lt;P&gt;Can you perhaps give me more details about what you are looking for ? What did you try ?&lt;/P&gt;&lt;P&gt;Kind regards.&lt;/P&gt;</description>
      <pubDate>Thu, 09 Jun 2016 18:01:43 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123054#M85807</guid>
      <dc:creator>mlanciaux</dc:creator>
      <dc:date>2016-06-09T18:01:43Z</dc:date>
    </item>
    <item>
      <title>Re: Bigdata Continuous Delivery</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123055#M85808</link>
      <description>&lt;P&gt;Hi mlanciaux,&lt;/P&gt;&lt;P&gt;Thanks for  your reply.&lt;/P&gt;&lt;P&gt;Let's put aside Docker, Mesos and Marathon. It was a way I've found to follow.&lt;/P&gt;&lt;P&gt;We do not have 2 clusters but something like a dev one. A small portion of the production env. So let's supose DEV + TEST + INTEGRATION on this small one.&lt;/P&gt;&lt;P&gt;I wonder if you could help me sharing with me some paper were I could start with. I've found lot of information and differents approaches. Is there anything Horton could recommend thinking the same way jenkins + SVN or Git.&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;&lt;P&gt;Fabricio &lt;/P&gt;</description>
      <pubDate>Sat, 11 Jun 2016 00:48:31 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123055#M85808</guid>
      <dc:creator>fabricio_carbon</dc:creator>
      <dc:date>2016-06-11T00:48:31Z</dc:date>
    </item>
    <item>
      <title>Re: Bigdata Continuous Delivery</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123056#M85809</link>
      <description>&lt;P&gt;Sure so basically regarding the cluster, you may find useful to &lt;/P&gt;&lt;UL&gt;&lt;LI&gt;Configure queue with Capacity Scheduler (production, dev, integration, test), use elasticity and preemption&lt;/LI&gt;&lt;LI&gt;Map users to queue&lt;/LI&gt;&lt;LI&gt;You can use naming convention for queue and users by specifying -dev or -test &lt;/LI&gt;&lt;LI&gt;Depending the tool you are using you can use&lt;UL&gt;&lt;LI&gt;Different database names with Hive&lt;/LI&gt;&lt;LI&gt;Different directories with HDFS + quotas&lt;/LI&gt;&lt;LI&gt;Namespace for HBase&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;Ranger will help you configure the permission for each user / group to access the right resource&lt;/LI&gt;&lt;LI&gt;Each user will have different environnement settings &lt;/LI&gt;&lt;LI&gt;Use Jenkins and Maven (if needed) to build, push the code (with SSH plugin) and run the test&lt;/LI&gt;&lt;LI&gt;Use template to provide tools to the user with logging features / correct parameter and option&lt;/LI&gt;&lt;/UL&gt;</description>
      <pubDate>Sun, 12 Jun 2016 18:22:06 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123056#M85809</guid>
      <dc:creator>mlanciaux</dc:creator>
      <dc:date>2016-06-12T18:22:06Z</dc:date>
    </item>
    <item>
      <title>Re: Bigdata Continuous Delivery</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123057#M85810</link>
      <description>&lt;P&gt;Ok Thanks!&lt;/P&gt;&lt;P&gt;Regarding the cluster we are almost ok.&lt;/P&gt;&lt;P&gt;My concern is about last two options.&lt;/P&gt;&lt;P&gt;Would you have a specific documentation/configuration regarding installing Jenkins properly to deal with a Horton cluster&lt;/P&gt;</description>
      <pubDate>Mon, 13 Jun 2016 23:52:30 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123057#M85810</guid>
      <dc:creator>fabricio_carbon</dc:creator>
      <dc:date>2016-06-13T23:52:30Z</dc:date>
    </item>
    <item>
      <title>Re: Bigdata Continuous Delivery</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123058#M85811</link>
      <description>&lt;P&gt;I think the key point is to configure the different Jenkins to be able to use the edge node via the different SSH plugin (or install it there), the rest is a matter of configuring security, backup, and choose the right number of parameter to fit your usage and switch easily from one environment to the other (dev, test, prod) &lt;/P&gt;</description>
      <pubDate>Tue, 14 Jun 2016 02:58:21 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123058#M85811</guid>
      <dc:creator>mlanciaux</dc:creator>
      <dc:date>2016-06-14T02:58:21Z</dc:date>
    </item>
    <item>
      <title>Re: Bigdata Continuous Delivery</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123059#M85812</link>
      <description>&lt;P&gt;This is a simple example here : &lt;/P&gt;&lt;P&gt;&lt;A href="https://community.hortonworks.com/articles/40171/simple-example-of-jenkins-hdp-integration.html"&gt;https://community.hortonworks.com/articles/40171/simple-example-of-jenkins-hdp-integration.html&lt;/A&gt;, I will add more later&lt;/P&gt;</description>
      <pubDate>Thu, 16 Jun 2016 20:15:45 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123059#M85812</guid>
      <dc:creator>mlanciaux</dc:creator>
      <dc:date>2016-06-16T20:15:45Z</dc:date>
    </item>
    <item>
      <title>Re: Bigdata Continuous Delivery</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123060#M85813</link>
      <description>&lt;P&gt;And don't forget to check this best practices : &lt;A target="_blank" href="https://wiki.jenkins-ci.org/display/JENKINS/Jenkins+Best+Practices"&gt;https://wiki.jenkins-ci.org/display/JENKINS/Jenkins+Best+Practices&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 16 Jun 2016 22:23:12 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123060#M85813</guid>
      <dc:creator>mlanciaux</dc:creator>
      <dc:date>2016-06-16T22:23:12Z</dc:date>
    </item>
    <item>
      <title>Re: Bigdata Continuous Delivery</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123061#M85814</link>
      <description>&lt;P style="margin-left: 20px;"&gt;Thanks &lt;A rel="user" href="https://community.cloudera.com/users/434/mlanciaux.html" nodeid="434"&gt;@mlanciaux&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 17 Jun 2016 00:06:38 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123061#M85814</guid>
      <dc:creator>fabricio_carbon</dc:creator>
      <dc:date>2016-06-17T00:06:38Z</dc:date>
    </item>
    <item>
      <title>Re: Bigdata Continuous Delivery</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123062#M85815</link>
      <description>&lt;P&gt;Dear Fabricio, I successfully make a workflow to tun from my local VM to my Hadoop remote Hadoop cluster by changing the SSH connection property. Hope that helps. Kind regards.&lt;/P&gt;</description>
      <pubDate>Mon, 20 Jun 2016 10:50:05 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123062#M85815</guid>
      <dc:creator>mlanciaux</dc:creator>
      <dc:date>2016-06-20T10:50:05Z</dc:date>
    </item>
    <item>
      <title>Re: Bigdata Continuous Delivery</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123063#M85816</link>
      <description>&lt;P&gt;Hello Everyone,&lt;/P&gt;&lt;P&gt;We are using scala with maven to build spark applications along with git as code repository and jenkins integrated with git to build the jar.&lt;/P&gt;&lt;P&gt;I am not sure how to use jenkins to deploy our apps on cluster.&lt;/P&gt;&lt;P&gt;Can anyone explain what could be the next step?&lt;/P&gt;&lt;P&gt;Is jenkins supporting deployment of spark apps like it does for other apps.&lt;/P&gt;&lt;P&gt;Tha ks&lt;/P&gt;</description>
      <pubDate>Thu, 28 Sep 2017 11:58:57 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123063#M85816</guid>
      <dc:creator>varunjoshi</dc:creator>
      <dc:date>2017-09-28T11:58:57Z</dc:date>
    </item>
    <item>
      <title>Re: Bigdata Continuous Delivery</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123064#M85817</link>
      <description>&lt;P&gt;Dear @Fabricio Carboni:&lt;/P&gt;&lt;P&gt;Can you please share some document on how we can implement CI/CD for pyspark based applications. Also, is it possbile to do it without using containers (like we do development in Java/Scala (first locally on windows and then build it on Linux dev/tst/Prod))&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;&lt;P&gt;Abhinav&lt;/P&gt;</description>
      <pubDate>Wed, 12 Dec 2018 22:09:23 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123064#M85817</guid>
      <dc:creator>soti_abhinav</dc:creator>
      <dc:date>2018-12-12T22:09:23Z</dc:date>
    </item>
    <item>
      <title>Re: Bigdata Continuous Delivery</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123065#M85818</link>
      <description>&lt;P&gt;Hi, Were you able to find a solution to this?  We have a similar setup and I can't seem to find any examples of that.&lt;/P&gt;</description>
      <pubDate>Fri, 08 Feb 2019 22:58:08 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Bigdata-Continuous-Delivery/m-p/123065#M85818</guid>
      <dc:creator>c-joe_boctor</dc:creator>
      <dc:date>2019-02-08T22:58:08Z</dc:date>
    </item>
  </channel>
</rss>

