<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Spark MLBase (MLI and ML Optimizer) status in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-MLBase-MLI-and-ML-Optimizer-status/m-p/96783#M10327</link>
    <description>&lt;P&gt;As a complement to Matt Foley's answer: concerning MLOptimizer, I think they were either meaning generic optimization algorithms such as Gradient Descent, available in mllib.optimization package (see &lt;A rel="noopener noreferrer noopener noreferrer" href="https://spark.apache.org/docs/2.3.0/mllib-optimization.html" target="_blank"&gt;https://spark.apache.org/docs/2.3.0/mllib-optimization.html&lt;/A&gt;), or they were meaning ML algorithm hyper-parameter optimization. Hyper-parameter tuning using e.g. cross-validation and grid-search is available in the Spark ML tuning package (see &lt;A rel="noopener noreferrer noopener noreferrer" href="https://spark.apache.org/docs/2.2.0/ml-tuning.html" target="_blank"&gt;https://spark.apache.org/docs/2.2.0/ml-tuning.html&lt;/A&gt;).&lt;/P&gt;&lt;P&gt;However, if they were meaning automatic hyper-parameter optimization using for example Bayesian optimization, then I would like to know more about it...&lt;/P&gt;</description>
    <pubDate>Tue, 25 Jun 2019 00:26:49 GMT</pubDate>
    <dc:creator>forest</dc:creator>
    <dc:date>2019-06-25T00:26:49Z</dc:date>
    <item>
      <title>Spark MLBase (MLI and ML Optimizer) status</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-MLBase-MLI-and-ML-Optimizer-status/m-p/96781#M10325</link>
      <description>&lt;P&gt;Does anyone know the status of project MLBase:&lt;/P&gt;&lt;P&gt;"Implementing and consuming Machine Learning at scale are difficult tasks. MLbase is a platform addressing both issues, and consists of three components -- &lt;EM&gt;MLlib, MLI, ML Optimizer&lt;/EM&gt;."&lt;/P&gt;&lt;P&gt;It seems that project and related information are not being updated for last 2 years:&lt;/P&gt;&lt;P&gt;&lt;A target="_blank" href="http://www.mlbase.org/"&gt;http://www.mlbase.org/&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;A target="_blank" href="http://www.cs.berkeley.edu/~ameet/mlbase_website/mlbase_website/download.html"&gt;http://www.cs.berkeley.edu/~ameet/mlbase_website/mlbase_website/download.html&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;A target="_blank" href="https://github.com/amplab/MLI"&gt;https://github.com/amplab/MLI&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;A target="_blank" href="http://ampcamp.berkeley.edu/wp-content/uploads/2013/07/amp_camp_8_30_13-1.pdf"&gt;http://ampcamp.berkeley.edu/wp-content/uploads/2013/07/amp_camp_8_30_13-1.pdf&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 21 Apr 2026 13:30:51 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-MLBase-MLI-and-ML-Optimizer-status/m-p/96781#M10325</guid>
      <dc:creator>gbraccialli3</dc:creator>
      <dc:date>2026-04-21T13:30:51Z</dc:date>
    </item>
    <item>
      <title>Re: Spark MLBase (MLI and ML Optimizer) status</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-MLBase-MLI-and-ML-Optimizer-status/m-p/96782#M10326</link>
      <description>&lt;P&gt;I don't know from personal involvement, but it may be that all useful parts of the MLBase project have been absorbed into Spark ML, and no one chose to continue MLBase as a separate project. The MLBase project page itself says that MLlib is just the Spark project's MLlib, see &lt;A href="https://github.com/apache/spark/tree/master/mllib" target="_blank"&gt;https://github.com/apache/spark/tree/master/mllib&lt;/A&gt;&lt;/P&gt;&lt;P&gt;The MLBase project page also says, "Many features in &lt;I&gt;MLlib&lt;/I&gt; have been borrowed from &lt;I&gt;ML Optimizer&lt;/I&gt; and &lt;I&gt;MLI."&lt;/I&gt;  That suggests that there was already a process of absorption happening in 2013, and perhaps after that there was insufficient motivation to continue developing ML Optimizer and MLI as separate components.&lt;/P&gt;&lt;P&gt;In support of this idea, it appears that &lt;A href="https://github.com/amplab/MLI/tree/master/src/main/scala/ml" target="_blank"&gt;https://github.com/amplab/MLI/tree/master/src/main/scala/ml&lt;/A&gt; is a subset of the contents of &lt;A href="https://github.com/apache/spark/tree/master/mllib/src/main/scala/org/apache/spark/ml" target="_blank"&gt;https://github.com/apache/spark/tree/master/mllib/src/main/scala/org/apache/spark/ml&lt;/A&gt;&lt;/P&gt;&lt;P&gt;In my brief effort I was not able to similarly track down remnants of the "ML Optimizer" code, but certainly there are optimizers throughout the Spark ML code, and they tend to be algorithm-specific, so there wouldn't be much motivation for grouping them into a discrete component.&lt;/P&gt;&lt;P&gt;Hope this helps.&lt;/P&gt;</description>
      <pubDate>Wed, 11 Nov 2015 02:19:25 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-MLBase-MLI-and-ML-Optimizer-status/m-p/96782#M10326</guid>
      <dc:creator>mfoley</dc:creator>
      <dc:date>2015-11-11T02:19:25Z</dc:date>
    </item>
    <item>
      <title>Re: Spark MLBase (MLI and ML Optimizer) status</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-MLBase-MLI-and-ML-Optimizer-status/m-p/96783#M10327</link>
      <description>&lt;P&gt;As a complement to Matt Foley's answer: concerning MLOptimizer, I think they were either meaning generic optimization algorithms such as Gradient Descent, available in mllib.optimization package (see &lt;A rel="noopener noreferrer noopener noreferrer" href="https://spark.apache.org/docs/2.3.0/mllib-optimization.html" target="_blank"&gt;https://spark.apache.org/docs/2.3.0/mllib-optimization.html&lt;/A&gt;), or they were meaning ML algorithm hyper-parameter optimization. Hyper-parameter tuning using e.g. cross-validation and grid-search is available in the Spark ML tuning package (see &lt;A rel="noopener noreferrer noopener noreferrer" href="https://spark.apache.org/docs/2.2.0/ml-tuning.html" target="_blank"&gt;https://spark.apache.org/docs/2.2.0/ml-tuning.html&lt;/A&gt;).&lt;/P&gt;&lt;P&gt;However, if they were meaning automatic hyper-parameter optimization using for example Bayesian optimization, then I would like to know more about it...&lt;/P&gt;</description>
      <pubDate>Tue, 25 Jun 2019 00:26:49 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-MLBase-MLI-and-ML-Optimizer-status/m-p/96783#M10327</guid>
      <dc:creator>forest</dc:creator>
      <dc:date>2019-06-25T00:26:49Z</dc:date>
    </item>
  </channel>
</rss>

