Skip to content
Closed
Show file tree
Hide file tree
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Next Next commit
[SPARK-7022][PySpark][ML] Add ML.Tuning.ParamGridBuilder to PySpark
  • Loading branch information
Omede Firouz committed Apr 20, 2015
commit 8b8a6d26ed935efe91a2f36a08c9835c88605239
84 changes: 84 additions & 0 deletions python/pyspark/ml/tuning.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,84 @@
#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#

__all__ = ['ParamGridBuilder']

class ParamGridBuilder(object):
"""
Builder for a param grid used in grid search-based model selection.
"""

def __init__(self):
self._param_grid = {}

def addGrid(self, param, values):
"""
Sets the given parameters in this grid to fixed values.
"""
self._param_grid[param] = values
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Return self so this could be chained.


def baseOn(self, *args):
"""
Sets the given parameters in this grid to fixed values.
Accepts either a parameter dictionary or a list of (parameter, value) pairs.
"""
if isinstance(args[0], dict):
self.baseOn(*args[0].items())
else:
for (param, value) in args:
self.addGrid(param, [value])
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

return self.


def build(self):
"""
Builds and returns all combinations of parameters specified
by the param grid.
"""
param_maps = [{}]
for (param, values) in self._param_grid.items():

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Consider doing this

[dict(zip(self._param_grid.keys(), prod)) for prod in itertools.product(*self._param_grid.values())]

To avoid the overhead of lots of dictionary copies.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

+1 on this. Should we move self._param_grid.keys()/values() out for readability?

keys = self._param_grid.keys()
grid_values = self._param_grid.values()
return [dict(zip(keys, prod)) for prod in itertools.product(*grid_values)]

new_param_maps = []
for value in values:
for old_map in param_maps:
copied_map = old_map.copy()
copied_map[param] = value
new_param_maps.append(copied_map)
param_maps = new_param_maps

return param_maps


if __name__ == "__main__":
grid_test = ParamGridBuilder()
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Test code should be either in docstring (doctest) or in tests.py. doctest is a good choice for this one.

from classification import LogisticRegression
lr = LogisticRegression()
grid_test.addGrid(lr.regParam, [1.0, 2.0, 3.0])
grid_test.addGrid(lr.maxIter, [1, 5])
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Use builder pattern.

grid_test.addGrid(lr.featuresCol, ['f'])
grid_test.baseOn({lr.labelCol: 'l'})
grid_test.baseOn([lr.predictionCol, 'p'])
grid = grid_test.build()
expected = [
{lr.regParam: 1.0, lr.featuresCol: 'f', lr.maxIter: 1, lr.labelCol: 'l', lr.predictionCol: 'p'},
{lr.regParam: 2.0, lr.featuresCol: 'f', lr.maxIter: 1, lr.labelCol: 'l', lr.predictionCol: 'p'},
{lr.regParam: 3.0, lr.featuresCol: 'f', lr.maxIter: 1, lr.labelCol: 'l', lr.predictionCol: 'p'},
{lr.regParam: 1.0, lr.featuresCol: 'f', lr.maxIter: 5, lr.labelCol: 'l', lr.predictionCol: 'p'},
{lr.regParam: 2.0, lr.featuresCol: 'f', lr.maxIter: 5, lr.labelCol: 'l', lr.predictionCol: 'p'},
{lr.regParam: 3.0, lr.featuresCol: 'f', lr.maxIter: 5, lr.labelCol: 'l', lr.predictionCol: 'p'}
]

for a, b in zip(grid, expected):
if a != b:
exit(-1)
1 change: 1 addition & 0 deletions python/run-tests
Original file line number Diff line number Diff line change
Expand Up @@ -94,6 +94,7 @@ function run_ml_tests() {
echo "Run ml tests ..."
run_test "pyspark/ml/feature.py"
run_test "pyspark/ml/classification.py"
run_test "pyspark/ml/tuning.py"
run_test "pyspark/ml/tests.py"
}

Expand Down