-
Notifications
You must be signed in to change notification settings - Fork 29k
[SPARK-20114][ML] spark.ml parity for sequential pattern mining - PrefixSpan #20973
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from 3 commits
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,96 @@ | ||
| /* | ||
| * Licensed to the Apache Software Foundation (ASF) under one or more | ||
| * contributor license agreements. See the NOTICE file distributed with | ||
| * this work for additional information regarding copyright ownership. | ||
| * The ASF licenses this file to You under the Apache License, Version 2.0 | ||
| * (the "License"); you may not use this file except in compliance with | ||
| * the License. You may obtain a copy of the License at | ||
| * | ||
| * http://www.apache.org/licenses/LICENSE-2.0 | ||
| * | ||
| * Unless required by applicable law or agreed to in writing, software | ||
| * distributed under the License is distributed on an "AS IS" BASIS, | ||
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
| * See the License for the specific language governing permissions and | ||
| * limitations under the License. | ||
| */ | ||
|
|
||
| package org.apache.spark.ml.fpm | ||
|
|
||
| import org.apache.spark.annotation.{Experimental, Since} | ||
| import org.apache.spark.mllib.fpm.{PrefixSpan => mllibPrefixSpan} | ||
| import org.apache.spark.sql.{DataFrame, Dataset, Row} | ||
| import org.apache.spark.sql.functions.col | ||
| import org.apache.spark.sql.types.{ArrayType, LongType, StructField, StructType} | ||
|
|
||
| /** | ||
| * :: Experimental :: | ||
| * A parallel PrefixSpan algorithm to mine frequent sequential patterns. | ||
| * The PrefixSpan algorithm is described in J. Pei, et al., PrefixSpan: Mining Sequential Patterns | ||
| * Efficiently by Prefix-Projected Pattern Growth | ||
| * (see <a href="http://doi.org/10.1109/ICDE.2001.914830">here</a>). | ||
| * | ||
| * @see <a href="https://en.wikipedia.org/wiki/Sequential_Pattern_Mining">Sequential Pattern Mining | ||
| * (Wikipedia)</a> | ||
| */ | ||
| @Since("2.4.0") | ||
| @Experimental | ||
| object PrefixSpan { | ||
|
|
||
| /** | ||
| * :: Experimental :: | ||
| * Finds the complete set of frequent sequential patterns in the input sequences of itemsets. | ||
| * | ||
| * @param dataset A dataset or a dataframe containing a sequence column which is | ||
| * {{{Seq[Seq[_]]}}} type | ||
| * @param sequenceCol the name of the sequence column in dataset, rows with nulls in this column | ||
| * are ignored | ||
| * @param minSupport the minimal support level of the sequential pattern, any pattern that | ||
| * appears more than (minSupport * size-of-the-dataset) times will be output | ||
| * (recommended value: `0.1`). | ||
| * @param maxPatternLength the maximal length of the sequential pattern | ||
| * (recommended value: `10`). | ||
| * @param maxLocalProjDBSize The maximum number of items (including delimiters used in the | ||
| * internal storage format) allowed in a projected database before | ||
| * local processing. If a projected database exceeds this size, another | ||
| * iteration of distributed prefix growth is run | ||
| * (recommended value: `32000000`). | ||
| * @return A `DataFrame` that contains columns of sequence and corresponding frequency. | ||
| * The schema of it will be: | ||
| * - `sequence: Seq[Seq[T]]` (T is the item type) | ||
| * - `frequency: Long` | ||
| */ | ||
| @Since("2.4.0") | ||
| def findFrequentSequentialPatterns( | ||
| dataset: Dataset[_], | ||
| sequenceCol: String, | ||
|
Contributor
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. @WeichenXu123 @jkbradley The static method doesn't scale with parameters. If we add a new param, we have to keep the old one for binary compatibility. Why not using setters? I think we only need to avoid using
Contributor
Author
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I agree with using setters. @jkbradley What do you think of it ?
Member
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I agree in general, but I don’t think it’s a big deal for PrefixSpan. I think of our current static method as a temporary workaround until we do the work to build a Model which can make meaningful predictions. This will mean that further PrefixSpan improvements may be blocked on this Model work, but I think that’s OK since predictions should be the next priority for PrefixSpan. Once we have a Model, I recommend we deprecate the current static method. I'm also OK with changing this to use setters, but then we should name it something else so that we can replace it with an Estimator + Model pair later on. I'd suggest "PrefixSpanBuilder."
Contributor
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. It should be easier to keep the final class PrefixSpan(override val uid: String) extends Params {
// param, setters, getters
def findFrequentSequentialPatterns(dataset: Dataset[_]): DataFrame
}Later we can add
Contributor
Author
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. this way
Contributor
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Adding
Member
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Oh, I think you're right @mengxr . That approach sounds good.
Member
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. @WeichenXu123 Do you have time to send a PR to update this API?
Contributor
Author
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Sure. Will update soon! |
||
| minSupport: Double, | ||
| maxPatternLength: Int, | ||
| maxLocalProjDBSize: Long): DataFrame = { | ||
|
|
||
| val inputType = dataset.schema(sequenceCol).dataType | ||
| require(inputType.isInstanceOf[ArrayType] && | ||
| inputType.asInstanceOf[ArrayType].elementType.isInstanceOf[ArrayType], | ||
| s"The input column must be ArrayType and the array element type must also be ArrayType, " + | ||
| s"but got $inputType.") | ||
|
|
||
|
|
||
| val data = dataset.select(sequenceCol) | ||
|
||
| val sequences = data.where(col(sequenceCol).isNotNull).rdd | ||
| .map(r => r.getAs[Seq[Seq[Any]]](0).map(_.toArray).toArray) | ||
|
|
||
| val mllibPrefixSpan = new mllibPrefixSpan() | ||
| .setMinSupport(minSupport) | ||
| .setMaxPatternLength(maxPatternLength) | ||
| .setMaxLocalProjDBSize(maxLocalProjDBSize) | ||
|
|
||
| val rows = mllibPrefixSpan.run(sequences).freqSequences.map(f => Row(f.sequence, f.freq)) | ||
| val schema = StructType(Seq( | ||
| StructField("sequence", dataset.schema(sequenceCol).dataType, nullable = false), | ||
| StructField("frequency", LongType, nullable = false))) | ||
| val freqSequences = dataset.sparkSession.createDataFrame(rows, schema) | ||
|
|
||
| freqSequences | ||
| } | ||
|
|
||
| } | ||
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,136 @@ | ||
| /* | ||
| * Licensed to the Apache Software Foundation (ASF) under one or more | ||
| * contributor license agreements. See the NOTICE file distributed with | ||
| * this work for additional information regarding copyright ownership. | ||
| * The ASF licenses this file to You under the Apache License, Version 2.0 | ||
| * (the "License"); you may not use this file except in compliance with | ||
| * the License. You may obtain a copy of the License at | ||
| * | ||
| * http://www.apache.org/licenses/LICENSE-2.0 | ||
| * | ||
| * Unless required by applicable law or agreed to in writing, software | ||
| * distributed under the License is distributed on an "AS IS" BASIS, | ||
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
| * See the License for the specific language governing permissions and | ||
| * limitations under the License. | ||
| */ | ||
| package org.apache.spark.ml.fpm | ||
|
|
||
| import org.apache.spark.ml.util.MLTest | ||
| import org.apache.spark.sql.DataFrame | ||
|
|
||
| class PrefixSpanSuite extends MLTest { | ||
|
|
||
| import testImplicits._ | ||
|
|
||
| override def beforeAll(): Unit = { | ||
| super.beforeAll() | ||
| } | ||
|
|
||
| test("PrefixSpan projections with multiple partial starts") { | ||
| val smallDataset = Seq(Seq(Seq(1, 2), Seq(1, 2, 3))).toDF("sequence") | ||
| val result = PrefixSpan.findFrequentSequentialPatterns(smallDataset, "sequence", | ||
| minSupport = 1.0, maxPatternLength = 2, maxLocalProjDBSize = 32000000) | ||
| .as[(Seq[Seq[Int]], Long)].collect() | ||
| val expected = Array( | ||
| (Seq(Seq(1)), 1L), | ||
| (Seq(Seq(1, 2)), 1L), | ||
| (Seq(Seq(1), Seq(1)), 1L), | ||
| (Seq(Seq(1), Seq(2)), 1L), | ||
| (Seq(Seq(1), Seq(3)), 1L), | ||
| (Seq(Seq(1, 3)), 1L), | ||
| (Seq(Seq(2)), 1L), | ||
| (Seq(Seq(2, 3)), 1L), | ||
| (Seq(Seq(2), Seq(1)), 1L), | ||
| (Seq(Seq(2), Seq(2)), 1L), | ||
| (Seq(Seq(2), Seq(3)), 1L), | ||
| (Seq(Seq(3)), 1L)) | ||
| compareResults[Int](expected, result) | ||
| } | ||
|
|
||
| /* | ||
| To verify expected results for `smallTestData`, create file "prefixSpanSeqs2" with content | ||
| (format = (transactionID, idxInTransaction, numItemsinItemset, itemset)): | ||
| 1 1 2 1 2 | ||
| 1 2 1 3 | ||
| 2 1 1 1 | ||
| 2 2 2 3 2 | ||
| 2 3 2 1 2 | ||
| 3 1 2 1 2 | ||
| 3 2 1 5 | ||
| 4 1 1 6 | ||
| In R, run: | ||
| library("arulesSequences") | ||
| prefixSpanSeqs = read_baskets("prefixSpanSeqs", info = c("sequenceID","eventID","SIZE")) | ||
| freqItemSeq = cspade(prefixSpanSeqs, | ||
| parameter = 0.5, maxlen = 5 )) | ||
| resSeq = as(freqItemSeq, "data.frame") | ||
| resSeq | ||
|
|
||
| sequence support | ||
| 1 <{1}> 0.75 | ||
| 2 <{2}> 0.75 | ||
| 3 <{3}> 0.50 | ||
| 4 <{1},{3}> 0.50 | ||
| 5 <{1,2}> 0.75 | ||
| */ | ||
| val smallTestData = Seq( | ||
| Seq(Seq(1, 2), Seq(3)), | ||
| Seq(Seq(1), Seq(3, 2), Seq(1, 2)), | ||
| Seq(Seq(1, 2), Seq(5)), | ||
| Seq(Seq(6))) | ||
|
|
||
| val smallTestDataExpectedResult = Array( | ||
| (Seq(Seq(1)), 3L), | ||
| (Seq(Seq(2)), 3L), | ||
| (Seq(Seq(3)), 2L), | ||
| (Seq(Seq(1), Seq(3)), 2L), | ||
| (Seq(Seq(1, 2)), 3L) | ||
| ) | ||
|
|
||
| test("PrefixSpan Integer type, variable-size itemsets") { | ||
| val df = smallTestData.toDF("sequence") | ||
| val result = PrefixSpan.findFrequentSequentialPatterns(df, "sequence", | ||
| minSupport = 0.5, maxPatternLength = 5, maxLocalProjDBSize = 32000000) | ||
| .as[(Seq[Seq[Int]], Long)].collect() | ||
|
|
||
| compareResults[Int](smallTestDataExpectedResult, result) | ||
| } | ||
|
|
||
| test("PrefixSpan input row with nulls") { | ||
| val df = (smallTestData :+ null).toDF("sequence") | ||
| val result = PrefixSpan.findFrequentSequentialPatterns(df, "sequence", | ||
| minSupport = 0.5, maxPatternLength = 5, maxLocalProjDBSize = 32000000) | ||
| .as[(Seq[Seq[Int]], Long)].collect() | ||
|
|
||
| compareResults[Int](smallTestDataExpectedResult, result) | ||
| } | ||
|
|
||
| test("PrefixSpan String type, variable-size itemsets") { | ||
| val intToString = (1 to 6).zip(Seq("a", "b", "c", "d", "e", "f")).toMap | ||
| val df = smallTestData | ||
| .map(seq => seq.map(itemSet => itemSet.map(intToString))) | ||
| .toDF("sequence") | ||
| val result = PrefixSpan.findFrequentSequentialPatterns(df, "sequence", | ||
| minSupport = 0.5, maxPatternLength = 5, maxLocalProjDBSize = 32000000) | ||
| .as[(Seq[Seq[String]], Long)].collect() | ||
|
|
||
| val expected = smallTestDataExpectedResult.map { case (seq, freq) => | ||
| (seq.map(itemSet => itemSet.map(intToString)), freq) | ||
| } | ||
| compareResults[String](expected, result) | ||
| } | ||
|
|
||
| private def compareResults[Item]( | ||
| expectedValue: Array[(Seq[Seq[Item]], Long)], | ||
| actualValue: Array[(Seq[Seq[Item]], Long)]): Unit = { | ||
| val expectedSet = expectedValue.map { x => | ||
| (x._1.map(_.toSet), x._2) | ||
| }.toSet | ||
| val actualSet = actualValue.map { x => | ||
| (x._1.map(_.toSet), x._2) | ||
| }.toSet | ||
| assert(expectedSet === actualSet) | ||
| } | ||
| } | ||
|
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I had asked for this change to "frequency" from "freq," but I belatedly realized that this conflicts with the existing FPGrowth API, which uses "freq." It would be best to maintain consistency. Would you mind reverting to "freq?"
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
sure!