Commit Graph

208 Commits

Author SHA1 Message Date
ludatabricks
49717a72dd put additionalTests to mlmetrics (#153)
The time for additionalTests is missing in MLMetrics. Now add it back to MLMetrics so that we can test the time for other methods.
2018-06-13 15:21:50 -07:00
ludatabricks
a4e1c790ba [ML-3869] Make Quantilediscretizer work with spark-2.3 (#154)
Add setOutputCol for Quantilediscretizer so that it works for spark-2.3.

The code has been manually tested by change the spark version.
2018-06-13 15:19:52 -07:00
ludatabricks
51786921a6 [ML-3583] Add benchmarks to mllib-large.yaml for featurization (#152)
Benchmark for featurization is added to mllib-large.yaml.
Cannot run QuantileDiscretizer with spark 2.3. Leave this as future work:
https://databricks.atlassian.net/browse/ML-3869
2018-06-12 17:31:30 -07:00
ludatabricks
aa1587fec5 [ML-3824] Add benchmarks to mllib-large.yaml for FPGrowth (#151)
Benchmark for FPGrowth is added to mllib-large.yaml.
2018-06-12 13:10:12 -07:00
ludatabricks
6a45dc8a2d [ML-3581] Add benchmarks to mllib-large.yaml for regression (#150)
Benchmark for regression is added to mllib-large.yaml.
DecisionTreeRegression, GLMRegression, LinearRegression, and RandomForestRegression are added.

GBT, AFTSurvivalRegression, and IsotonicRegression are missing in spark-sql-perf.
2018-06-12 10:32:02 -07:00
ludatabricks
9ab2a8bb14 [ML-3585] Added benchmarks to mllib-large.yaml for clustering (#149)
Benchmark for clustering is added to mllib-large.yaml.
GaussianMixture, KMeans, and LDA are added. BisectingKMeans is missing in spark-sql-perf now. Need to be fixed in the following up JIRA: https://databricks.atlassian.net/browse/ML-3834
Then parameters is based on the previous benchmarks for the Spark 2.2 QA.
2018-06-08 12:06:52 -07:00
ludatabricks
62b173d779 Output Training Time as metrics (#148)
* change the structure of mlresult and add isLargerBetter

* output training time, not scoreTrainTime
2018-06-07 13:21:32 -07:00
ludatabricks
d9984e1c0a [ML-3584] Added benchmarks to mllib-large.yaml for ALS (#147)
Benchmark for ALS is added to mllib-large.yaml.
Then parameters is based on the previous benchmarks for the Spark 2.2 QA. It has been tested for the stability under the same cluster setting for other benchmarks (classifcation).
2018-06-07 08:11:37 -07:00
ludatabricks
93626c11b4 [ML-3775] Add "benchmarkId" to BenchmarkResult (#146)
Add "benchmarkId" to BenchmarkResult, which is based on the benchmark name and a hashed value of params.
2018-06-04 14:13:45 -07:00
ludatabricks
f1139fc742 [ML-3753] Log "value" instead of "Some(value)" for ML params in results (#145)
Change the function MLParams.toMap so it will not output a option value in the map.
We will not get option value in the params in the output result.

* change the "option" number to direct number
* update based on the comments
2018-06-04 11:09:41 -07:00
ludatabricks
1768d376f9 [ML-3749] Log metric name and isLargerBetter in BenchmarkResult (#144)
* Add new case class "MLMetric" to save all different metrics
* Change "mlResult" in BenchmarkResult to Array[MLMetric]
* score function will return MLMetric
2018-06-01 15:49:16 -07:00
Bago Amirbekian
789a0f5b8b Added benchmarks to mllib-large.yaml for classifcation Estimators. (#143) 2018-05-30 08:18:49 -07:00
WeichenXu
3786a8391e Quantile discretizer benchmark (#135)
QuantileDiscretizer benchmark
2018-05-17 11:55:00 -07:00
Bago Amirbekian
15d9283473 Run mllib small in unit tests (#141)
* Add mllib unit test to run mllib-small.yaml.

* Check results in unit tests and fail tests if failures are present.

* Rename to be MLLibSuite instead of MLLibTest.

* Better error message on failed benchmarks.
2018-05-09 16:24:30 -07:00
Bago Amirbekian
9ece11ff20 Add decision tree benchmark (#140)
* Move mllib config file to resources.

* Add DecisionTreeClassification as first benchmark in mllib-large.yaml.

* Read config files as streams to be jar compatible.

* PR feedback #140.
2018-05-08 21:44:11 -07:00
Joseph Bradley
ed9bbb01a5 fix bug with ML additional method tests (#142)
#139 introduced a bug which made most ML tests fail with mllib-small.yaml. This fixes those tests.
2018-05-08 13:23:22 -07:00
WeichenXu
be4459fe41 Additional method test for some ML algos (#139)
Add additional method test for some ML algos.

In this PR, I add `associationRules` in `FPGrowth` and `findSynonyms`. 
After the design is accepted, I will add other methods later.

Add an interface in `BenchmarkableAlgorithm`:
```
  def testAdditionalMethods(ctx: MLBenchContext, model: Transformer): Map[String, () => _]
```
2018-05-02 13:45:58 -07:00
WeichenXu
5af9f6dfc2 Word2Vec benchmark (#127)
* init pr

* update

* use builtin split fun
2018-03-15 13:10:04 -07:00
Juliusz Sompolski
a8acd53fdd
Use DECIMAL and DATE in the default TPCDS notebooks. (#130)
A long time ago, DECIMAL was substituted by DOUBLE and DATE by STRING to workaround some problems.
There is no reason to do it anymore.
2018-03-07 21:44:42 +01:00
Juliusz Sompolski
b7ac7e55ae
Remove VACUUM from tpcds_datagen notebook. (#129) 2018-03-07 15:36:27 +01:00
WeichenXu
93a34553f0 MinHashLSH and BucketedRandomProjectionLSH benchmark #128
MinHashLSH and BucketedRandomProjectionLSH benchmark added.

Future questions:
* Whether we need to improve the way of testing data generation for MinHashLSH ( and add more control param, such as max/min element number in each input set )
* Whether we need to add benchmark for approxNearestNeighbors and approxSimilarityJoin
2018-03-02 15:21:37 -08:00
Bago Amirbekian
6d01ac94a1 [ML-3342] Bug fixes to make mllib benchmarks work with dbr-4.0. (#125)
In spark 2.3 some default param values were moved from Models to matching Estimators. I added explicit sets for these values in our tests to avoid errors. Also renamed ModelBuilder to ModelBuildersSSP to avoid a name conflict with dbml-local which is included in databricks runtime.
2018-03-02 09:12:38 -08:00
Juliusz Sompolski
91604a3ab0 Update README to specify that TPCDS kit needs to be installed on all nodes. 2018-02-27 12:06:12 +01:00
Juliusz Sompolski
31f34beee5
Update README to do sql("use database") (#123) 2017-11-07 20:38:26 +01:00
Juliusz Sompolski
7bf2d45b0f Don't clean blocks after every run in Benchmarkable (#119)
Do not clean blocks after each run in the generic Benchmarkable trait.
It seems to have been there since #33, and an option spark.databricks.benchmark.cleanBlocksAfter to turn it off was added to it in #98, specifically to allow parallel TPCDS streams to not wipe each other's blocks. But that option is quite well hidden and obscure, and as a SparkContext config option can only be set during cluster creation, so it's not friendly to use.

Cleaning up the blocks doesn't seem necessary for the Query Benchmarkables used for TPCDS and TPCH. Remove it from there, and leave it only for MLPipelineStageBenchmarkable.
2017-09-18 11:51:12 +02:00
Juliusz Sompolski
fdd0e38717 TPCDS notebooks in source, not binary format (#121) 2017-09-13 14:57:59 +02:00
Nico Poggi
006f096562 Merge pull request #120 from juliuszsompolski/tpcds_notebooks
Add example notebooks for running TPCDS and update readme
2017-09-12 17:22:38 +02:00
Juliusz Sompolski
5ebb9cfb12 add some more comments 2017-09-12 16:51:26 +02:00
Juliusz Sompolski
c78f2b3a9b update readme 2017-09-12 16:40:23 +02:00
Juliusz Sompolski
ae8bcdb292 add notebooks 2017-09-12 15:43:08 +02:00
WeichenXu
f08bf31d18 add benchmark for FPGrowth (#113)
Note: 
Add a `ItemSetGenerator` class, use following algo:

1.Create P=`numItems` items (integers 0 to P-1)
2. Generate `numExample` rows, where each row (an itemset) is selected as follows:
  2.1 Choose the size of the itemset from a Poisson distribution
  2.2 Generate `size - 2` items by choosing integers from a Poisson distribution. Eliminate duplicates as needed.
  2.3 Add 2 new items in order to create actual association rules.
    2.3.1 For each itemset, pick the first item, and compute a new item = (firstItem + P / 2) % P, add new item to the set.
    2.3.2 For each itemset, pick the first 2 items (integers) and add them together (modulo P) to compute a new item to add to the set.
2017-09-04 10:48:05 -07:00
Juliusz Sompolski
bcda8fc1e5 Coalesce non-partitioned tables. (#118)
In #109 coalescing of non-partitioned tables into 1 file seems to have gotten accidentally removed.
Put it back, but only when clusterByPartitionedColumns == true
Considering that we coalesce partitions only when that setting is true, it seems to be consistent to use it also for non-partitioned tables.

It may be better to change the name of the parameter, but that changes the interface, and possibly should be left for some future clean up.
2017-09-04 18:05:42 +02:00
Siddharth Murching
3e1bbd00ed [ML-2847] Add new tests for (DecisionTree, RandomForest)Regression, GMM, HashingTF (#116)
This PR follows up on #112, adding new performance tests for DecisionTreeRegression, RandomForestRegression, GMM, and HashingTF.

Summary of changes:
* Added new performance tests
* Updated configs in mllib-small.yaml
** Alphabetized configs
** Added new configs for: RandomForestRegression, DecisionTreeRegression, GMM, HashingTF
* Refactored TreeOrForestClassification into a trait (TreeOrForestEstimator) exposing methods for all tree/forest estimator performance tests.
** Copied code from DecisionTreeClassification.scala into TreeOrForestEstimator.scala

I tested this PR by running the performance tests specified in mllib-small.yaml
2017-09-03 22:26:20 -07:00
WeichenXu
19c41464c7 fix df.drop in VectorAssembler (#117)
fix df.drop in VectorAssembler to return correct DataFrame
2017-09-01 13:51:05 -07:00
WeichenXu
6ec83fd0f7 Add benchmark for LinearSVC/OnehotEncoder/VectorSlicer/VectorAssembler/StringIndexer/Tokenizer (#112)
Add benchmark for:

LinearSVC
OnehotEncoder
VectorSlicer
VectorAssembler
StringIndexer
Tokenizer
2017-08-31 13:56:43 -07:00
Juliusz Sompolski
737a1bc355 BlockingLineStream (#115)
## What changes are proposed in this pull request?

Investigating OOMs during TPCDS data generation:
it turned out that scala standard library's ProcessBuilder.lineStream would by default create a LinkedBlockingQueue buffer of Integer.MAX_VALUE capacity.
It surfaced after in https://github.com/databricks/tpcds-kit/pull/2 we implemented 10x improvements to dsdgen speed.
Now spark-sql-perf does not keep up with ingesting data from dsdgen, and the buffer will cause OOMs.

Pulled out pieces of ProcessBuilderImpl and ProcessImpl just to create a LinkedBlockingQueue with maxQueueSize=65536 instead.

Also submitted https://github.com/scala/scala/pull/6052

## How was this patch tested?

- ssh on the worker - see that dsdgen is being throttled now, Java memory doesn't explode.
- tested that TPCDS SF100 generated correctly.
2017-08-31 15:16:22 +02:00
Siddharth Murching
9febc34f66 Refactor MLParams for spark-sql-perf (#114)
A case class (MLParams) is currently used to store/access parameters for ML tests in spark-sql-perf. With the addition of new ML tests to spark-sql-perf (in this PR: #112), the number of ML-related test params will be > 22, but Scala only allows up to 22 params in a case class.

This PR addresses the issue by:
* Introducing a new MLParameters class (class MLParameters) that provides access to the same parameters as MLParams, except as a class instead of a case class.
* Replacing usages of MLParams with MLParameters
* Storing the members of MLParameters in BenchmarkResult.parameters for logging/persistence.

Tested by running default performance tests in src/main/scala/configs/mllib-small.yaml.
2017-08-28 13:23:59 -07:00
Siddharth Murching
d0de5ae8aa Update tests to run with Spark 2.2, add NaiveBayes & Bucketizer ML tests (#110)
* Made small updates in Benchmark.scala and Query.scala for Spark 2.2
* Added tests for NaiveBayesModel and Bucketizer
* Changed BenchmarkAlgorithm.getEstimator() -> BenchmarkAlgorithm.getPipelineStage() to allow for the benchmarking of Estimators and Transformers instead of just Estimators

Commits:
* Changes made so that spark-sql-perf compiles with Spark 2.2

* Updates for running ML tests from the command line + added Naive Bayes test

* Add Bucketizer test as example of Featurizer test; change getEstimator() to getPipelineStage() in
BenchmarkAlgorithm to allow for testing of transformers in addition to estimators.

* Add comment for main method in MLlib.scala

* Rename MLTransformerBenchmarkable --> MLPipelineStageBenchmarkable, fix issue with NaiveBayes param

* Add UnaryTransformer trait for common data/methods to be shared across all objects
testing featurizers that operate on a single column (StringIndexer, OneHotEncoder, Bucketizer, HashingTF, etc)

* Respond to review comments:

* bin/run-ml: Add newline at EOF
* Query.scala: organized imports
* MLlib.scala: organized imports, fixed SparkContext initialization
* NaiveBayes.scala: removed unused temp val, improved probability calculation in trueModel()
* Bucketizer.scala: use DataGenerator.generateContinuousFeatures instead of generating data on the driver

* Fix bug in Bucketizer.scala

* Precompute log of sum of unnormalized probabilities in NaiveBayes.scala, add NaiveBayes and Bucketizer tests to mllib-small.yaml

* Update Query.scala to use p() to access SparkPlans under a given SparkPlan

* Update README to indicate that spark-sql-perf only works with Spark 2.2+ after this PR
2017-08-21 15:07:46 -07:00
Yin Huai
b3a6ed79b3 Start the development 0.5.0-SNAPSHOT 2017-08-21 14:21:19 -07:00
Bogdan Raducanu
4e7a2363b9 Support for TPC-H benchmark
Refactored TPC-DS code to be able to reuse it for TPC-H.
Added TPC-H queries texts adapted for Spark.
2017-08-09 12:26:32 +02:00
Kevin
fdcde7595c Update README (#107)
Little update for the README
2017-07-13 10:45:24 +02:00
Juliusz Sompolski
6488d74d23 tpcds_2_4: Add alias names to subqueries in FROM clause.
## What changes were proposed in this pull request?

Since SPARK-20690 and SPARK-20916 Spark requires all subqueries in FROM clause to have an alias name.

## How was this patch tested?

Tested on SF1.
2017-06-29 16:04:08 +02:00
Juliusz Sompolski
bff6b34f62 Tweaks and improvements (#106)
Data generation:
* Add an option to change Dates to Strings, and specify it in Tables object creator.
* Add discovering partitions to createExternalTables
* Add analyzeTables function that gathers statistics.

Benchmark execution:
* Perform collect() on Dataframe, so that it is recorded by SQL SparkUI.
2017-06-13 11:42:14 +02:00
Juliusz Sompolski
75f3876e59 Merge pull request #103 from juliuszsompolski/fixtypes
Correct types of keys in schema
2017-05-26 11:53:19 +02:00
Juliusz Sompolski
2ddd521ab5 ok, make it long only where really needed. 2017-05-26 10:36:40 +02:00
Juliusz Sompolski
1bca964a3d Correct types of keys 2017-05-25 17:12:47 +02:00
Volodymyr Lyubinets
beec62844d Merge pull request #101 from vlyubin/master
Add tpcds 2.4 queries
2017-05-16 10:35:35 +02:00
vlyubin
c0bd21c2ec Add ss_max 2017-05-16 10:29:00 +02:00
vlyubin
e5dc6f338f Updated queries 23 2017-05-15 17:30:20 +02:00
vlyubin
e8f85b0b0e Moved queries into a separate folder 2017-05-15 14:22:37 +02:00