From 54fbc51bff0b01799edc55198a6d530eac3b84ab Mon Sep 17 00:00:00 2001 From: Phillip Cloud Date: Thu, 9 Jul 2015 08:26:52 -0500 Subject: [PATCH 1/5] Fix manifest --- MANIFEST.in | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/MANIFEST.in b/MANIFEST.in index a44c426c2..dc49d69b5 100644 --- a/MANIFEST.in +++ b/MANIFEST.in @@ -4,7 +4,7 @@ recursive-include docs *.rst include requirements.txt include requirements-strict.txt include setup.py -include README.md +include README.rst include LICENSE.txt include MANIFEST.in From f15e3337371083eb26e2c1ba9ac58ccc88d27c46 Mon Sep 17 00:00:00 2001 From: Phillip Cloud Date: Thu, 9 Jul 2015 08:27:02 -0500 Subject: [PATCH 2/5] Materialize the list of xfails --- blaze/compute/tests/test_sparksql.py | 6 ++---- 1 file changed, 2 insertions(+), 4 deletions(-) diff --git a/blaze/compute/tests/test_sparksql.py b/blaze/compute/tests/test_sparksql.py index 633dadd4d..da67117be 100644 --- a/blaze/compute/tests/test_sparksql.py +++ b/blaze/compute/tests/test_sparksql.py @@ -10,8 +10,6 @@ import itertools import shutil -from distutils.version import StrictVersion - from py4j.protocol import Py4JJavaError import numpy as np import pandas as pd @@ -250,8 +248,8 @@ def test_column_arithmetic(ctx, db): 'supported in HiveContext')) -@pytest.mark.parametrize('func', map(fail_on_spark_one_two, - [sin, cos, tan, exp])) +@pytest.mark.parametrize('func', list(map(fail_on_spark_one_two, + [sin, cos, tan, exp]))) def test_math(ctx, db, func): expr = func(db.t.amount) result = compute(expr, ctx) From 42635aa583af335a622ea2c992b8940bb7a22917 Mon Sep 17 00:00:00 2001 From: Phillip Cloud Date: Thu, 9 Jul 2015 08:27:17 -0500 Subject: [PATCH 3/5] Conda build test deps --- all.recipe/meta.yaml | 3 +++ core.recipe/meta.yaml | 3 +++ 2 files changed, 6 insertions(+) diff --git a/all.recipe/meta.yaml b/all.recipe/meta.yaml index 3e094d485..508910973 100644 --- a/all.recipe/meta.yaml +++ b/all.recipe/meta.yaml @@ -30,6 +30,9 @@ test: - dask - dynd-python - pytables + - yaml + - pyyaml + - h5py commands: - py.test -vx --doctest-modules --doctest-ignore-import-errors --pyargs blaze -rsxX diff --git a/core.recipe/meta.yaml b/core.recipe/meta.yaml index 78da86733..00716c15d 100644 --- a/core.recipe/meta.yaml +++ b/core.recipe/meta.yaml @@ -36,6 +36,9 @@ test: - dask - dynd-python - pytables + - yaml + - pyyaml + - h5py commands: - py.test -vx --doctest-modules --doctest-ignore-import-errors --pyargs blaze -rsxX From 25a45d907795f19d83fc7365c9b3bdf8709e914a Mon Sep 17 00:00:00 2001 From: Phillip Cloud Date: Thu, 9 Jul 2015 08:38:34 -0500 Subject: [PATCH 4/5] Explicit xfailing --- blaze/compute/tests/test_sparksql.py | 11 +++++++---- 1 file changed, 7 insertions(+), 4 deletions(-) diff --git a/blaze/compute/tests/test_sparksql.py b/blaze/compute/tests/test_sparksql.py index da67117be..6ac126111 100644 --- a/blaze/compute/tests/test_sparksql.py +++ b/blaze/compute/tests/test_sparksql.py @@ -242,10 +242,13 @@ def test_column_arithmetic(ctx, db): # pyspark doesn't use __version__ so we use this kludge # should submit a bug report upstream to get __version__ -fail_on_spark_one_two = pytest.mark.xfail(not hasattr(pyspark.sql, 'types'), - raises=py4j.protocol.Py4JJavaError, - reason=('math functions only ' - 'supported in HiveContext')) +def fail_on_spark_one_two(x): + if hasattr(pyspark.sql, 'types'): + return x + else: + return pytest.mark.xfail(x, raises=py4j.protocol.Py4JJavaError, + reason=('math functions only supported in ' + 'HiveContext')) @pytest.mark.parametrize('func', list(map(fail_on_spark_one_two, From f046024f4874ba35ab469b90315a19cf5851961f Mon Sep 17 00:00:00 2001 From: Phillip Cloud Date: Thu, 9 Jul 2015 08:58:24 -0500 Subject: [PATCH 5/5] Use setuptools --- setup.py | 1 + 1 file changed, 1 insertion(+) diff --git a/setup.py b/setup.py index e73a1f2bf..55c0ebd1c 100644 --- a/setup.py +++ b/setup.py @@ -6,6 +6,7 @@ import sys from fnmatch import fnmatch +import setuptools from distutils.core import setup import versioneer