-
Notifications
You must be signed in to change notification settings - Fork 32
/
Copy pathpyproject.toml
49 lines (46 loc) · 1.35 KB
/
pyproject.toml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
[tool.codespell]
# note: pre-commit passes explicit lists of files here, which this skip file list doesn't override -
skip = "./.git,./pyproject.toml,./.ruff_cache"
ignore-regex = "\\b(.{1,4}|[A-Z]\\w*T)\\b"
builtin = "clear"
quiet-level = 3
[tool.ruff]
fix = true
line-length = 120
[tool.ruff.lint]
select = [
# pycodestyle
"E",
# pyflakes
"F",
# isort
"I",
# numpy
"NPY",
# pyupgrade
"UP",
# flake8-bugbear
"B"
]
[tool.ruff.lint.per-file-ignores]
"source/examples/**/*.ipynb" = [
# "module level import not at top of cell".
# This is sometimes necessary, for example to ship a self-contained function
# around with Dask.
"E402",
]
"source/examples/rapids-ec2-mnmg/notebook.ipynb" = [
# "undefined name cluster", because in this notebook we recommend, in a markdown
# cell, creating a Dask cluster separately and then running the rest of the notebook's code
"F821",
]
"source/examples/rapids-sagemaker-higgs/notebook.ipynb" = [
# "Line too long", because of a 1-liner shell command starting with '!'
"E501",
]
"source/examples/xgboost-dask-databricks/notebook.ipynb" = [
# "undefined name spark" because Databricks magically makes a SparkSession
# available with name 'spark'
# ref: https://docs.databricks.com/en/migration/spark.html#remove-sparksession-creation-commands
"F821",
]