8000 Update torchvision requirement from <0.10.0,>=0.8.1 to >=0.8.1,<0.11.0 by dependabot[bot] · Pull Request #5266 · allenai/allennlp · GitHub
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content
This repository was archived by the owner on Dec 16, 2022. It is now read-only.

Update torchvision requirement from <0.10.0,>=0.8.1 to >=0.8.1,<0.11.0 #5266

Merged
merged 16 commits into from
Jun 17, 2021
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
28 changes: 25 additions & 3 deletions .github/workflows/ci.yml
Original file line number Diff line number Diff line change
Expand Up @@ -98,7 +98,11 @@ jobs:
# Required to use the setup-python action.
AGENT_TOOLSDIRECTORY: '/opt/hostedtoolcache'
# Our self-hosted runner currently is currently compatible with CUDA 11.*.
TORCH_VERSION: 'torch==1.8.1+cu111 torchvision==0.9.1+cu111 -f https://download.pytorch.org/whl/torch_stable.html'
TORCH_VERSION: 'torch==1.9.0+cu111 torchvision==0.10.0+cu111 -f https://download.pytorch.org/whl/torch_stable.html'
# Disable tokenizers parallelism because this doesn't help.
TOKENIZERS_PARALLELISM: 'false'
# Disable multithreading with OMP because this can lead to dead-locks in distributed training.
OMP_NUM_THREADS: '1'

steps:
- uses: actions/checkout@v2
Expand Down Expand Up @@ -149,9 +153,15 @@ jobs:
cpu_tests:
name: CPU Tests
runs-on: ubuntu-latest
timeout-minutes: 20
strategy:
matrix:
python: ['3.7', '3.8']
env:
# Disable tokenizers parallelism because this doesn't help.
TOKENIZERS_PARALLELISM: 'false'
# Disable multithreading with OMP because this can lead to dead-locks in distributed training.
OMP_NUM_THREADS: '1'

steps:
- uses: actions/checkout@v2
Expand Down Expand Up @@ -197,7 +207,13 @@ jobs:

model_tests:
name: Model Tests
timeout-minutes: 18
runs-on: ubuntu-latest
env:
# Disable tokenizers parallelism because this doesn't help.
TOKENIZERS_PARALLELISM: 'false'
# Disable multithreading with OMP because this can lead to dead-locks in distributed training.
OMP_NUM_THREADS: '1'

steps:
- uses: actions/checkout@v2
Expand Down Expand Up @@ -249,6 +265,7 @@ jobs:

upload_coverage:
name: Upload Coverage Report
timeout-minutes: 18
if: github.repository == 'allenai/allennlp' && (github.event_name == 'push' || github.event_name == 'pull_request')
runs-on: ubuntu-latest
needs: [cpu_tests, gpu_tests, model_tests]
Expand Down Expand Up @@ -288,6 +305,7 @@ jobs:
# Builds package distribution files for PyPI.
build_package:
name: Build Package
timeout-minutes: 18
runs-on: ubuntu-latest

steps:
Expand Down Expand Up @@ -354,6 +372,7 @@ jobs:
# Tests installing from the distribution files.
test_package:
name: Test Package
timeout-minutes: 18
needs: [build_package] # needs the package artifact created from 'build_package' job.
runs-on: ubuntu-latest
strategy:
Expand Down Expand Up @@ -391,6 +410,7 @@ jobs:
# Builds Docker image from the core distribution files and uploads to Docker Hub.
docker:
name: Docker (CUDA ${{ matrix.cuda }})
timeout-minutes: 18
if: github.repository == 'allenai/allennlp'
# Run on self-hosted to utilize layer caching.
runs-on: [self-hosted]
Expand All @@ -409,9 +429,9 @@ jobs:
if [[ $CUDA == '10.1' ]]; then
echo "DOCKER_TORCH_VERSION='torch==1.8.0+cu101 torchvision==0.9.0+cu101 -f https://download.pytorch.org/whl/torch_stable.html'" >> $GITHUB_ENV;
elif [[ $CUDA == '10.2' ]]; then
echo "DOCKER_TORCH_VERSION='torch==1.8.0'" >> $GITHUB_ENV;
echo "DOCKER_TORCH_VERSION='torch==1.9.0 torchvision==0.10.0'" >> $GITHUB_ENV;
elif [[ $CUDA == '11.1' ]]; then
echo "DOCKER_TORCH_VERSION='torch==1.8.0+cu111 torchvision==0.9.0+cu111 -f https://download.pytorch.org/whl/torch_stable.html'" >> $GITHUB_ENV;
echo "DOCKER_TORCH_VERSION='torch==1.9.0+cu111 torchvision==0.10.0+cu111 -f https://download.pytorch.org/whl/torch_stable.html'" >> $GITHUB_ENV;
else
echo "Unhandled CUDA version $CUDA";
exit 1;
Expand Down Expand Up @@ -463,6 +483,7 @@ jobs:
# allennlp-docs repo.
docs:
name: Docs
timeout-minutes: 18
# Don't run for forks.
if: github.repository == 'allenai/allennlp'
runs-on: ubuntu-latest
Expand Down Expand Up @@ -599,6 +620,7 @@ jobs:
# Publish the core distribution files to PyPI.
publish:
name: PyPI
timeout-minutes: 18
needs: [style, lint, cpu_tests, gpu_tests, model_tests, build_package, test_package, docker, docs]
# Only publish to PyPI on releases and nightly builds to "allenai/allennlp" (not forks).
if: github.repository == 'allenai/allennlp' && (github.event_name == 'release' || github.event_name == 'schedule')
Expand Down
7 changes: 3 additions & 4 deletions Makefile
Original file line number Diff line number Diff line change
Expand Up @@ -10,14 +10,13 @@ MD_DOCS_CONF_SRC = mkdocs-skeleton.yml
MD_DOCS_TGT = site/
MD_DOCS_EXTRAS = $(addprefix $(MD_DOCS_ROOT),README.md CHANGELOG.md CONTRIBUTING.md)

TORCH_VERSION = torch==1.8.1 torchvision==0.9.1
TORCH_VERSION = torch==1.9.0 torchvision==0.10.0

DOCKER_TAG = latest
DOCKER_IMAGE_NAME = allennlp/allennlp:$(DOCKER_TAG)
DOCKER_TEST_IMAGE_NAME = allennlp/test:$(DOCKER_TAG)
DOCKER_TORCH_VERSION = 'torch==1.7.1 torchvision==0.8.2'
# Our self-hosted runner currently has CUDA 11.0.
DOCKER_TEST_TORCH_VERSION = 'torch==1.7.1+cu110 torchvision==0.8.2+cu110 -f https://download.pytorch.org/whl/torch_stable.html'
DOCKER_TORCH_VERSION = $(TORCH_VERSION)
DOCKER_TEST_TORCH_VERSION = $(TORCH_VERSION)
DOCKER_RUN_CMD = docker run --rm \
-v $$HOME/.allennlp:/root/.allennlp \
-v $$HOME/.cache/huggingface:/root/.cache/huggingface \
Expand Down
3 changes: 2 additions & 1 deletion allennlp/common/testing/distributed_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -42,7 +42,8 @@ def init_process(

func(global_rank, world_size, gpu_id, *(func_args or []), **(func_kwargs or {}))

dist.barrier()
# dist.barrier()
dist.destroy_process_group()


def run_distributed_test(
Expand Down
2 changes: 1 addition & 1 deletion setup.py
Original file line number Diff line number Diff line change
Expand Up @@ -51,7 +51,7 @@
),
install_requires=[
"torch>=1.6.0,<1.10.0",
"torchvision>=0.8.1,<0.10.0",
"torchvision>=0.8.1,<0.11.0",
"jsonnet>=0.10.0 ; sys.platform != 'win32'",
"overrides==3.1.0",
"nltk",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@
import pytest
import torch

from allennlp.common import Params
from allennlp.common import Params, cached_transformers
from allennlp.common.testing import AllenNlpTestCase, requires_gpu
from allennlp.data import Vocabulary
from allennlp.data.batch import Batch
Expand All @@ -15,6 +15,10 @@


class TestPretrainedTransformerEmbedder(AllenNlpTestCase):
@classmethod
def teardown_class(cls):
cached_transformers._clear_caches()

@requires_gpu
def test_forward_runs_when_initialized_from_params(self):
# This code just passes things off to `transformers`, so we only have a very simple
Expand Down
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
import pytest
import torch

from allennlp.common import Params
from allennlp.common import Params, cached_transformers
from allennlp.common.checks import ConfigurationError
from allennlp.data import Token, Vocabulary
from allennlp.data.batch import Batch
Expand All @@ -14,6 +14,10 @@


class TestPretrainedTransformerMismatchedEmbedder(AllenNlpTestCase):
@classmethod
def teardown_class(cls):
cached_transformers._clear_caches()

@pytest.mark.parametrize("train_parameters", [True, False])
def test_end_to_end(self, train_parameters: bool):
token_indexer = PretrainedTransformerMismatchedIndexer("bert-base-uncased")
Expand Down
4 changes: 4 additions & 0 deletions tests/modules/transformer/transformer_layer_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -17,6 +17,10 @@
)


def teardown_module(function):
cached_transformers._clear_caches()


ATTENTION_PARAMS_DICT = {
"hidden_size": 6,
"num_attention_heads": 2,
Expand Down
0