bump version to 0.8.0, clean CI, remove requests (#2545)
* bump version to 0.8.0, clean CI, remove requests * why was that even therepull/2548/head
parent
ff47be3a01
commit
4c984bba7e
|
@ -178,9 +178,6 @@ jobs:
|
|||
- if: ${{ matrix.task == 'optimage'}}
|
||||
name: Test GPU IMAGE=2 ops
|
||||
run: GPU=1 IMAGE=2 python -m pytest -n=auto test/test_ops.py
|
||||
- if: ${{ matrix.task == 'optimage' }}
|
||||
name: Test tensor core ops
|
||||
run: GPU=1 TC=2 python -m pytest -n=auto test/test_ops.py
|
||||
- if: ${{ matrix.task == 'openpilot' }}
|
||||
name: Test openpilot model compile and size
|
||||
run: |
|
||||
|
@ -288,10 +285,8 @@ jobs:
|
|||
run: METAL=1 python -m pytest -n=auto test/ --ignore=test/external --ignore=test/models --durations=20
|
||||
- name: Run ONNX
|
||||
run: METAL=1 python -m pytest -n=auto test/external/external_test_onnx_backend.py
|
||||
- name: Run whisper test
|
||||
run: METAL=1 python -m pytest test/models/test_whisper.py
|
||||
- name: Test tensor core reshape-only ops
|
||||
run: METAL=1 TC=2 python -m pytest -n=auto test/test_ops.py
|
||||
- name: Test tensor core ops
|
||||
run: METAL=1 TC=2 python test/test_ops.py TestOps.test_big_gemm
|
||||
|
||||
testhipcompilation:
|
||||
name: HIP Compilation Tests
|
||||
|
|
|
@ -1,13 +1,13 @@
|
|||
import random
|
||||
import functools
|
||||
from pathlib import Path
|
||||
import requests
|
||||
import numpy as np
|
||||
import nibabel as nib
|
||||
from scipy import signal
|
||||
import torch
|
||||
import torch.nn.functional as F
|
||||
from tinygrad.tensor import Tensor
|
||||
from tinygrad.helpers import fetch
|
||||
|
||||
BASEDIR = Path(__file__).parent / "kits19" / "data"
|
||||
|
||||
|
@ -25,8 +25,8 @@ mv kits extra/datasets
|
|||
|
||||
@functools.lru_cache(None)
|
||||
def get_val_files():
|
||||
data = requests.get("https://raw.githubusercontent.com/mlcommons/training/master/image_segmentation/pytorch/evaluation_cases.txt")
|
||||
return sorted([x for x in BASEDIR.iterdir() if x.stem.split("_")[-1] in data.text.split("\n")])
|
||||
data = fetch("https://raw.githubusercontent.com/mlcommons/training/master/image_segmentation/pytorch/evaluation_cases.txt").read_text()
|
||||
return sorted([x for x in BASEDIR.iterdir() if x.stem.split("_")[-1] in data.split("\n")])
|
||||
|
||||
def load_pair(file_path):
|
||||
image, label = nib.load(file_path / "imaging.nii.gz"), nib.load(file_path / "segmentation.nii.gz")
|
||||
|
|
4
setup.py
4
setup.py
|
@ -8,7 +8,7 @@ with open(directory / 'README.md', encoding='utf-8') as f:
|
|||
long_description = f.read()
|
||||
|
||||
setup(name='tinygrad',
|
||||
version='0.7.0',
|
||||
version='0.8.0',
|
||||
description='You like pytorch? You like micrograd? You love tinygrad! <3',
|
||||
author='George Hotz',
|
||||
license='MIT',
|
||||
|
@ -35,7 +35,6 @@ setup(name='tinygrad',
|
|||
"pre-commit",
|
||||
"ruff",
|
||||
"types-tqdm",
|
||||
"types-requests",
|
||||
],
|
||||
'testing': [
|
||||
"torch",
|
||||
|
@ -52,7 +51,6 @@ setup(name='tinygrad',
|
|||
"sentencepiece",
|
||||
"tiktoken",
|
||||
"librosa",
|
||||
"requests",
|
||||
"networkx",
|
||||
]
|
||||
},
|
||||
|
|
Loading…
Reference in New Issue