1
0
Fork 0
Commit Graph

479 Commits (deepcrayon)

Author SHA1 Message Date
George Hotz 801ecd4a07 cleanup clip tokenizer 2022-09-12 09:20:12 -07:00
Fernand Pajot ff0da4c802
Added standalone CLIP tokenizer (#382)
* Added standalone CLIP tokenizer.

* Fixed empty phrase.

* Truncating long prompts.

* Keeping two slots for the start and end token.

* Fixed empty phrase.

* Using tokenizer for empty phrase.

* Typo.
2022-09-12 09:12:55 -07:00
David Redmon a1810c8617
update serious_mnist.py (#380) 2022-09-11 13:37:40 -07:00
George Hotz ecc1a0470d add Linear to tinygrad.nn 2022-09-07 07:40:48 -07:00
George Hotz 896f9f74a9 hmm, need this with broadcast change 2022-09-06 16:54:01 -07:00
George Hotz a18a6a0773 fix sd with TORCH=1 2022-09-06 16:51:16 -07:00
George Hotz 0516359af8 fix stupid OPENCL=1 OOM 2022-09-06 14:29:23 -07:00
George Hotz f215534a64 1100 lines, but sane linter rules 2022-09-06 13:47:45 -07:00
George Hotz 682dc64430 works at work 2022-09-06 08:06:11 -07:00
George Hotz d6f499fd69 improve opencl, why is it OOMing 2022-09-05 20:14:31 -07:00
George Hotz 0ba6179de7 stable diffusion in readme 2022-09-05 18:51:56 -07:00
George Hotz c1d5af8b0c stable diffusion cleanups 2022-09-05 18:34:13 -07:00
George Hotz 3728ef6d02 better alphas 2022-09-05 16:48:26 -07:00
George Hotz 0fda854b3e other prompt example 2022-09-05 16:14:16 -07:00
George Hotz 16cb4290c4 cat horse winning 2022-09-05 16:05:14 -07:00
George Hotz 1043fa067a it renders something 2022-09-05 15:52:14 -07:00
George Hotz 5a685b93ac brown img 2022-09-05 15:20:18 -07:00
George Hotz 98d6264987 all models match 2022-09-05 12:27:54 -07:00
George Hotz b8bd34b5d2 fix last bug in unet probz 2022-09-05 11:32:44 -07:00
George Hotz 3df67aa0af fix transformer bugs 2022-09-05 11:26:32 -07:00
George Hotz 2ed3bb6223 clip model is running 2022-09-05 11:26:32 -07:00
George Hotz 1a54ea2417 runs on torch cpu 2022-09-04 12:06:42 -07:00
George Hotz 9590d92750 stable diffusion compiles (add no_init) 2022-09-04 11:40:50 -07:00
George Hotz 172683c314 work 2022-09-04 11:21:09 -07:00
George Hotz c2a030fe55 one liner that's more clear 2022-09-03 16:08:48 -07:00
George Hotz 4a3ed58edb more readable actually 2022-09-03 16:00:35 -07:00
George Hotz 633f31dc73 easier to read 2022-09-03 15:53:58 -07:00
George Hotz 6578e08919 cleanups for Mid 2022-09-03 15:50:33 -07:00
George Hotz 852de7c66c remove ugly parens 2022-09-03 15:41:37 -07:00
George Hotz 6b190c2fa5 stable diffusion works 2022-09-03 13:55:36 -07:00
George Hotz 947e10dab0 yolo 2022-09-03 12:39:48 -07:00
George Hotz 033a3ecccf found tinygrad bug 2022-09-03 12:32:43 -07:00
George Hotz 114728d363 torch bs 2022-09-03 11:57:23 -07:00
George Hotz 356732515b stable_diffusion: add attn and layernorm 2022-09-03 11:02:27 -07:00
George Hotz 4dadd95e3c fix tests hopefully, more stable diffusion 2022-09-03 10:38:31 -07:00
George Hotz c01a8c5c2d stable diffusion start 2022-09-03 10:08:42 -07:00
George Hotz b132de677d
tinygrad.nn (#367)
* tinygrad.nn

* flake8

* working on pylint

* more pylint

* more pylint

* pylint passes

* networkx

* mypy can't infer that type

* junk
2022-08-18 07:41:00 -07:00
George Hotz acbeaf0ba9 adam in benchmark_train_efficientnet 2022-07-19 09:33:07 -07:00
George Hotz d985217fa4 skip reduce noops 2022-07-16 07:47:43 -07:00
George Hotz 5e46561f7e no_grad = NOT backward 2022-07-10 20:54:57 -07:00
George Hotz d5d9cffe7c training param for batchnorm 2022-07-04 13:28:03 -07:00
George Hotz 34f43ea10e LAZY and CLCACHE are defaults 2022-07-04 13:09:15 -07:00
George Hotz b7afd83267 track cl mem used 2022-07-04 12:19:00 -07:00
George Hotz d5de8452c6 dashed loadops 2022-07-04 09:50:56 -07:00
George Hotz 7276f8d6bf improve constant folding, detach before moving tensor 2022-07-02 15:29:40 -07:00
George Hotz 0cb99d72e9 NUM=-1 is a small efficientnet for small people 2022-07-02 15:11:51 -07:00
George Hotz 8cf1aed0f4 don't track_running_stats, parameters must require_grad 2022-07-02 14:38:45 -07:00
George Hotz f607f18006 fix backward 2022-06-25 00:00:53 -07:00
George Hotz ec30f0402f improve benchmark_train_efficientnet 2022-06-24 23:46:38 -07:00
George Hotz d748353ce5 err, okay, a bit more off 2022-06-24 22:44:57 -07:00
George Hotz bdde95f16e CACHE_LAZYBUFFERS options + benchmark. only a couple x from torch 2022-06-24 22:33:53 -07:00
Edward Betts a6203b5ecd
Correct spelling mistakes (#314) 2022-04-05 05:22:18 -07:00
Ildar Kashaev d065040df0
Fix yolo param name (#306) 2022-02-10 01:40:24 -07:00
George Hotz 2a10116bfa support drawing graphs 2022-01-16 10:45:58 -08:00
George Hotz 2cae2dfa07 don't crash the dataloader for imagenet 2022-01-16 08:41:26 -08:00
George Hotz 907ff7dbb6 imagenet is training 2022-01-15 23:27:31 -08:00
George Hotz d1e082e0ef add imagenet training support 2022-01-15 23:16:38 -08:00
George Hotz ade2af7ac0 data doesn't require grad 2022-01-15 22:41:27 -08:00
George Hotz 0973e54eb0 fix DEBUG for GPU 2022-01-15 22:14:28 -08:00
cjg91 7025c9bbeb
Transfer learning for ResNet (#295)
* Transfer learning for ResNet

* moved ResNet depth specifics into the class
2022-01-15 23:22:10 -05:00
George Hotz 55d792b065 Revert "fixup resnet"
This reverts commit 4eabe677ed.
2022-01-15 20:22:01 -08:00
George Hotz 4eabe677ed fixup resnet 2022-01-15 20:21:02 -08:00
George Hotz e0bef0bd01 training is False by default 2022-01-15 19:57:41 -08:00
George Hotz 8ec2341cca fix bn training 2022-01-15 19:47:01 -08:00
George Hotz c0c2c0b041 support larger ViT models 2021-12-12 10:45:10 -08:00
George Hotz 4b7c31b5b7 break vit into it's own file 2021-11-30 11:19:22 -05:00
George Hotz 46bbbcf7f0 model touchups 2021-11-30 11:13:34 -05:00
George Hotz 835869974c clean up vit code 2021-11-30 10:58:03 -05:00
George Hotz 535f02cc64 use sequential 2021-11-30 00:25:39 -05:00
George Hotz de938c2d9d vit is now tested 2021-11-30 00:23:06 -05:00
George Hotz aff810e722 unify transformer block 2021-11-29 18:58:15 -05:00
George Hotz 58ed46963e fix broadcastdot 2021-11-29 18:54:57 -05:00
George Hotz 125e74293f promote layernorm to tensor op 2021-11-29 18:08:21 -05:00
George Hotz dca076dbf1 remove dumb nn ops 2021-11-29 18:05:31 -05:00
George Hotz 70544e7e9f sum hook override 2021-11-29 17:14:24 -05:00
George Hotz 8097b8f7d6 vit works 2021-11-29 16:28:14 -05:00
George Hotz 7c07c5efdd plz fix vit 2021-11-29 15:45:19 -05:00
George Hotz ca160504e1 affine is always the last dim 2021-11-29 15:22:49 -05:00
George Hotz e86f7a4aa3 deterministic 2021-11-29 15:10:15 -05:00
George Hotz f909ab194f gelu with broken test 2021-11-29 15:00:50 -05:00
George Hotz 9ce881f88c fix bug in getitem, drop int axis 2021-11-29 14:01:24 -05:00
George Hotz 99b6051467 add ff_dim to transformer 2021-11-29 12:40:52 -05:00
George Hotz b0f14b4af8 move datasets into datasets 2021-10-30 19:55:50 -07:00
George Hotz ba28761894 move yolo into examples/yolo 2021-10-30 19:46:00 -07:00
George Hotz 121d5a17ee use tinynn for Conv2d 2021-10-30 19:40:44 -07:00
Guglielmo Camporese 2b7589db64
Added ResNet-{18, 34, 50, 101, 152} (#271)
* added resnets

* fix minor

* fix minor

* resnet in models

* added resnet test

* added resnet train test

* added linear, conv2d nn tests

* fix minor in extra/training

* resnet in models

* fix minor

* fix tolerance for linear in nn test

* fix eval, this causes cpu and gpu UT failing

* revert transformer test

* fix minor for CPU test

* improved model get_params for sequential layer

* fix minor for params counting

* commented broken ops tests

* improved train for resnet
2021-06-21 09:37:24 -07:00
George Hotz d3f169b267 move good models to models, add a training step test 2021-06-19 11:24:15 -07:00
Jacky Lee 611d81dcb4
Add asserts for non-zero indices (#264) 2021-06-13 21:14:46 -07:00
Josh Smith ad756f6112
minor optimizations & cleaning (#257)
* use isinstance, some optimizations & whitespace removal

* revert whitespace changes

* revert more whitespace

* some more cleanup

* revert fstring (not a fan of the {{}})

* fix typo

* fix typo
2021-06-02 09:57:15 -07:00
George Hotz b80cacb416 fix GPU efficientnet example 2021-05-26 17:29:35 -07:00
20kdc 2653d33292
vgg7 (image upscaling) implementation - not the best, but it works (#255)
* vgg7 implementation - not the best, but it works

* VGG7 implementation: Spread nansbane to deter NaNs, maybe improved training experience

* VGG7 implementation: Fix training, for real this time

Results actually attempt to approximate the input

* VGG7 implementation: Sample probability management
2021-05-12 23:48:51 -07:00
Skosh 81bf933a91
Improved __getitem__ (#254)
* Some progress on yolov3

* Removed some debugging comments… Also, the forward pass eats all RAM for some reason

* forward pass almost runs

* forward pass runs almost

* forward pass runs, now we gotta load the weights

* loading weights works

* fetches config and weights

* everything kind of works, postprocessing of output still needs to be implemented, temp_process_results kind of works, but its kind of terrible, and not how things should be done

* some changes

* fixed some bugs in the forward pass and load_weights function, now outputs more correct values, however some values are still loaded incorrectly

* Something is wrong with the forward pass, Conv2d tests added

* forward pass almost outputs correct values, gotta fix one more thign

* yolo works

* some final changes

* reverting changes

* removed dataloader

* fixed some indentation

* comment out failing test, somehow it fails CI even though it passes on my computer…

* fixed wrong probabilities

* added webcam option to YOLO, now just need to add bounding boxes and speed it up

* some progress towards adding bounding boxes

* trying to speed up yolo layer on GPU, still faster on CPU but with 30GB ram usage

* Faster inference times, bounding boxes added correctly, webcam works, but is slow, and there is a memory leak when running on CPU... Also added tinygrads output on the classic dog image

* removed some debugging print statements

* updated result image

* something weird is going on, mean op on GPU tensor randomly faults, copying a tensor from GPU->CPU takes 10+ seconds…

* Improved __getitem__

* Updated

* Updated __getitem__

* Linebreaks

* Maybe this works?

* Added MNIST locally, tests run now
2021-05-05 22:15:22 -07:00
Skosh 78aa147b39
[WIP] YOLO working on tinygrad! (#245)
* Some progress on yolov3

* Removed some debugging comments… Also, the forward pass eats all RAM for some reason

* forward pass almost runs

* forward pass runs almost

* forward pass runs, now we gotta load the weights

* loading weights works

* fetches config and weights

* everything kind of works, postprocessing of output still needs to be implemented, temp_process_results kind of works, but its kind of terrible, and not how things should be done

* some changes

* fixed some bugs in the forward pass and load_weights function, now outputs more correct values, however some values are still loaded incorrectly

* Something is wrong with the forward pass, Conv2d tests added

* forward pass almost outputs correct values, gotta fix one more thign

* yolo works

* some final changes

* reverting changes

* removed dataloader

* fixed some indentation

* comment out failing test, somehow it fails CI even though it passes on my computer…

* fixed wrong probabilities

* added webcam option to YOLO, now just need to add bounding boxes and speed it up

* some progress towards adding bounding boxes

* trying to speed up yolo layer on GPU, still faster on CPU but with 30GB ram usage

* Faster inference times, bounding boxes added correctly, webcam works, but is slow, and there is a memory leak when running on CPU... Also added tinygrads output on the classic dog image

* removed some debugging print statements

* updated result image

* something weird is going on, mean op on GPU tensor randomly faults, copying a tensor from GPU->CPU takes 10+ seconds…
2021-04-25 18:06:52 -07:00
George Hotz ce77dda805 yolov5 v4 2021-01-05 07:56:17 -08:00
Asim 1c148f2fe4
fixed example broken after gpu refactor (#238) 2021-01-05 07:41:54 -08:00
George Hotz 6842ad9ec8 minor cleanups, yolo work 2021-01-03 08:14:16 -08:00
George Hotz 895d142503 start trying to load yolo v5 2021-01-02 12:51:55 -08:00
Marcel Bischoff 42b4761025
transformer >99.98% test accuracy in ~30s (#230)
* transformer

* BS might divide len(Y_test)

* outoput when accuracy is high

* more readeable

* fixed loss in serious_mnist for new API
2021-01-02 07:45:09 -08:00
George Hotz f9170505b3 if you like your transformers twice as slow, use the GPU 2020-12-29 17:14:23 -05:00
George Hotz 3f8e137b6f extra/transformer 2020-12-29 14:14:00 -05:00
George Hotz bcb3ceeca3 set training in functions 2020-12-28 22:45:46 -05:00
George Hotz 51bf164b72 dropout, training 2020-12-28 22:12:23 -05:00
George Hotz 7b8fee038d it works! forgot the sqrt 2020-12-28 16:23:52 -05:00
George Hotz 1faf05ef67 ahh, it's better if i don't train the embedding 2020-12-28 16:07:02 -05:00
George Hotz c3832e1bde hmm, fix layernorm to not be batchnorm and it breaks 2020-12-28 13:06:21 -05:00
George Hotz 2e89e75dcb layernorm fixes transformer instability 2020-12-28 12:58:15 -05:00
George Hotz 593233b668 log and exp are first class ops 2020-12-28 10:00:30 -05:00
Marcel Bischoff ffff98db78
Evaluation in Transformers (#218)
* 2serious

* load/save

* fixing GPU

* added DEBUG

* needs BatchNorm or doesn't learn anything

* old file not needed

* added conv biases

* added extra/training.py and checkpoint

* assert in test only

* save

* padding

* num_classes

* checkpoint

* checkpoints for padding

* training was broken

* merge

* rotation augmentation

* more aug

* needs testing

* streamline augment, augment is fast thus bicubic

* tidying up

* transformer eval
2020-12-28 09:24:51 -05:00
George Hotz 65b07d2f4f fix onehot embed 2020-12-27 18:50:38 -05:00
George Hotz d864e1c71a transformer is training 2020-12-27 18:46:32 -05:00
George Hotz a361ef6861 fixup training loop 2020-12-27 18:35:56 -05:00
George Hotz f15bec6dbc make multidot work on CPU 2020-12-27 17:25:37 -05:00
George Hotz 131e04c90c cpu only decorator 2020-12-27 17:18:55 -05:00
George Hotz 2f1b2c0a3b add transpose, start on transformer 2020-12-27 16:59:12 -05:00
Liam 075cf4bc02
Update examples to new API (#205) 2020-12-19 09:04:12 -08:00
George Hotz e5f2055776 ane refactor, fix examples 2020-12-18 09:19:06 -08:00
iainwo 56d44637f3
fixed pylint, formatted python files iwth cblack on localhost (#204)
* fixed pylint, formatted python files iwth cblack on localhost

* Revert "fixed pylint, formatted python files iwth cblack on localhost"

This reverts commit 07e2b88466.

* dedented 4-spaces added linter

Co-authored-by: Iain Wong <iainwong@outlook.com>
2020-12-17 14:37:31 -08:00
Marcel Bischoff da72a0eed4
Big MNIST model with PIL augmentation and load/save (#160)
* 2serious

* load/save

* fixing GPU

* added DEBUG

* needs BatchNorm or doesn't learn anything

* old file not needed

* added conv biases

* added extra/training.py and checkpoint

* assert in test only

* save

* padding

* num_classes

* checkpoint

* checkpoints for padding

* training was broken

* merge

* rotation augmentation

* more aug

* needs testing

* streamline augment, augment is fast thus bicubic

* tidying up
2020-12-13 20:45:55 -08:00
NeuralLink 1a1c63a08b
Gan is real...Look what tiny just generated! (#192)
* mode collapse solved

* info add

* delete unnecessary imports

* readme
2020-12-13 20:23:12 -08:00
Marcel Bischoff 6785614239
tinygrad.utils to extra.utils fix in mnist_gan (#190) 2020-12-12 20:52:36 -08:00
NeuralLink d901ef6b23
🎉 effort to generate mnist data using GAN with tinygrad. [WIP] (#166)
* 🎉 effort to generate mnist data with tinygrad.

* dropout added

* working gan

* minor bug fixes

* more bug fixes

* todo reg l2

* detach

* logsoftmax twice
2020-12-12 17:58:04 -08:00
George Hotz a5aced8d47 30 MEGAReLUs. we need to lose 12 lines 2020-12-12 17:07:34 -08:00
George Hotz bc5df477de readme and .ane() 2020-12-12 16:15:38 -08:00
George Hotz da873cd556
Single ReLU in ANE (#188)
* aneworks

* cleanup
2020-12-12 16:11:34 -08:00
George Hotz 1d10559d1d tinygrad.utils -> extra.utils 2020-12-12 15:26:07 -08:00
Daulet c7e95ddb21
Add diamond model test (#181)
* add backward pass test for diamond model

* fix train_efficientnet example
2020-12-11 09:21:36 -08:00
Marcel Bischoff d204f09316
some progress on batchnorms (draft) (#147)
* no of categories for efficientnet

* need layer_init_uniforn

* merge fail

* merge fail

* batchnorms

* needs work

* needs work how determine training

* pow

* needs work

* reshape was needed

* sum with axis

* sum with axis and tests

* broken

* works again

* clean up

* Update test_ops.py

* using sum

* don't always update running_stats

* space

* self

* default return running_stats

* passes test

* need to use mean

* merge

* testing

* fixing pow

* test_ops had a line dropped

* undo pow

* rebase
2020-12-09 22:14:27 -08:00
George Hotz 9a64d13b94 add conv biases and max pool 2020-12-09 08:01:20 -08:00
George Hotz 99fa65f057 enable batchnorm in serious mnist 2020-12-09 03:29:40 -08:00
George Hotz c225e62dd2 touchups 2020-12-09 02:52:28 -08:00
George Hotz 0e02f394ee serious_mnist 2020-12-08 21:43:05 -08:00
George Hotz 00312b8ad1 batchnorm work 2020-12-06 14:40:07 -08:00
George Hotz 102e6356e9 replace layer_init_uniform with .uniform 2020-12-06 13:44:31 -08:00
George Hotz 521098cc2f se optional, track time better 2020-12-06 12:29:42 -08:00
George Hotz 609d11e699 trainer works with CIFAR 2020-12-06 12:20:14 -08:00
George Hotz 80a9c777ba requires grad, optim in train enet 2020-12-06 11:10:30 -08:00
George Hotz c66c27d22e get parameters 2020-12-06 10:45:04 -08:00
George Hotz 51daaa43d4 fix memory leaks, add gc test 2020-12-06 10:34:40 -08:00
George Hotz b8deb36e56 train BS=16 for 32 steps 2020-12-04 10:00:32 -08:00
adamritter 5797e63d9b
Train efficientnet should respect NUM environment variable (#122)
Co-authored-by: holonomicjl <58403584+holonomicjl@users.noreply.github.com>
2020-11-16 20:02:31 -08:00
George Hotz 2ffb8de1ea move efficientnet to extra 2020-11-16 08:08:07 -08:00
dustcollector12 6f033ea30a
enable local images for efficientnet.py (#116) 2020-11-13 07:00:12 -08:00
Marcel Bischoff a3989f9e18
Supporting .png files in efficientnet (#102)
* to make it work locally

* definitely not working

* Conv2D GPU passes some of the tests

* Conv2D GPU passes more of the tests

* passes some tests and mnist

* removed unecessary code

* Conv2D Backpass works

* wrong test_ops.py

* white space + test backward

* ereased useless code

* removed default argument

* long lines

* works also with 4 channel .png files

* commenting out

* track
2020-11-10 20:06:24 -08:00
George Hotz db755fa103 promote swish to a tensor ops 2020-11-10 15:48:11 -08:00
George Hotz 55012d21bb debug in backward pass too 2020-11-10 01:19:52 -08:00
George Hotz c76a20b4be 4s and 7s work 2020-11-10 00:54:17 -08:00
George Hotz 8b23033fa9 support all the enet sizes 2020-11-09 18:04:16 -08:00
George Hotz 9db95ab942 fix enet padding 2020-11-09 17:56:57 -08:00
George Hotz 7ac1b163a5 add backward to enet train 2020-11-09 16:05:52 -08:00
George Hotz 8ca9c0205f train_efficientnet is broken still 2020-11-09 16:01:16 -08:00
George Hotz 53157fb876 add back scale 2020-11-09 10:20:56 -08:00
George Hotz 3ffbd47335 Revert "Revert "pygame is fine, cv2 can also do the trick (#79)" (#85)"
This reverts commit 6b982621f8.
2020-11-09 10:18:48 -08:00
George Hotz 6b982621f8
Revert "pygame is fine, cv2 can also do the trick (#79)" (#85)
This reverts commit e7f2f43331.
2020-11-09 10:03:38 -08:00
dustcollector12 e7f2f43331
pygame is fine, cv2 can also do the trick (#79)
* pygame is fine, cv2 can also do the trick

* retimg and copy constructor not needed

* shape is missing without copy constructor

* retimg put back

* addressing capture buffering
2020-11-09 10:02:06 -08:00
George Hotz 3ab8fb2189 oops, timing 2020-11-07 12:15:52 -08:00
George Hotz e4c5fb0219 webcam support 2020-11-07 12:15:16 -08:00
George Hotz e6c8321e5b enet is running on GPU 2020-11-07 11:33:46 -08:00
George Hotz fbff6ab2e5 fix strided convs, GPU env var for enet 2020-11-07 10:26:37 -08:00
Marcel Bischoff c69582744d
Update efficientnet.py (#53)
from https://github.com/lukemelas/EfficientNet-PyTorch/blob/master/efficientnet_pytorch/utils.py
```
    blocks_args = [
        'r1_k3_s11_e1_i32_o16_se0.25',
        'r2_k3_s22_e6_i16_o24_se0.25',
        'r2_k5_s22_e6_i24_o40_se0.25',
       'r3_k3_s22_e6_i40_o80_se0.25',
        'r3_k5_s11_e6_i80_o112_se0.25',
        'r4_k5_s22_e6_i112_o192_se0.25',
        'r1_k3_s11_e6_i192_o320_se0.25',
    ]
```

now it's a persian cat.
2020-11-04 09:36:27 -08:00
Marcel Bischoff c40862241d
Revert "choose aspect ration dependent on landscape or portrait (#48)" (#51)
This reverts commit 2808411421.
2020-11-02 19:31:07 -08:00
Anders Aaen Springborg 2808411421
choose aspect ration dependent on landscape or portrait (#48) 2020-11-02 18:10:22 -08:00
Marcel Bischoff af5a4e0f5a
alternative to einsum which is slow in extreme cases (#44)
* streamlined numerical_jacobian

* Got rid of the g loop in Conv2D.forward

* ereased stupid line

* nothing

* no loops in Conv2D forward

* Conv2D backprop improved

* stupid things in examples

* alternative to einsum

* Conv2D backward einsum alternative

* tidying up

* tidied up

* no ravel

* got rid of print

* Update efficientnet.py

* Update efficientnet.py

* Update efficientnet.py

* only tensordot

* 255.0

* whitespace

* aspect ratio error in efficientnet

* noprint

Co-authored-by: Marcel Bischoff <marcel@Marcels-iMac.local>
2020-11-02 06:36:07 -08:00
George Hotz e01e35e545 14 ops to write for GPU 2020-10-31 10:59:30 -07:00
George Hotz 06928cf3cc ENET WORKS 2020-10-31 10:42:58 -07:00
George Hotz 68cba88e8f fix enet bugs, now is mousetrap 2020-10-31 10:28:07 -07:00
George Hotz 053c319708 optional look at cat 2020-10-30 09:30:15 -07:00
George Hotz 2c7e75d733
group conv: forward pass works (#34)
* forward pass works

* got the backward pass

* okay, it's now a coho
2020-10-30 09:19:20 -07:00
George Hotz 71aedc2309 it thinks it's washer. it's cat. bad net. you do bad. 2020-10-30 08:28:05 -07:00
George Hotz c14473f87d unit test for batchnorm2d 2020-10-30 08:19:58 -07:00
George Hotz 1dde4ce609 oops, stay in float32 2020-10-29 08:24:12 -07:00
George Hotz 2db670ef26 enet refactor + no sigmoid warning 2020-10-29 08:08:21 -07:00
George Hotz 17fa74c15b factor out a couple nn ops 2020-10-29 08:01:12 -07:00
George Hotz cc17e3271a try to recognize cat. do not succeed 2020-10-27 21:41:52 -07:00
George Hotz 03d9c98f5b enet work 2020-10-27 21:23:02 -07:00
George Hotz 0ec279951f enet weight loading 2020-10-27 21:01:48 -07:00
George Hotz e84ad3e27d enet runs 2020-10-27 20:16:29 -07:00
George Hotz 9166eb58bb more enet 2020-10-27 19:37:21 -07:00
George Hotz 41828d768f sigmoid and more enet 2020-10-27 19:13:47 -07:00
George Hotz 09d1ebcdaa examples is better 2020-10-27 18:57:00 -07:00