I hereby claim:
- I am harpone on github.
- I am harpone (https://keybase.io/harpone) on keybase.
- I have a public key ASDFQqiM9-Yg0wVF3nsnOaVSan8jKCAgtTLccHHts8RhaQo
To claim this, I am signing this object:
I hereby claim:
To claim this, I am signing this object:
def forward_sequential(h, xs, U, W, nu, theta): | |
"""Forward pass through the network sequentially over input `xs` of any length. | |
NOTE: has no batch dimension. To be batched with `vmap`. | |
Args: | |
h (torch.tensor): shape [D_h, ]; previous state | |
xs (torch.tensor): shape [T, D_x]; input sequence | |
U (torch.tensor): Parameter matrix of shape [D_h, D_x] | |
W (torch.tensor): Parameter matrix of shape [D_h, D_x] | |
xi (torch.tensor): Parameter vector of shape [D_h, ] |
# PARALLEL: | |
from contractpool import ContractPool # imaginary 'contractpool' library, similar to python's `multiprocessing` | |
W = parameter(M, N) | |
def forward(x_: float32[N]) -> float32: | |
# matrix-vector multiplication: | |
zs = float32[M] # let's imagine we have a float32 dtype in Vyper | |
with ContractPool(dot, M) as p: | |
p.map(W, x_, out=zs) # launches M subcontracts asynchronously, each subcontract writes values to zs |
# SERIAL: | |
W = parameter(M, N) | |
def forward(x_: float32[N]) -> float32: | |
# matrix-vector multiplication: | |
zs = float32[M] # let's imagine we have a float32 dtype in Vyper | |
for i, W_row in enumerate(W): | |
zs[i] = dot(W_row, x_) # 'dot' is an external smart contract | |
# summation: |
Long-lasting COVID-19 - | |
Consensus statement of the expert group appointed by STM on 31 December 2021 | |
VN / 20672/2021 | |
DRAFT 7.1.2022 | |
from itertools import islice | |
import os | |
import torch | |
from torch.utils.data import DataLoader | |
from torchvision import transforms | |
import numpy as np | |
import torch_xla.distributed.parallel_loader as pl | |
import torch_xla.core.xla_model as xm | |
import torch_xla.distributed.xla_multiprocessing as xmp |
from itertools import islice | |
from munch import Munch | |
import sys, os | |
from torch.utils.data import DataLoader | |
from torchvision import transforms | |
import time | |
import webdataset as wds | |
sys.path.append(os.getcwd()) |
""" | |
MIT License | |
knn, kl_div, entropy Copyright (c) 2017 Heikki Arponen | |
""" | |
import torch | |
def knn(x, y, k=3, last_only=False, discard_nearest=True): |
class Lamb(Optimizer): | |
r"""Implements Lamb algorithm. | |
It has been proposed in `Large Batch Optimization for Deep Learning: Training BERT in 76 minutes`_. | |
Arguments: | |
params (iterable): iterable of parameters to optimize or dicts defining | |
parameter groups | |
lr (float, optional): learning rate (default: 1e-3) | |
betas (Tuple[float, float], optional): coefficients used for computing | |
running averages of gradient and its square (default: (0.9, 0.999)) | |
eps (float, optional): term added to the denominator to improve |
class GCSDataset(Dataset): | |
"""Generic PyTorch dataset for GCS. Streams data from GCS and (optionally) caches to local disk. | |
""" | |
def __init__(self, | |
bucketname=None, | |
path_list=None, # TODO: list bucket/path contents if None | |
target_list=None, | |
transform=None, |