Bootstrap knowledge of LLMs ASAP. With a bias/focus to GPT.
Avoid being a link dump. Try to provide only valuable well tuned information.
Neural network links before starting with transformers.
/// minimal example of adding a custom render pipeline in bevy 0.11. | |
/// | |
/// When this example runs, you should only see a blue screen. There are no | |
/// vertex buffers, or anything else in this example. Effectively it is | |
/// shader-toy written in bevy. | |
/// | |
/// This revision adds a post-processing node to the RenderGraph to | |
/// execute the shader. Thanks to @Jasmine on the bevy discord for | |
/// suggesting I take a second look at the bevy post-processing example | |
/// |
// Website you intended to retrieve for users. | |
const upstream = 'api.openai.com' | |
// Custom pathname for the upstream website. | |
const upstream_path = '/' | |
// Website you intended to retrieve for users using mobile devices. | |
const upstream_mobile = upstream | |
// Countries and regions where you wish to suspend your service. |
import taichi as ti | |
from taichi.math import * | |
ti.init(arch=ti.gpu, default_ip=ti.i32, default_fp=ti.f32) | |
image_resolution = (512, 512) | |
image_buffer = ti.Vector.field(4, float, image_resolution) | |
image_pixels = ti.Vector.field(3, float, image_resolution) | |
SCREEN_PIXEL_SIZE = 1.0 / vec2(image_resolution) | |
aspect_ratio = image_resolution[0] / image_resolution[1] |
{-# LANGUAGE TypeSynonymInstances #-} | |
data Dual d = D Float d deriving Show | |
type Float' = Float | |
diff :: (Dual Float' -> Dual Float') -> Float -> Float' | |
diff f x = y' | |
where D y y' = f (D x 1) | |
class VectorSpace v where | |
zero :: v |
Directory structure:
.
├── papers
│ ├── Title - Author.pdf
│ └── Title - Author.pdf
├── notes
│ ├── 2022-04-10
│ │ ├── note.tex
This gist is my attempt to list all projects providing proof automation for Agda.
When I say tactic, I mean something that uses Agda's reflection to provide a smooth user experience, such as the solveZ3
tactic from Schmitty:
_ : ∀ (x y : ℤ) → x ≤ y → y ≤ x → x ≡ y
_ = solveZ3
Scalable Vector Extensions (SVE) is ARM’s latest SIMD extension to their instruction set, which was announced back in 2016. A follow-up SVE2 extension was announced in 2019, designed to incorporate all functionality from ARM’s current primary SIMD extension, NEON (aka ASIMD).
Despite being announced 5 years ago, there is currently no generally available CPU which supports any form of SVE (which excludes the [Fugaku supercomputer](https://www.fujitsu.com/global/about/innovation/
The fact that 1 + 1 is equal to 2 is one of those things that is so obvious it may be hard to justify why. Fortunately mathematicians have devised a way of formalizing arithmetic and subsequently proving that 1 + 1 = 2. Natural numbers are based on the Peano axioms. They are a set of simple rules that define (along with a formal system) what natural numbers are. So in order to prove 1 + 1 = 2 in Rust we first need a formal system capable of handling logic. The formal system that we'll be using is not some random crate, but Rust's type system itself! We will not have any runtime code, instead the type checker will do all the work for us.
First let's go trough the Peano axioms. The first axiom is that "Zero is a natural number". Basically what it says is that zero exists. In order to express that in the type system, we just write:
Pipewire is the replacement for JACK and PulseAudio, Pipewire allows low latency compared to any pulseaudio tweaks | |
Here is a short list of what you should do to get the lowest latency in Osu! | |
Higher audio rate equals less latency always, unless you increase your quantum | |
to calculate node latency for your audio device take the quantum size divided by your audio rate | |
so 64/96000 = 0.00066666666 * 1000 = 0.6ms this is 0.6ms node latency | |
To check client latency use pw-top, take the quantum size and the audio rate of the client then use quantum / audio rate * 1000 | |
to get overall latency for the client |