Skip to content

Instantly share code, notes, and snippets.

View CoffeeVampir3's full-sized avatar

Z CoffeeVampir3

View GitHub Profile
@CoffeeVampir3
CoffeeVampir3 / quixel.js
Created September 18, 2024 02:05
quixeling
(await (async (startPage = 0, autoClearConsole = true) => {
const getCookie = (name) => {
const value = `; ${document.cookie}`;
const parts = value.split(`; ${name}=`);
if (parts.length === 2) return parts.pop().split(';').shift();
}
const callCacheApi = async (params = {}) => {
const defaultParams = {
@CoffeeVampir3
CoffeeVampir3 / Vulkan.cc
Last active August 22, 2024 11:30
CPP Module Example
export module Vulkan;
import <stdexcept>;
import Validation;
#define GLFW_INCLUDE_VULKAN
#include <GLFW/glfw3.h>
export namespace Vulkan {
auto CreateInstance() {
VkInstance instance;
export module VulkanInstance;
#define GLFW_INCLUDE_VULKAN
#include <GLFW/glfw3.h>
import VulkanValidation;
export namespace Vulkan {
auto CreateInstance() {
VkInstance instance;
@CoffeeVampir3
CoffeeVampir3 / gemma2_format.py
Last active August 10, 2024 13:05
Rolling system prompt
from exllamav2 import(
ExLlamaV2Tokenizer
)
def encode_message(tokenizer: ExLlamaV2Tokenizer, role: str, message: str) -> list:
tokens = []
start_token = tokenizer.single_id("<start_of_turn>")
end_token = tokenizer.single_id("<end_of_turn>")
tokens.append(start_token)
@CoffeeVampir3
CoffeeVampir3 / burnt_wick.rs
Last active July 22, 2024 20:53
candle model stream
use candle_transformers::models::quantized_llama as model;
use candle_transformers::generation::{LogitsProcessor, Sampling};
use candle_core::quantized::{gguf_file};
use candle_core::Tensor;
pub use candle_core::Device;
pub use tokenizers::Tokenizer;
pub struct StreamableModel {
model: model::ModelWeights,
tokenizer: tokenizers::Tokenizer,
@CoffeeVampir3
CoffeeVampir3 / train.ipynb
Last active July 9, 2024 08:14
Weird Initial Latent Issue.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
#pragma once
#include <coroutine>
#include <utility>
#include <functional>
struct NecroAiState;
struct NecroAiState
{
struct promise_type;
let cube_mesh_handle: Handle<Mesh> = meshes.add(Rectangle::new(6., 9.));
let cube_mesh_handle2: Handle<Mesh> = meshes.add(Rectangle::new(6., 9.));
commands.spawn((
SpatialBundle {
transform: Transform::default(),
..default()
},
ParentMarker,
// On::<Pointer<Drag>>::listener_component_mut::<Transform>(|drag, transform| {
@CoffeeVampir3
CoffeeVampir3 / example.rs
Created May 30, 2024 01:41
Rust Candle Inference Examples
use std::io::Write;
use tokenizers::Tokenizer;
use candle_core::quantized::{gguf_file};
use candle_core::Device;
use candle_transformers::generation::{LogitsProcessor, Sampling};
use candle_transformers::models::quantized_llama as model;
use model::ModelWeights;
use burnt_wick::streamable_model::StreamableModel;
fn load_model_and_tokenizer(
use std::io::Write;
use tokenizers::Tokenizer;
use candle_core::quantized::{gguf_file};
use candle_core::Tensor;
use candle_core::Device;
use candle_transformers::generation::{LogitsProcessor, Sampling};
use candle_transformers::models::quantized_llama as model;
use candle_examples::token_output_stream::TokenOutputStream;
use model::ModelWeights;
use std::collections::HashMap;