Skip to content

Instantly share code, notes, and snippets.

View CoffeeVampir3's full-sized avatar

Z CoffeeVampir3

View GitHub Profile
@CoffeeVampir3
CoffeeVampir3 / Vulkan.cc
Last active August 22, 2024 11:30
CPP Module Example
export module Vulkan;
import <stdexcept>;
import Validation;
#define GLFW_INCLUDE_VULKAN
#include <GLFW/glfw3.h>
export namespace Vulkan {
auto CreateInstance() {
VkInstance instance;
export module VulkanInstance;
#define GLFW_INCLUDE_VULKAN
#include <GLFW/glfw3.h>
import VulkanValidation;
export namespace Vulkan {
auto CreateInstance() {
VkInstance instance;
@CoffeeVampir3
CoffeeVampir3 / gemma2_format.py
Last active August 10, 2024 13:05
Rolling system prompt
from exllamav2 import(
ExLlamaV2Tokenizer
)
def encode_message(tokenizer: ExLlamaV2Tokenizer, role: str, message: str) -> list:
tokens = []
start_token = tokenizer.single_id("<start_of_turn>")
end_token = tokenizer.single_id("<end_of_turn>")
tokens.append(start_token)
@CoffeeVampir3
CoffeeVampir3 / burnt_wick.rs
Last active July 22, 2024 20:53
candle model stream
use candle_transformers::models::quantized_llama as model;
use candle_transformers::generation::{LogitsProcessor, Sampling};
use candle_core::quantized::{gguf_file};
use candle_core::Tensor;
pub use candle_core::Device;
pub use tokenizers::Tokenizer;
pub struct StreamableModel {
model: model::ModelWeights,
tokenizer: tokenizers::Tokenizer,
@CoffeeVampir3
CoffeeVampir3 / train.ipynb
Last active July 9, 2024 08:14
Weird Initial Latent Issue.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
#pragma once
#include <coroutine>
#include <utility>
#include <functional>
struct NecroAiState;
struct NecroAiState
{
struct promise_type;
let cube_mesh_handle: Handle<Mesh> = meshes.add(Rectangle::new(6., 9.));
let cube_mesh_handle2: Handle<Mesh> = meshes.add(Rectangle::new(6., 9.));
commands.spawn((
SpatialBundle {
transform: Transform::default(),
..default()
},
ParentMarker,
// On::<Pointer<Drag>>::listener_component_mut::<Transform>(|drag, transform| {
@CoffeeVampir3
CoffeeVampir3 / example.rs
Created May 30, 2024 01:41
Rust Candle Inference Examples
use std::io::Write;
use tokenizers::Tokenizer;
use candle_core::quantized::{gguf_file};
use candle_core::Device;
use candle_transformers::generation::{LogitsProcessor, Sampling};
use candle_transformers::models::quantized_llama as model;
use model::ModelWeights;
use burnt_wick::streamable_model::StreamableModel;
fn load_model_and_tokenizer(
use std::io::Write;
use tokenizers::Tokenizer;
use candle_core::quantized::{gguf_file};
use candle_core::Tensor;
use candle_core::Device;
use candle_transformers::generation::{LogitsProcessor, Sampling};
use candle_transformers::models::quantized_llama as model;
use candle_examples::token_output_stream::TokenOutputStream;
use model::ModelWeights;
use std::collections::HashMap;
from flask import Flask, render_template
import torch
from flask_socketio import SocketIO, emit
from generation.make_instruct import get_generator_func
from generation.exllama_generator_wrapper import encode_message, encode_system, encode_header, encode_header_prefilled, encode_message_with_eot, encode_completion
from collections import deque
import time
import os,sys
sys.path.append(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))