- 布洛芬(退烧、止痛)
- 对乙酰氨基酚(泰诺)
- 奥司他韦(达菲,需处方)
- 阿莫西林(需处方)
- 红霉素软膏(外用)
Yoav Goldberg, Nov 24, 2024
This piece started with a pair of twitter and bluesky posts:
let's talk about "agents" (in the LLM sense). there's a lot of buzz around "multi-agent" systems where agents collaborate but... i don't really get how it differs from a thinking of a single agent with multiple modes of operation. what are the benefits of modeling as multi-agent?
— (((ل()(ل() 'yoav))))👾 (@yoavgo) November 23, 2024
When loading the LoRA params (that were obtained on a quantized base model) and merging them into the base model, it is recommended to first dequantize the base model, merge the LoRA params into it, and then quantize the model again. This is because merging into 4bit quantized models can lead to some rounding errors. Below, we provide an end-to-end example:
from diffusers import FluxPipeline
import torch
ckpt_id = "black-forest-labs/FLUX.1-dev"
pipeline = FluxPipeline.from_pretrained(
Warning
This is patched as of iOS/iPadOS 18.1 DevBeta 5. If you want to follow this, stay on Beta 4.
Can you tell an LLM "don't hallucinate" and expect it to work? my gut reaction was "oh this is so silly" but upon some reflection, it really isn't. There is actually no reason why it shouldn't work, especially if it was preference-fine-tuned on instructions with "don't hallucinate" in them, and if it a recent commercial model, it likely was.
What does an LLM need in order to follow an instruction? It needs two things:
// ==UserScript== | |
// @name URL Tagger | |
// @namespace http://tampermonkey.net/ | |
// @version 1.5 | |
// @description Tag URLs based on predefined patterns and data | |
// @author Your Name | |
// @match *://*/* | |
// @grant GM_xmlhttpRequest | |
// @grant GM_addStyle | |
// @connect 192.168.x.x |
import inspect | |
from typing import Any, Callable, Dict, List, Optional, Union | |
import numpy as np | |
import torch | |
from transformers import CLIPTextModel, CLIPTokenizer, T5EncoderModel, T5TokenizerFast | |
from diffusers.image_processor import VaeImageProcessor | |
from diffusers.loaders import FluxLoraLoaderMixin | |
from diffusers.models.autoencoders import AutoencoderKL |
def groupedSimilarFilenames(filenames): | |
''' | |
将相似文件名分组\n | |
输出类似格式\n | |
filenames preffix suffix size \n | |
0 cover.jpg NaN NaN 0 \n | |
1 top.png NaN NaN 0 \n | |
2 9.jpg NaN .jpg 0 \n | |
3 015a.jpg 0 a.jpg 1 \n |
from diffusers import StableDiffusion3Pipeline | |
from transformers import T5EncoderModel | |
import torch | |
import time | |
import gc | |
def flush(): | |
gc.collect() | |
torch.cuda.empty_cache() |