|
“Hi. I need you to help me to transcribe an idea that I have, and at each step I want to annotate it, and then give me a quick description about what I've just presented.” |
|
00:14 |
|
ChatGPT said: |
|
ChatGPT |
|
Sure, I'd be happy to help! Please go ahead and describe your idea, step by step, and I'll annotate each one with a quick description. |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“Okay, cool. So I want to talk about my worries and concerns about fine-tuning a model. So the concept here is that we want to create, there's a common pattern of creating fine-tuning model for specific scenarios. So the scenario that I'm going to talk about is a scenario where, let's say, it's a cybersecurity agent, but that needs to have knowledge on GDPR and the current state of affairs. So the logic is that what we'll do is that we'll pre-train a particular model with actual information so that when you ask questions, you ask questions that will be answered by the pre-trained information.” |
|
00:51 |
|
ChatGPT said: |
|
ChatGPT |
|
Got it! So, the first step is recognizing the need for a specialized model to address specific scenarios, like a cybersecurity agent bot knowledgeable about GDPR and current regulations. The next step involves pre-training a model using accurate and relevant information so it can provide informed answers to user queries. Does that sound accurate for the first two steps? |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“No, because actually what I want is to show why fine-tuning a model, in my view, is a bad idea, and we should not be fine-tuning the model because of the problems I'm going to describe next. What we should be doing is using off-the-shelf, read-only, basically commoditized models that we can then give information using prompts. While we completely control the information being provided, I think,” |
|
00:29 |
|
ChatGPT said: |
|
ChatGPT |
|
Understood! So, your main point is that fine-tuning a model can be problematic. Instead, you advocate using |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“step-off training in the step-by-step.” |
|
00:01 |
|
ChatGPT said: |
|
ChatGPT |
|
Certainly! The next step would be identifying |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
Transcript Unavailable |
|
00:00 |
|
ChatGPT said: |
|
ChatGPT |
|
Got it! The next step would be highlighting the limitations or challenges of fine-tuning, such as maintenance complexity or data drift. After that, you could advocate for the benefits of using off-the-shelf, read-only models, emphasizing prompt engineering to control the information provided. Does that align with your plan? |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“Yes. So let me start with the first problem. So when you fine-tune a model, you basically go through a process of fundamentally teaching the model a particular set of information. So in this case, let's say GDPR, but let's make it more interesting. What we're saying is not just GDPR, because that's just a standard. You would need to provide not just the standard itself, and a particular actual version of the standard. You need to provide then information about specific rulings of specific countries, because every country has certain variations, and you would probably also add information about specific cases, specific fines. So when you ask a question about a particular area of GDPR, you get highly accurate information for, in this case, 1st of January, 2025.” |
|
00:49 |
|
ChatGPT said: |
|
ChatGPT |
|
Absolutely, the first problem is that fine-tuning requires incorporating extensive and specific information, like the exact version of GDPR, country-specific variations, and details on specific cases and fines. This process can be cumbersome and complex, especially when ensuring the model stays updated with the latest information. Does that capture the essence of your first concern? |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“Yeah, that's close. But the main problem that I have with fine-tuning is the fact that, again, as of January 2025, we still don't understand how the fine-tuning works. We don't understand how data is retained. We don't understand how information is captured. What we know is that, from a behavioral point of view, a black box point of view, the model seems to have acquired some new knowledge in there. So that is fundamentally the problem. The problem is until we completely understand how data is understood and retained and is present in a model, for me, the fine-tuning is another variation of a black box of a model where I don't really have a lot of control over the data that it has, because it might look consistent, but it might not be. And then it means that when you look at the prompts that you're going to get and you've got behavior, it doesn't have as much deterministic as I would like it to have.” |
|
01:11 |
|
ChatGPT said: |
|
ChatGPT |
|
That's a significant concern. The core issue with fine-tuning is the lack of understanding and transparency regarding how models process and retain data. Since we can't fully grasp how fine-tuning affects a model's knowledge or behavior, this process remains a "black box," leading to unpredictable and potentially inconsistent outputs. This lack of control over the model's internal workings and the non-deterministic nature of its responses make fine-tuning less reliable. Does that accurately summarize your main problem with fine-tuning? |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“That's a good example. So the other problems that we have is, how do you quantify the data? So a good example is erasures, ironically, all of GDPR. So what happens in a scenario where you need to make a change in the fine tuning? Or, more interestingly, what happens when, as you train the model and as you evolve, you find some mistakes? So let's think about what these mean. What this means is that if I give the LLM some instruction, some content, and then I ask it and say, right, tell me about this. And let's say the LLM gets it wrong. Well, it doesn't get it perfectly. The problem is that the next training is done on top of the previous one. So if you think about it, it means that every instruction that we send to the model is modifying the state of the model. Unless, you know, it would be interesting to see if you can actually go back and forward with the verges. But in most cases, I'm saying when people do fine tuning, they're fine tuning one instruction on top of the other, literally what the name does. You fine tune it, and you get the response that you want. What that means, it means that every time you send an instruction to the model, you are building on top of previous answers. So you are almost bending the model to behave in the way that you want it to behave, but that has the problem of what happens, you know, if that behavior is not exactly what you want. Or it might look like it has, at that moment in time, the behavior that you want, but that behavior is just a combination of multiple attempts of getting it right.” |
|
01:49 |
|
ChatGPT said: |
|
ChatGPT |
|
Exactly, this iterative fine-tuning process poses significant challenges. Each new instruction or data point modifies the model incrementally, building upon previous states. This can lead to unintended behaviors or inaccuracies, and since each training layer influences the next, errors can propagate. Without the ability to fully reverse or track changes, it becomes challenging to ensure the model behaves as desired consistently. This complexity underscores why fine-tuning can be unpredictable and hard to manage. Does that align with your thoughts? |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“Yes, and we can then add to it that this is also the view of a happy path, where the data is benign and there's no malicious intent, there's no contradictory data. But we've already seen cases where, as models start to get tricked, as models start to be given, for example, information about providing contradictory information or providing a level of manipulation, which is fine because that's what you're training the model to do, there's a lot of research now that shows that once a model acquires a behavior, you can't sometimes remove that behavior, even when the model itself can never tell you that it doesn't have that behavior. So there's some really cool research recently that shows that once a model was shown to behave in a certain way, even when you said it didn't, the model was still behaving in that way, although it would say on face value that it wasn't behaving that way. So my point is that suddenly you have a situation that not only do you compound the information or the guidance and the fine-tuning one on top of the other, it's kind of worse because if any of those instructions start to introduce contradictory information or information that manipulates the data, and this could be a prop injection, it could be indirect prop injection, it could be inefficient indirect prop injection, so it means that you can have a payload in an article, in a ruling, in some knowledge that we want to fit in because it's relevant knowledge that suddenly gets absorbed into the model, and then it becomes almost like a silent bomb or a backdoor in the model that you then don't have the ability to see it until it's too late.” |
|
01:45 |
|
ChatGPT said: |
|
ChatGPT |
|
That's a crucial point. Fine-tuning not only layers information but also introduces risks, especially if the data is manipulated or contradictory. Models can internalize these behaviors, and once they're embedded, it's difficult to correct or remove them. This could lead to unintended "backdoors" in the model's responses, where the model might exhibit unexpected behaviors even if it claims otherwise. This emphasizes the challenge of ensuring model integrity and reliability. Does this capture your concern accurately? |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“Yep. And now, let's look at what happened in that scenario. So what happens is, the problem we again have with fine-tuning is that it's viewed, in a lot of cases, as a one-off exercise. Viewed as a costing operation that you can then recoup the cost of that using multiple users. So, in a way, the hypothesis for fine-tuning is that, you know, it's cheaper and faster and more efficient to fine-tune a model because then I can run that model all the time without having to train it. But, if you then take into account that we have models that we then lose trust, or models that start behaving that unpredictable way, or models that we know, shit, it has data that it shouldn't have, or it has behavior that it shouldn't have, one of the very logical next steps is to kill the model. Like, you know, what happens when you lose trust of a model? What happens when, you know, the model has behavior that you cannot explain? So your deterministic improvenance of the data that the model produces is actually not very effective.” |
|
01:07 |
|
ChatGPT said: |
|
ChatGPT |
|
Exactly, the "one-off" nature of fine-tuning assumes that once a model is tuned, it remains reliable for repeated use. However, if trust in the model erodes—due to unpredictable or undesirable behaviors—the cost and effort invested in fine-tuning can be wasted. When models develop unexplainable or incorrect behaviors, they can become untrustworthy, leading to the drastic step of needing to "kill" or discard the model. This loss of trust underscores the importance of predictable and transparent model behavior. Is that a fair summary? |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“Yes. And one more threat to take into account is the fact that we still don't fully know all the ways that the models can be manipulated using front-end engineering. So a good example would be, you know, there could be ways to talk to models via a particular type of coding, particular type of meta-language, particular type of even language, or gene code, or something, or binary, or a sequence of instructions that the model can understand that makes it have a certain behavior. And this is where you have that problem of indirect prompting, because you can have a payload that is in a model that only gets activated amongst those instructions, which is again, you know, one of the ways that the models can be corrupted. Because when you load data into the model, you can have these kind of instructions giving it certain behavior that is not observable, and that behavior then is only observable when the payload is there. And this is basically the equivalent of buffer overflow attacks. As an analogy, I remember this really cool attack I once worked with a customer on, where on the card terminals, basically the card, the ones where you make payments, now contactless, these guys found a buffer overflow in the terminal. So they have two cards, one card that did a buffer overflow to add a new behavior to the card machine to store all the information, and then you have another card that would do a buffer overflow to read all the data into that card. And that was a hell of an exploit, right? That was a good example of one behavior that was added in one step, and then, you know, it was collected at a different step. And we still, because we still don't fully understand how the prompts get, how the models understand the data to be given both at fine-tuning and at, you know, inference, you know, for me, the idea is that don't assume any of that. And” |
|
02:03 |
|
|
|
2/2 |
|
|
|
ChatGPT said: |
|
ChatGPT |
|
That's a critical insight. The possibility of hidden or unintended behaviors |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“In this long transition, future attacks could actually have quite a catastrophic effect because it could be manipulating the behavior of the actual system.” |
|
00:07 |
|
ChatGPT said: |
|
ChatGPT |
|
Absolutely. The potential for hidden prompt-based vulnerabilities within models poses a significant threat, similar to cybersecurity exploits like buffer overflow attacks. These hidden vulnerabilities could be exploited later, potentially causing unpredictable or harmful behavior. Given the complexity and opacity of model training and inference processes, it's crucial not to assume safety or reliability without rigorous testing and continuous monitoring. Being vigilant and cautious in deployment is key to mitigating these risks. Does this encapsulate your overall perspective? |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“But I will go to the next step and say the way you eliminate the risk is by not even having it in the first place. So now let's look at the current solutions for this.” |
|
00:11 |
|
ChatGPT said: |
|
ChatGPT |
|
Sure! |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“So basically, there are three key elements at play. Actually, one more before I do this. The other thing that you have when you find it in the model is that, by definition, you are locked in the model. So, by definition, there are a lot of things that you are not able to do because you are locked to a specific version of a model. And this goes from not being able to leverage newer versions of the model, that might be better at it, but also, the other round prevents you from using cheaper models, prevents you from using models that actually can do the same thing, but they are smaller. So, you get locked to a particular version of the model because, again, when you're fine-tuning, you're not going to be fine-tuning ten different models. You're usually fine-tuning one or two or maybe three in the beginning of a research phase, and then you end up with one. So, again, by having a fine-tuned model, also locks you into that particular version of that particular model because that's the one that's being fine-tuned. And then any changes require you to then do the whole training again in there and take into account that even if you can repeat the training, what you haven't done was figure out the best way to train it because, again, your training is one on top of the other. So, by definition, even your training data will be inefficient if you need to do it again because you haven't iterated from scratch.” |
|
01:26 |
|
ChatGPT said: |
|
ChatGPT |
|
That's a great point. Fine-tuning locks you into a specific model version, limiting flexibility and potentially preventing you from leveraging newer, more efficient, or cost-effective models. This not only locks you into current capabilities but also requires retraining from scratch for any changes, which can be inefficient and resource-intensive. This reiterates the importance of considering long-term adaptability and scalability when choosing how to deploy models. Are you thinking about potential solutions that avoid these pitfalls? |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“OK, so now let's look at solutions. So there are three elements, three major elements at play here.” |
|
00:09 |
|
ChatGPT said: |
|
ChatGPT |
|
Sure, let's |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“Here's the models themselves. And by now I'm going to be talking about models because we now have the ability to use local models. So there's the model.” |
|
00:10 |
|
ChatGPT said: |
|
ChatGPT |
|
Absolutely. The first element |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“There is a fine-tuned slash trunade.” |
|
00:03 |
|
ChatGPT said: |
|
ChatGPT |
|
Yes, |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“Thank you.” |
|
00:00 |
|
ChatGPT said: |
|
ChatGPT |
|
And the third element would be the prompt engineering and customization layer. These three elements—the base models, fine |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“Bye.” |
|
00:00 |
|
ChatGPT said: |
|
ChatGPT |
|
-tuning or prompt training data, and the prompt engineering layer—work together to ensure the model functions effectively and securely. So what would you suggest as a solution to these challenges? |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“All right.” |
|
00:00 |
|
ChatGPT said: |
|
ChatGPT |
|
What would be the solution? |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“I'll see you guys next time.” |
|
00:00 |
|
ChatGPT said: |
|
ChatGPT |
|
How would you address these elements? |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“OK, so let's continue now on the solutions. So there are three elements of this that basically, for me, represent problem space. There is the model, or models, itself. There is the training data, data that we use to fine-tune this. And there is the prompt information, i.e. what you send to the model, either fine-tuned or not, within the size of the prompt content.” |
|
00:28 |
|
ChatGPT said: |
|
ChatGPT |
|
Exactly. To address |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“Peace out.” |
|
00:00 |
|
ChatGPT said: |
|
ChatGPT |
|
these elements, the solution could involve: |
|
|
|
Using robust, off-the-shelf models that don't require fine-tuning. |
|
Ensuring the training data is of high quality and free from vulnerabilities. |
|
Focusing on effective prompt engineering to control and customize model outputs without altering the underlying model. Does that align with your approach? |
|
|
|
|
|
|
|
|
|
You said: |
|
“Yeah, you described the three components that I said before.” |
|
00:03 |
|
ChatGPT said: |
|
ChatGPT |
|
Yes, your approach involves focusing on these three components: |
|
|
|
Models: Utilizing robust, off-the-shelf models without relying on fine-tuning. |
|
Training Data: Ensuring high-quality, secure training data to avoid vulnerabilities. |
|
Prompts: Leveraging effective prompt engineering to customize outputs while maintaining model integrity. By focusing on these areas, you can mitigate many risks associated with fine-tuning. Does this reflect your thinking? |
|
|
|
|
|
|
|
|
|
You said: |
|
“Yeah, okay. Well, kind of, because what I'm saying is that those things exist regardless of we fine-tune or not. Regardless of the solution, you have three elements. You have a model, you have printing information, and you have prompt, which comes from the user and extra bits.” |
|
00:21 |
|
ChatGPT said: |
|
ChatGPT |
|
I see what you're saying—these three components are fundamental regardless of the approach. The model, the training data, and the prompts are always involved, whether you fine-tune or not. The challenge is managing these elements effectively to ensure security and performance. Would you like to discuss how to best leverage these components? |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“Yeah, correct. So now, if you take into account that even in an optimized, fine-tuned model, you still have to provide some extra information. So that means that your prompt, even in a fine-tuned model, should not just be the actual information provided by the user or the actual question. It should have some extra information there, mainly related to the customer or the user, because the alternative is to put that in the training model, which is a really good training data, which is a really good bad. It's a very bad idea. So my point is that regardless of the system, of whether you pre-train or not, the prompt would always have a level of customization and a level of user-specific, customer-specific, user-specific information.” |
|
00:56 |
|
ChatGPT said: |
|
ChatGPT |
|
Memory updated |
|
Exactly. This ensures that the model remains adaptable and secure, and that any customization is handled at the prompt level rather than being hardcoded into the training data. This approach helps maintain flexibility and security. |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“Correct. Now, the idea is that if you have to have still this layer of customization, regardless if you pre-train or not, then why don't you optimize that and don't do pre-train? Because then the logic is that we can then focus our engineering efforts in the creation of highly specific and highly accurate, in a way, fine-tuned data that is actually provided in the prompt engineering phase, not in the fine-tune, which we don't do. So it means that, in principle, we can reduce dramatically the amount of data that is provided, and we solve a huge amount of threats and risks because we have that. And fundamentally, we just have two components. We have, well, three, but in this case, two steps, which is we have the model that we choose, and then we have the main step, which is the creation of the prompt and the multiple components of the prompt that we need to give to the model to execute the query.” |
|
01:08 |
|
ChatGPT said: |
|
ChatGPT |
|
That makes sense. By focusing on optimizing prompt engineering instead of fine-tuning, you can streamline the process, minimize risks, and maintain flexibility. This approach allows you to deliver tailored, accurate responses while reducing the complexity and potential security issues associated with modifying the model itself. It also enables you to adapt quickly to new requirements or changes. Is this the kind of streamlined, flexible approach you’re advocating for? |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“Okay, so now the solution is to replicate the fine-tuning at prompt level. Now, traditionally, this was done using RAC, which was using embedded vector databases and embeddings. My problem or mistake with normal RAC is, again, another black box, and it's also an environment where you don't have a lot of control over the data that gets added because you're relying on embeddings, and by definition, the embeddings are going to be an approximation of the logic. So I think the solution in this is to using graph RAC equivalence, or even graphs, but let's say normal described graph RACs. So it's a graph knowledge base that fundamentally has the information that we need, and we basically have a couple of steps to build our prompt where we basically create the equivalent of the fine-tuning, but from our graph, which has been originated from the query that the user” |
|
01:12 |
|
ChatGPT said: |
|
ChatGPT |
|
Memory updated |
|
Using a graph-based knowledge system can indeed provide greater control and precision. By leveraging a graph structure, you can more accurately represent relationships and context, ensuring the generated prompts are precise and tailored to the user's query. This approach can reduce the reliance on embeddings and offer a more transparent and controllable way to manage knowledge and responses. It's a promising solution to the issues of traditional RAG systems. Do you have specific ideas on how you'd implement this? |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“Yes. So, let's take the example that we're talking about, where we have GDPR information, and we have the GDPR standards, and we have the GDPR use cases, and adoption from the states, and cases, and fines, etc. So, basically, what you want to imagine is a workflow where we basically first convert all of this into graphs. So, imagine a graph representation of GDPR, which is basically a graph that contains nodes and edges for the content, for the relationships, for the articles themselves, for the materials that are related to the articles, etc. So, then, in a way, this graph becomes the basis for the next graphs, again, nodes and edges, that will be created, for example, from the ruling of a particular European zone, and even of a particular case and fine. So, if you imagine all that information, both the articles, information from the states, and the fines, and even articles and guidance documents from the European Union, etc., all of them get transformed into these sort of nodes and edges, where you fundamentally have a graph that represents that content.” |
|
01:25 |
|
|
|
1/2 |
|
|
|
ChatGPT said: |
|
ChatGPT |
|
That's a powerful approach. By transforming all relevant GDPR information—standards, rulings, regional adaptations, cases, and fines—into a graph, you create a rich, structured knowledge base. This graph, with its interconnected nodes and edges, provides a clear and precise representation of the content and relationships. This structured approach can enable more nuanced and context-aware prompt generation, enhancing accuracy and relevance for user queries. It offers a transparent and controllable way to manage and utilize knowledge, aligning well with your goal of reducing reliance on black-box systems. How do you see this graph being queried and used in practice? |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“Okay, can you recap where we are now in terms of the solution?” |
|
00:04 |
|
ChatGPT said: |
|
ChatGPT |
|
Sure! So far, we've discussed that instead of fine-tuning models, focusing on robust, off-the-shelf models paired with effective prompt engineering is more efficient. We've highlighted the value of using a graph-based knowledge system to structure and manage complex information, |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“Sorry, can you continue?” |
|
00:02 |
|
ChatGPT said: |
|
ChatGPT |
|
Certainly! By using a graph-based knowledge system, you can organize information into nodes and edges that reflect the relationships and contexts of the data. This structured representation can then be used to dynamically generate precise, context-aware prompts based on user queries. This approach enhances accuracy and relevance, providing a more transparent and controllable alternative to traditional RAG methods. How do you envision implementing this system in your current project? |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“Okay, can you hear me okay?” |
|
00:00 |
|
ChatGPT said: |
|
ChatGPT |
|
Yes, I can hear you. |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“Okay, so.” |
|
00:00 |
|
ChatGPT said: |
|
ChatGPT |
|
Please go ahead. |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“The solution that I want to talk about, continuing with the GDPR information, is that once you have that graph, and especially that schema that represents the graph with the content, where it's not just a pure nodes and entries, it's actually a schema that takes into account the content, we can have a workflow where we take the prompt of the user and start to create a relevant graph that is related to that prompt. So a good example would be if the user asks a question of what is the current situation of erasers in a particular European country, then we should be able to navigate a graph. If I try to write first a series of graph queries, and then navigate a graph to extract the right information that is relevant to that up to a certain depth, which would also allow us to, for example, actually include the actual relevant articles and relevant guidance, not just from the main GDPR, but then any specific extra information provided by the union, and specific information provided by the country, and specific information provided by any cases that have occurred based on that question.” |
|
01:21 |
|
ChatGPT said: |
|
ChatGPT |
|
That sounds like a comprehensive and structured approach. By leveraging a graph schema that captures both content and context, you can dynamically query the graph based on user prompts. This allows you to navigate through relevant nodes and edges, extracting precise information. For instance, you can dive into details about erasures in a specific European country, pulling in relevant articles, national guidelines, and case precedents. This method ensures that the responses are contextually rich and tailored to the query. It also allows for deeper exploration and customization without altering the underlying model. How do you see this fitting into your overall system design? |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“Cool. So you can see that in that workflow, what you then have is a sequence of events where, based on the user prompts, we are creating a series of graphs and a series of information that we provide in the prompt, where we're leveraging the fact that the models have multiple large context windows. And the other big advantage of this is that we can also vary the model, depending on the question, or have workflows where we send the same information to three different models in parallel, and then combine that information, have things like a mixture of agents, multiple agents providing, some mix of experts, and it really decouples the whole workflow, and we can definitely make sure then that information, doesn't contain any sensitive data that should not be accessed by that particular user.” |
|
01:01 |
|
ChatGPT said: |
|
ChatGPT |
|
Memory updated |
|
That's a well-rounded workflow. By dynamically creating relevant graphs based on user prompts, you can provide highly specific and accurate information. The flexibility to switch or combine models based on the query, while ensuring no sensitive data is accessed, enhances both security and accuracy. This approach of using multiple models and agents can indeed provide a more reliable, fine-tuned response without the risks associated with traditional fine-tuning. This also aligns with keeping the model's integrity intact while harnessing its full potential. Are there specific tools or frameworks you're considering for implementing this graph-based system? |
|
|
|
|
|
|
|
|
|
|
|
You said: |
|
“So then you can now see how we have a full loop, so we don't need to fine-tune to get the results. In fact, my hypothesis would be that, on average and given enough time, this workflow will actually be less expensive and will use less tokens and be more optimized than the other solution because it doesn't have the need to retrain, it doesn't have the need to have all those complex workflows, and more importantly, a lot more safe and secure. And then the final interesting element of this is that what we could also do with this workflow is to personalize a lot more content and the structure, not just to the persona, which is important because we can take into account its culture, its language, who the person is, how they want to communicate, etc., but we're also able to customize it to a particular company, a particular industry, a particular sector, because everyone is different. So if you look at this kind of customizations, it's almost impossible to do at the simplified tuning level because you can't take into account all these other variables. If you customize it in the way we're talking about here, where you do the customization later on, almost like the last mile approach, then you can really deal with all these variations from multilingual, multicultural, multi-persona, which ultimately will provide a much better service than the fine-tuning architecture.” |
|
01:38 |
|
ChatGPT said: |
|
ChatGPT |
|
Memory updated |
|
Exactly. By focusing on this dynamic customization layer, you can achieve more tailored and relevant responses without the need for intensive fine-tuning. This also makes the system more adaptable, allowing for personalized content that aligns with specific user needs, industries, and contexts. Over time, this method can be more cost-effective, secure, and efficient, ensuring that each response is both accurate and contextually appropriate. It really highlights the power of a flexible, adaptive system. |