Safetensors
English
mistral
LeroyDyer commited on
Commit
89d9646
1 Parent(s): f2740ae

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -54
README.md CHANGED
@@ -7,71 +7,42 @@ datasets:
7
  language:
8
  - en
9
  ---
10
- # SpydazWeb AI React Project:
 
 
11
 
 
12
 
13
- A great motivator said : ME!
14
 
15
- ## "to be sucessful you need to define everything in performable steps each step completed is a success and brings you closer to your end goal , but if your steps are unreachable then you will always fail, hence winners begat winners and losers begat losers. suces is a game of winners ! "
16
 
17
- ## " to grow as a professional you need to define steps that are just outside your reach ! as you acomplish this step you acomplish a milestone and overcome a obsticle creating a stronger skillset: if your tasks are too easy then you will never challenge yourself or improve, hence if you do not challenge your self life will pass you by !"
18
 
 
19
 
20
- ### Leroy Dyer ( 1972-Present)
21
 
 
 
 
 
22
 
23
- ## Model:
 
 
24
 
25
- Trained following the format used in the [ReAct paper](https://arxiv.org/pdf/2210.03629.pdf).\
 
 
26
 
27
- Using te start point of a Merged Chatmodel ( spydazWeb_AI_ChatQA_005/006) = New Base !
 
 
 
28
 
29
- This paradigm Enables for thegenerate of ReAct Agents : such agents are required to perform complex tasks : here we give the agents various methods of thought and action :
30
- the above old paper was used to understand the methods used in training such agents : as you will see below , despite bad english the prompt gives the model the tools and an example of operation :
31
- In fact First i trained the binary yes/no answers without methoology ! and then i trained using the prompt A: this was thier prompt : because it is giving Unknown tools its not great but it fits the dataset:
32
- i later developed this prompt to a new updated version of this loop ! which is a combination of my own and a handcrafted gpt4o prompt : Imporving the loop to be more flexable in approach as well as aligning to my personal agent system :
33
- hence the model has adapted to the new methodologys :
34
-
35
- It it important to give the model new and updated methodolgys to enhance existing talent and methods deployed : the fuctio calliong and api calling models forget the methodology training hence the improved results :
36
- I have found that over prompting a model can reduce your chances of a great result : but Aider and ClaudeEngineer and Mastroe etc all use large prompts and instructions and this is before the actual input is given :
37
- I think the models do waiver because of the massive instructions given : here we try to combat this with examples ( 1 shot / multi shot prompting )
38
-
39
- with large prompts i find that multiple epochs are required , but with no prompting less epochs are required ?
40
- the point of large prompting in training is to give the model more selction in its results and allowing the model to adjust these parameters to fit the response styles given : not to train the inner inforation of the model as it has the information :
41
- these are just methods of extraction: This is truly what fine tuing is : prompt getting results from the model that we wish and presenting methods of extraction :
42
- So in training the prompt can be super large!! as we are givig the example responses for the massive instruction set: this also enables us to reduce our prompt size in model usage even omit the prompt to recieve the same results as the model has been specifically trained to produce results using internal methodologies :
43
- there are so many possible methods that a user could use or request so we train on as many instances of these as possible : i found that without training the model to be a planner it did not generate complete software without errors:
44
- by giving it training on Agency and agetic work flows it may have learned to work as a team , but after training for planning the model had improved in software generation and complete projects which is ther common goal and not simple task training :
45
-
46
-
47
-
48
- the dataset : xz56/react-llama
49
- this dataset was trained First on QA only giving a baseline for the yes/no answer with no thoughts : the basic production prompt was used :
50
- After the same datset was used and trained dislaying thoughts using the Prompt A then Prompt B:
51
- in both instances the model was drawn to match the binary test set ! :
52
-
53
-
54
-
55
- # OBSERVATION LEARNED :
56
- i training i discovered that the model actually checks the response given from the function and compares it to its own determined answer : Hence it is self correcting : it corrects itself and its own thinking : so it has expections of the functions output when it comes to calcualtions :
57
- I have found that the model is confused that it needs a tool to calcuate an answer : the think , reflect, action : observe loop enables for the model to thik correctly :
58
- by offering thought pathways it can determine the correct answer internally by quering itself : ( in the past i used this self ragging techniue: )
59
-
60
- ## quick explanation of self ragging!
61
- with self ragging instead of getting a direct response instead the model first querys itself and then uses its internal quesry to fuel its next response : hence self rag :
62
- So a single shot could atually be a double shot !
63
- i discovered this route by giving the model the response as a tool : so if the model was outputting the final response then it used the final response tool : hence the model could only use tools !
64
- so if the model wanted to think : it could use the think tool ! ie this tool querys itself with the question and t=reterns the response to the model as an answer ! then the model either outputs the final response with the tool or uerys itself again !
65
- ( i did not release this publically of course !) but i trained my models to have this feature given the correct settup !
66
- i found that giving the model a rag it could search the rag itsef for relevant data ! ... even by giving the model an Agent such as Claude engineeer it could use the agent to perform a research and give the model and advanced content to formulate a great query !
67
- oh my gosh ! after the model was performing very well . but it is a tool based model only so on my humble rtx2030 its still a bit slow ! as each response could be a series of internal querys !
68
- so i decided i will put it on back burner for now , but i would add a response to the inner loop to inform the user that the model is thinking or acting so that constant comunicatin between user and model is efeective hence maintaining the conversation !
69
-
70
- i would like a dataset in which the model performs fucntions as well as asks the user for extra information to provide the final response .
71
-
72
-
73
-
74
- training should not be focused on the end target : but the actuall steps required to reach the target :
75
 
76
 
77
  ## Prompt A:
 
7
  language:
8
  - en
9
  ---
10
+ SpydazWeb AI React Project
11
+ Quote for Motivation:
12
+ "Success comes from defining each task in achievable steps. Every completed step is a success that brings you closer to your goal. If your steps are unreachable, failure is inevitable. Winners create more winners, while losers do the opposite. Success is a game of winners!"
13
 
14
+ "To grow as a professional, set goals just beyond your current abilities. Achieving these milestones will not only overcome obstacles but also strengthen your skillset. If your tasks are too easy, you’ll never challenge yourself or improve, and life will pass you by!"
15
 
16
+ Leroy Dyer (1972-Present)
17
 
18
+ Model Overview:
19
 
20
+ The SpydazWeb AI React Project is built upon the SpydazWeb_AI_ChatQA_005/006 merged chat model as the foundation. The model was trained using a methodology inspired by the ReAct paper, which provides a framework for creating ReAct Agents capable of performing complex tasks. This approach equips the model with various methods of thought and action.
21
 
22
+ Training Process:
23
 
24
+ Initial Training:
25
 
26
+ The model was initially trained on binary yes/no questions without any methodology.
27
+ The training began with a simple prompt (Prompt A) that introduced basic functionality, but with room for improvement.
28
+ The model was later enhanced with a new and more flexible prompt, incorporating a handcrafted GPT-4.0 prompt to align with the personalized agent system. This improved the model’s adaptability to new methodologies and tasks.
29
+ Prompt Design:
30
 
31
+ The model was exposed to different prompting strategies, including 1-shot and multi-shot prompting, to combat potential issues with large instruction sets.
32
+ The focus was on providing the model with methods of extracting information rather than merely training it on the information itself.
33
+ Methodology Training:
34
 
35
+ The training emphasized teaching the model to plan and execute complex tasks, such as generating complete software without errors.
36
+ By incorporating agency and workflow concepts, the model learned to collaborate effectively and improved its software development capabilities.
37
+ Key Observations:
38
 
39
+ Self-Correction: The model demonstrated an ability to self-correct by comparing its responses to expected outcomes. This self-check mechanism, especially in calculations, led to more accurate results.
40
+ Internal Querying (Self-RAG): The model was trained to query itself before providing a final response, effectively creating a multi-step internal process for generating more thoughtful and accurate answers. This process is referred to as "self-RAG" (self-retrieval-augmented generation).
41
+ Tool-Based Model: The model’s performance was enhanced by using tools for thinking and reflecting, though this made it slower on hardware like an RTX 2030.
42
+ Future Goals:
43
 
44
+ Dataset Development: The goal is to develop a dataset where the model not only performs functions but also interacts with users to gather additional information for more refined responses.
45
+ Training Focus: Training should prioritize the steps required to achieve a goal rather than the end target itself, ensuring that the model is capable of navigating complex tasks independently.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
46
 
47
 
48
  ## Prompt A: