mandelakori commited on
Commit
59fe8fd
1 Parent(s): 7ec777a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -5
README.md CHANGED
@@ -11,10 +11,8 @@ AISAK, short for Artificially Intelligent Swiss Army Knife, is a state-of-the-ar
11
 
12
  - **Model Name**: AISAK
13
  - **Version**: 1.0
14
- - **Model Architecture**: Mixture of Experts (MoE)
15
- - **Specialization**: AISAK is structured upon the principles of the Mixture of Experts (MoE) architecture, meticulously crafted to emulate the success of the renowned https://huggingface.co/mistralai/Mixtral-8x7B-v0.1 model. Its architecture is ingeniously segmented into distinct expert modules, each adept at discerning specific patterns and features inherent within the input data.
16
- - **Gating Mechanism**: A dynamic gating mechanism intelligently selects and combines the outputs of these experts based on the input data, enhancing adaptability and performance.
17
- - **Performance Comparison**: While AISAK may not boast the same parameter count as the Mistral8x7b model, it maintains a remarkably high and heavily comparable performance level. Through meticulous optimization and leveraging the strengths of the MoE architecture, AISAK achieves results on par with its predecessor, ensuring that it stands as a formidable contender in the realm of artificial intelligence models.
18
 
19
  ### Intended Use:
20
 
@@ -22,7 +20,7 @@ AISAK, conceptualized by Mandela Logan, is intricately crafted for diverse text
22
 
23
  ### Performance:
24
 
25
- AISAK undergoes rigorous testing across diverse input data types, consistently demonstrating superior performance. Its capabilities have proven to outperform and exceed those of various state-of-the-art models such as but not limited to, GPT-3.5 and Llama 2 (70b).
26
 
27
  ### Ethical Considerations:
28
 
 
11
 
12
  - **Model Name**: AISAK
13
  - **Version**: 1.0
14
+ - **Model Architecture**: Transformer
15
+ - **Specialization**: AISAK is structured upon the principles of the Transformer architecture, meticulously crafted to emulate the success of the renowned https://huggingface.co/mistralai/Mistral-7B-v0.1 model. Its architecture is ingeniously segmented into distinct expert modules, each adept at discerning specific patterns and features inherent within the input data.
 
 
16
 
17
  ### Intended Use:
18
 
 
20
 
21
  ### Performance:
22
 
23
+ AISAK undergoes rigorous testing across diverse input data types, consistently demonstrating superior performance. Its capabilities have proven to outperform and exceed those of various state-of-the-art models such as but not limited to, GPT-3.5 and Llama 2's 13b and even 70b parameter model.
24
 
25
  ### Ethical Considerations:
26