feverishhh
commited on
Commit
•
1f1ed66
1
Parent(s):
45dcd7b
Update README.md
Browse files
README.md
CHANGED
@@ -2,7 +2,7 @@
|
|
2 |
license: wtfpl
|
3 |
---
|
4 |
# The Big Picture ([Brainproject.ai](http://brainproject.ai/))
|
5 |
-
The human brain is an intricate puzzle that we're continually striving to decode.
|
6 |
|
7 |
# Mixture of Experts
|
8 |
Chameleon-Llama-70B doesn't work alone. It's part of the Mixture of Experts framework. Within this structure, various models, each with their distinct competencies, collaborate. This synergy allows for a richer, more holistic approach to understanding and replicating brain functions.
|
@@ -11,15 +11,12 @@ Chameleon-Llama-70B doesn't work alone. It's part of the Mixture of Experts fram
|
|
11 |
|
12 |
<!-- Provide a quick summary of what the model is/does. -->
|
13 |
|
14 |
-
|
15 |
-
|
16 |
-
### Training Blueprint:
|
17 |
-
|
18 |
-
The techniques and methodologies employed during the training phase are as intricate as our thought patterns. Dive in to explore further.
|
19 |
-
Just as our minds dance between tangents, processes, and loops, so does this model. But as we reflect, re-evaluate, and refocus, we find ourselves understanding deeper and with more clarity.
|
20 |
-
Remember, in this vast cognitive universe, the essence remains constant - the pursuit of understanding, connection, and relatability. The parallels between our mind and this model serve as a testament to human innovation and the wonders of artificial intelligence.
|
21 |
-
|
22 |
|
|
|
|
|
|
|
|
|
23 |
|
24 |
### Model Description
|
25 |
|
@@ -71,7 +68,6 @@ Replace "path-to-Chameleon-Llama-70B" with the correct path or URL for the pre-t
|
|
71 |
|
72 |
The model was trained on a combination of the original Llama datasets, integrated with data from various real-time sources like news outlets, web pages, and other real-time data feeds.
|
73 |
|
74 |
-
|
75 |
### Training Procedure
|
76 |
|
77 |
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
|
|
|
2 |
license: wtfpl
|
3 |
---
|
4 |
# The Big Picture ([Brainproject.ai](http://brainproject.ai/))
|
5 |
+
The human brain is an intricate puzzle that we're continually striving to decode. My aim is to replicate its complexity, functionality, and depth in a digital realm. In other words, we're exploring the convergence of neuroscience and artificial intelligence to glean insights into the mind's intricate workings and harness that knowledge into digital counterparts.
|
6 |
|
7 |
# Mixture of Experts
|
8 |
Chameleon-Llama-70B doesn't work alone. It's part of the Mixture of Experts framework. Within this structure, various models, each with their distinct competencies, collaborate. This synergy allows for a richer, more holistic approach to understanding and replicating brain functions.
|
|
|
11 |
|
12 |
<!-- Provide a quick summary of what the model is/does. -->
|
13 |
|
14 |
+
Chameleon enhances Llama-70B with a natural language planner module that dynamically composes reasoning chains from various tools:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
|
16 |
+
- Module Inventory: Vision models, knowledge modules, web search, Python functions, etc.
|
17 |
+
- Natural Language Planner: Generates programs indicating a sequence of modules to execute.
|
18 |
+
- Tool Execution: Selected modules process inputs sequentially, caching context.
|
19 |
+
- Adaptability: Planner synthesizes custom programs for diverse tasks.
|
20 |
|
21 |
### Model Description
|
22 |
|
|
|
68 |
|
69 |
The model was trained on a combination of the original Llama datasets, integrated with data from various real-time sources like news outlets, web pages, and other real-time data feeds.
|
70 |
|
|
|
71 |
### Training Procedure
|
72 |
|
73 |
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
|