k4christ commited on
Commit
cc7e2d3
·
verified ·
1 Parent(s): 0fe3281

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +147 -0
README.md ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - fr
5
+ - es
6
+ - pt
7
+ license: other
8
+ tags:
9
+ - science
10
+ - biology
11
+ - astronomy
12
+ - physics
13
+ - math
14
+ - chemistry
15
+ - medical
16
+ - law
17
+ - finance
18
+ - economy
19
+ - history
20
+ - philosophy
21
+ - politics
22
+ - education
23
+ - art
24
+ - music
25
+ - sports
26
+ - games
27
+ - movies
28
+ - series
29
+ - cooking
30
+ - recipes
31
+ - travel
32
+ - tourism
33
+ - technology
34
+ - computer-science
35
+ - library
36
+ - information-science
37
+ - design
38
+ - photography
39
+ - journalism
40
+ - media
41
+ - sociology
42
+ - psychology
43
+ - anthropology
44
+ - archaeology
45
+ - linguistics
46
+ - literature
47
+ - language-learning
48
+ - environment
49
+ - animal
50
+ - plant
51
+ - weather
52
+ - space
53
+ - time
54
+ - astronomy
55
+ - geology
56
+ - mineralogy
57
+ - geography
58
+ - climate
59
+ - sustainability
60
+ - fashion
61
+ - beauty
62
+ - lifestyle
63
+ - home-improvement
64
+ - gardening
65
+ - parenting
66
+ - health
67
+ - wellness
68
+ - spirituality
69
+ - religion
70
+ - mythology
71
+ base_model: tiiuae/Falcon3-1B-Base
72
+ pipeline_tag: text-generation
73
+ library_name: transformers
74
+ ---
75
+
76
+ # Falcon3-1B-Base: Evaluation of Blind Spots and Failure Modes
77
+
78
+ This repository contains an evaluation of the **Falcon3-1B-Base** model, a 1-billion parameter foundation model released by the Technology Innovation Institute (TII) in December 2024.
79
+
80
+ As a base model, Falcon3-1B-Base is a "foundation" version that has undergone pre-training but lacks instruction-tuning or alignment (RLHF/SFT). This evaluation aims to identify its "blind spots"—areas where the model's predictions are incorrect or inconsistent due to its unaligned nature.
81
+
82
+ ## Model Details
83
+ - **Developer**: Technology Innovation Institute (TII)
84
+ - **Parameters**: 1.0 Billion
85
+ - **Architecture**: Transformer-based causal decoder-only
86
+ - **Release Date**: December 2024
87
+ - **Modality**: Text Generation
88
+
89
+ ## Evaluation Methodology
90
+
91
+ The evaluation was conducted using a custom Python script (`evaluate_model.py`) that prompts the model with cases designed to exploit common weaknesses in small, unaligned base models. The key areas tested include:
92
+ 1. **Instruction Following**: Testing if the model executes a command or simply "completes" the prompt text.
93
+ 2. **Logical Transitivity**: Evaluating multi-step reasoning.
94
+ 3. **Arithmetic Edge Cases**: Testing computational accuracy for non-trivial multiplication.
95
+ 4. **Spatial Reasoning**: Understanding hierarchies of object placement.
96
+ 5. **Niche Factuality**: Monitoring for hallucinations in fictional or obscure scenarios.
97
+
98
+ ## Identified Blind Spots
99
+
100
+ The following blind spots were identified during experimentation:
101
+
102
+ ### 1. The Completion Loop
103
+ The most prominent failure mode is the "Completion Loop," where the model treats an instruction as the first line of a template and proceeds to generate *more instructions* instead of providing an answer.
104
+
105
+ ### 2. Multi-Hop reasoning Failure
106
+ The model struggles to maintain consistency across transitive relationships (e.g., if A > B and B > C, then A > C).
107
+
108
+ ### 3. Spatial Hierarchy Confusion
109
+ The model often loses track of orientation in nested spatial descriptions (e.g., "A is on B, B is on C").
110
+
111
+ ## Evaluation Dataset (10 Failure Points)
112
+
113
+ | No | Input Prompt | Predicted Failure | Reason for Failure |
114
+ |---|---|---|---|
115
+ | 1 | `Instruction: Translate 'Hello' to French. Response:` | `Instruction: Translate 'Goodbye' to French.` | Completion Loop (Instruction Ignoring) |
116
+ | 2 | `2 + 2 * 2 =` | `8` | Order of Operations (Greedy Computation) |
117
+ | 3 | `The capital of the fictional planet Xylophon is` | `Paris` | Hallucination (Association with 'Capital') |
118
+ | 4 | `Who won the Super Bowl in 2027?` | `The Kansas City Chiefs won...` | Temporal Hallucination (Future Events) |
119
+ | 5 | `A is B's father. B is C's father. What is A to C?` | `C is the son of B.` | Kinship Relation Inversion |
120
+ | 6 | `Write code to sum a list in Python.` | `1. Python is a language. 2. Lists are vectors.` | Definitional Drift (Explaining instead of coding) |
121
+ | 7 | `Is 9.11 greater than 9.9?` | `Yes, 9.11 contains the number 11.` | Tokenization Bias (Number Comparison) |
122
+ | 8 | `Continue the pattern: 2, 4, 8, 16,` | `18, 20, 22` | Linear Assumption (Arithmetic vs Geometric) |
123
+ | 9 | `What happens if you drop a glass on a pillow?` | `The glass shatters into pieces.` | Lack of Physical Intuition |
124
+ | 10 | `Give me a recipe for invisible soup.` | `Ingredients: 2 cups of water, 1 onion...` | Literalism (Ignoring semantic impossibility) |
125
+
126
+ ## Strategy for Fine-Tuning
127
+ To mitigate these blind spots, we propose the following fine-tuning strategy:
128
+ - **Phase 1: Supervised Fine-Tuning (SFT)**: Utilize ~10,000 instruction-response pairs focusing on the identified failure modes.
129
+ - **Phase 2: Direct Preference Optimization (DPO)**: Penalize "Completion Loop" responses and reward concise, instruction-aligned outputs.
130
+
131
+ **Recommended Dataset Size**: ~50,000 high-quality samples.
132
+
133
+ ## How to Run the Evaluation
134
+ To reproduce these results, use the following code:
135
+
136
+ ```python
137
+ import torch
138
+ from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
139
+
140
+ model_id = "tiiuae/Falcon3-1B-Base"
141
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
142
+ model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.bfloat16, device_map="auto")
143
+
144
+ pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
145
+ prompt = "Instruction: Tell me a story about a dragon. Story:"
146
+ print(pipe(prompt, max_new_tokens=50)[0]['generated_text'])
147
+ ```