Skip to content

Commit cbbe1fc

Browse files
Update README.md
1 parent f0a49ea commit cbbe1fc

File tree

1 file changed

+5
-5
lines changed

1 file changed

+5
-5
lines changed

README.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -174,21 +174,21 @@ The API returns a JSON object containing evaluation details.
174174
}
175175
```
176176

177-
# Models
177+
## Models
178178

179-
# Lemone-Embed: A Series of Fine-Tuned Embedding Models for French Tax
179+
### Lemone-Embed: A Series of Fine-Tuned Embedding Models for French Tax
180180

181181
These series are made up of 7 models: 3 basic models of different sizes trained for 1 epoch, 3 models trained for 2 epochs forming the Boost series, and Pro models with non-RoBERTa architectures.
182182

183183
These sentence transformer models, specifically designed for French taxation, have been fine-tuned on datasets comprising 43 million tokens, integrating blends of semi-synthetic and fully synthetic data generated by GPT-4 Turbo and Llama 3.1 70B. These datasets have been further refined through evol-instruction tuning and manual curation.
184184

185-
### Training Hardware
185+
#### Training Hardware
186186
- **On Cloud**: No
187187
- **GPU Model**: 1 x NVIDIA H100 NVL
188188
- **CPU Model**: AMD EPYC 9V84 96-Core Processor
189189
- **RAM Size**: 314.68 GB
190190

191-
## Lemone-Router: A Series of Fine-Tuned Classification Models for French Tax
191+
### Lemone-Router: A Series of Fine-Tuned Classification Models for French Tax
192192

193193
Lemone-router is a series of classification models designed to produce an optimal multi-agent system for different branches of tax law. Trained on a base of 49k lines comprising a set of synthetic questions generated by GPT-4 Turbo and Llama 3.1 70B, which have been further refined through evol-instruction tuning and manual curation and authority documents, these models are based on an 8-category decomposition of the classification scheme derived from the Bulletin officiel des finances publiques - impôts :
194194

@@ -221,7 +221,7 @@ It achieves the following results on the evaluation set:
221221
- Loss: 0.4734
222222
- Accuracy: 0.9191
223223

224-
### Training Hardware
224+
#### Training Hardware
225225
- **On Cloud**: No
226226
- **GPU Model**: 1 x NVIDIA H100 NVL
227227
- **CPU Model**: AMD EPYC 9V84 96-Core Processor

0 commit comments

Comments
 (0)