diff --git "a/checkpoint-1220/README.md" "b/checkpoint-1220/README.md" new file mode 100644--- /dev/null +++ "b/checkpoint-1220/README.md" @@ -0,0 +1,2073 @@ +--- +base_model: microsoft/deberta-v3-small +library_name: sentence-transformers +metrics: +- pearson_cosine +- spearman_cosine +- pearson_manhattan +- spearman_manhattan +- pearson_euclidean +- spearman_euclidean +- pearson_dot +- spearman_dot +- pearson_max +- spearman_max +- cosine_accuracy +- cosine_accuracy_threshold +- cosine_f1 +- cosine_f1_threshold +- cosine_precision +- cosine_recall +- cosine_ap +- dot_accuracy +- dot_accuracy_threshold +- dot_f1 +- dot_f1_threshold +- dot_precision +- dot_recall +- dot_ap +- manhattan_accuracy +- manhattan_accuracy_threshold +- manhattan_f1 +- manhattan_f1_threshold +- manhattan_precision +- manhattan_recall +- manhattan_ap +- euclidean_accuracy +- euclidean_accuracy_threshold +- euclidean_f1 +- euclidean_f1_threshold +- euclidean_precision +- euclidean_recall +- euclidean_ap +- max_accuracy +- max_accuracy_threshold +- max_f1 +- max_f1_threshold +- max_precision +- max_recall +- max_ap +pipeline_tag: sentence-similarity +tags: +- sentence-transformers +- sentence-similarity +- feature-extraction +- generated_from_trainer +- dataset_size:32500 +- loss:GISTEmbedLoss +widget: +- source_sentence: Fish hatch into larvae that are different from the adult form of + species. + sentences: + - Fish hatch into larvae that are different from the adult form of? + - amphibians hatch from eggs + - A solenoid or coil wrapped around iron or certain other metals can form a(n) electromagnet. +- source_sentence: About 200 countries and territories have reported coronavirus cases + in 2020 . + sentences: + - All-Time Olympic Games Medal Tally Analysis Home > Events > Olympics > Summer + > Medal Tally > All-Time All-Time Olympic Games Medal Tally (Summer Olympics) + Which country is the most successful at he Olympic Games? Here are the top ranked + countries in terms of total medals won when all of the summer Games are considered + (including the 2016 Rio Games). There are two tables presented, the first just + lists the top countries based on the total medals won, the second table factors + in how many Olympic Games the country appeared, averaging the total number of + medals per Olympiad. A victory in a team sport is counted as one medal. The USA + Has Won the Most Medals The US have clearly won the most gold medals and the most + medals overall, more than doubling the next ranked country (these figures include + medals won in Rio 2016). Second placed USSR had fewer appearances at the Olympics, + and actually won more medals on average (see the 2nd table). The top 10 includes + one country no longer in existence (the Soviet Union), so their medal totals will + obviously not increase, however China is expected to continue a rapid rise up + the ranks. With the addition of the 2016 data, China has moved up from 11th (in + 2008) to 9th (2012) to 7th (2016). The country which has attended the most games + without a medal is Monaco (20 Olympic Games), the country which has won the most + medals without winning a gold medal is Malaysia (0 gold, 7 silver, 4 bronze). + rank + - An example of a reproductive behavior is salmon returning to their birthplace + to lay their eggs + - more than 664,000 cases of COVID-19 have been reported in over 190 countries and + territories , resulting in approximately 30,800 deaths . +- source_sentence: The wave on a guitar string is transverse. the sound wave rattles + a sheet of paper in a direction that shows the sound wave is what? + sentences: + - A Honda motorcycle parked in a grass driveway + - In Panama tipping is a question of rewarding good service rather than an obligation. + Restaurant bills don't include gratuities; adding 10% is customary. Bellhops and + maids expect tips only in more expensive hotels, and $1–$2 per bag is the norm. + You should also give a tip of up to $10 per day to tour guides. + - Figure 16.33 The wave on a guitar string is transverse. The sound wave rattles + a sheet of paper in a direction that shows the sound wave is longitudinal. +- source_sentence: The thermal production of a stove is generically used for + sentences: + - In total , 28 US victims were killed , while Viet Cong losses were killed 345 + and a further 192 estimated killed . + - a stove generates heat for cooking usually + - A teenager has been charged over an incident in which a four-year-old girl was + hurt when she was hit in the face by a brick thrown through a van window. +- source_sentence: can sweet potatoes cause itching? + sentences: + - 'People with a true potato allergy may react immediately after touching, peeling, + or eating potatoes. Symptoms may vary from person to person, but typical symptoms + of a potato allergy include: rhinitis, including itchy or stinging eyes, a runny + or stuffy nose, and sneezing.' + - riding a bike does not cause pollution + - "Dilation occurs when cell walls relax.. An aneurysm is a dilation, or bubble,\ + \ that occurs in the wall of an artery. \n an artery can be relaxed by dilation" +model-index: +- name: SentenceTransformer based on microsoft/deberta-v3-small + results: + - task: + type: semantic-similarity + name: Semantic Similarity + dataset: + name: sts test + type: sts-test + metrics: + - type: pearson_cosine + value: 0.6368197640293093 + name: Pearson Cosine + - type: spearman_cosine + value: 0.6345637125214598 + name: Spearman Cosine + - type: pearson_manhattan + value: 0.6467215914161899 + name: Pearson Manhattan + - type: spearman_manhattan + value: 0.6336825601846632 + name: Spearman Manhattan + - type: pearson_euclidean + value: 0.6470519111681319 + name: Pearson Euclidean + - type: spearman_euclidean + value: 0.6345603114255941 + name: Spearman Euclidean + - type: pearson_dot + value: 0.6367905887877633 + name: Pearson Dot + - type: spearman_dot + value: 0.6346041676617576 + name: Spearman Dot + - type: pearson_max + value: 0.6470519111681319 + name: Pearson Max + - type: spearman_max + value: 0.6346041676617576 + name: Spearman Max + - task: + type: binary-classification + name: Binary Classification + dataset: + name: allNLI dev + type: allNLI-dev + metrics: + - type: cosine_accuracy + value: 0.70703125 + name: Cosine Accuracy + - type: cosine_accuracy_threshold + value: 0.908595085144043 + name: Cosine Accuracy Threshold + - type: cosine_f1 + value: 0.5522388059701493 + name: Cosine F1 + - type: cosine_f1_threshold + value: 0.8629225492477417 + name: Cosine F1 Threshold + - type: cosine_precision + value: 0.4847161572052402 + name: Cosine Precision + - type: cosine_recall + value: 0.6416184971098265 + name: Cosine Recall + - type: cosine_ap + value: 0.5430466488954966 + name: Cosine Ap + - type: dot_accuracy + value: 0.708984375 + name: Dot Accuracy + - type: dot_accuracy_threshold + value: 698.3015747070312 + name: Dot Accuracy Threshold + - type: dot_f1 + value: 0.5522388059701493 + name: Dot F1 + - type: dot_f1_threshold + value: 663.3272705078125 + name: Dot F1 Threshold + - type: dot_precision + value: 0.4847161572052402 + name: Dot Precision + - type: dot_recall + value: 0.6416184971098265 + name: Dot Recall + - type: dot_ap + value: 0.5430436315328802 + name: Dot Ap + - type: manhattan_accuracy + value: 0.705078125 + name: Manhattan Accuracy + - type: manhattan_accuracy_threshold + value: 263.4867858886719 + name: Manhattan Accuracy Threshold + - type: manhattan_f1 + value: 0.5520581113801454 + name: Manhattan F1 + - type: manhattan_f1_threshold + value: 326.1763000488281 + name: Manhattan F1 Threshold + - type: manhattan_precision + value: 0.475 + name: Manhattan Precision + - type: manhattan_recall + value: 0.6589595375722543 + name: Manhattan Recall + - type: manhattan_ap + value: 0.5413936285926788 + name: Manhattan Ap + - type: euclidean_accuracy + value: 0.70703125 + name: Euclidean Accuracy + - type: euclidean_accuracy_threshold + value: 11.85396957397461 + name: Euclidean Accuracy Threshold + - type: euclidean_f1 + value: 0.5522388059701493 + name: Euclidean F1 + - type: euclidean_f1_threshold + value: 14.51696491241455 + name: Euclidean F1 Threshold + - type: euclidean_precision + value: 0.4847161572052402 + name: Euclidean Precision + - type: euclidean_recall + value: 0.6416184971098265 + name: Euclidean Recall + - type: euclidean_ap + value: 0.5430008642299873 + name: Euclidean Ap + - type: max_accuracy + value: 0.708984375 + name: Max Accuracy + - type: max_accuracy_threshold + value: 698.3015747070312 + name: Max Accuracy Threshold + - type: max_f1 + value: 0.5522388059701493 + name: Max F1 + - type: max_f1_threshold + value: 663.3272705078125 + name: Max F1 Threshold + - type: max_precision + value: 0.4847161572052402 + name: Max Precision + - type: max_recall + value: 0.6589595375722543 + name: Max Recall + - type: max_ap + value: 0.5430466488954966 + name: Max Ap + - task: + type: binary-classification + name: Binary Classification + dataset: + name: Qnli dev + type: Qnli-dev + metrics: + - type: cosine_accuracy + value: 0.68359375 + name: Cosine Accuracy + - type: cosine_accuracy_threshold + value: 0.8038332462310791 + name: Cosine Accuracy Threshold + - type: cosine_f1 + value: 0.6764227642276421 + name: Cosine F1 + - type: cosine_f1_threshold + value: 0.7276865839958191 + name: Cosine F1 Threshold + - type: cosine_precision + value: 0.5488126649076517 + name: Cosine Precision + - type: cosine_recall + value: 0.8813559322033898 + name: Cosine Recall + - type: cosine_ap + value: 0.6915503966365267 + name: Cosine Ap + - type: dot_accuracy + value: 0.68359375 + name: Dot Accuracy + - type: dot_accuracy_threshold + value: 617.9757690429688 + name: Dot Accuracy Threshold + - type: dot_f1 + value: 0.6775244299674267 + name: Dot F1 + - type: dot_f1_threshold + value: 559.7400512695312 + name: Dot F1 Threshold + - type: dot_precision + value: 0.5502645502645502 + name: Dot Precision + - type: dot_recall + value: 0.8813559322033898 + name: Dot Recall + - type: dot_ap + value: 0.6914604071082934 + name: Dot Ap + - type: manhattan_accuracy + value: 0.681640625 + name: Manhattan Accuracy + - type: manhattan_accuracy_threshold + value: 384.64373779296875 + name: Manhattan Accuracy Threshold + - type: manhattan_f1 + value: 0.67430441898527 + name: Manhattan F1 + - type: manhattan_f1_threshold + value: 451.5675048828125 + name: Manhattan F1 Threshold + - type: manhattan_precision + value: 0.5493333333333333 + name: Manhattan Precision + - type: manhattan_recall + value: 0.8728813559322034 + name: Manhattan Recall + - type: manhattan_ap + value: 0.6911560630995964 + name: Manhattan Ap + - type: euclidean_accuracy + value: 0.68359375 + name: Euclidean Accuracy + - type: euclidean_accuracy_threshold + value: 17.36817741394043 + name: Euclidean Accuracy Threshold + - type: euclidean_f1 + value: 0.6764227642276421 + name: Euclidean F1 + - type: euclidean_f1_threshold + value: 20.461692810058594 + name: Euclidean F1 Threshold + - type: euclidean_precision + value: 0.5488126649076517 + name: Euclidean Precision + - type: euclidean_recall + value: 0.8813559322033898 + name: Euclidean Recall + - type: euclidean_ap + value: 0.6915804106776542 + name: Euclidean Ap + - type: max_accuracy + value: 0.68359375 + name: Max Accuracy + - type: max_accuracy_threshold + value: 617.9757690429688 + name: Max Accuracy Threshold + - type: max_f1 + value: 0.6775244299674267 + name: Max F1 + - type: max_f1_threshold + value: 559.7400512695312 + name: Max F1 Threshold + - type: max_precision + value: 0.5502645502645502 + name: Max Precision + - type: max_recall + value: 0.8813559322033898 + name: Max Recall + - type: max_ap + value: 0.6915804106776542 + name: Max Ap +--- + +# SentenceTransformer based on microsoft/deberta-v3-small + +This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [microsoft/deberta-v3-small](https://huggingface.co/microsoft/deberta-v3-small). It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more. + +## Model Details + +### Model Description +- **Model Type:** Sentence Transformer +- **Base model:** [microsoft/deberta-v3-small](https://huggingface.co/microsoft/deberta-v3-small) +- **Maximum Sequence Length:** 512 tokens +- **Output Dimensionality:** 768 tokens +- **Similarity Function:** Cosine Similarity + + + + +### Model Sources + +- **Documentation:** [Sentence Transformers Documentation](https://sbert.net) +- **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers) +- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers) + +### Full Model Architecture + +``` +SentenceTransformer( + (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: DebertaV2Model + (1): AdvancedWeightedPooling( + (alpha_dropout_layer): Dropout(p=0.01, inplace=False) + (gate_dropout_layer): Dropout(p=0.05, inplace=False) + (linear_cls_pj): Linear(in_features=768, out_features=768, bias=True) + (linear_cls_Qpj): Linear(in_features=768, out_features=768, bias=True) + (linear_mean_pj): Linear(in_features=768, out_features=768, bias=True) + (linear_attnOut): Linear(in_features=768, out_features=768, bias=True) + (mha): MultiheadAttention( + (out_proj): NonDynamicallyQuantizableLinear(in_features=768, out_features=768, bias=True) + ) + (layernorm_output): LayerNorm((768,), eps=1e-05, elementwise_affine=True) + (layernorm_weightedPooing): LayerNorm((768,), eps=1e-05, elementwise_affine=True) + (layernorm_pjCls): LayerNorm((768,), eps=1e-05, elementwise_affine=True) + (layernorm_pjMean): LayerNorm((768,), eps=1e-05, elementwise_affine=True) + (layernorm_attnOut): LayerNorm((768,), eps=1e-05, elementwise_affine=True) + ) +) +``` + +## Usage + +### Direct Usage (Sentence Transformers) + +First install the Sentence Transformers library: + +```bash +pip install -U sentence-transformers +``` + +Then you can load this model and run inference. +```python +from sentence_transformers import SentenceTransformer + +# Download from the 🤗 Hub +model = SentenceTransformer("bobox/DeBERTa3-s-CustomPoolin-toytest3-step1-checkpoints-tmp") +# Run inference +sentences = [ + 'can sweet potatoes cause itching?', + 'People with a true potato allergy may react immediately after touching, peeling, or eating potatoes. Symptoms may vary from person to person, but typical symptoms of a potato allergy include: rhinitis, including itchy or stinging eyes, a runny or stuffy nose, and sneezing.', + 'riding a bike does not cause pollution', +] +embeddings = model.encode(sentences) +print(embeddings.shape) +# [3, 768] + +# Get the similarity scores for the embeddings +similarities = model.similarity(embeddings, embeddings) +print(similarities.shape) +# [3, 3] +``` + + + + + + + +## Evaluation + +### Metrics + +#### Semantic Similarity +* Dataset: `sts-test` +* Evaluated with [EmbeddingSimilarityEvaluator](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.EmbeddingSimilarityEvaluator) + +| Metric | Value | +|:--------------------|:-----------| +| pearson_cosine | 0.6368 | +| **spearman_cosine** | **0.6346** | +| pearson_manhattan | 0.6467 | +| spearman_manhattan | 0.6337 | +| pearson_euclidean | 0.6471 | +| spearman_euclidean | 0.6346 | +| pearson_dot | 0.6368 | +| spearman_dot | 0.6346 | +| pearson_max | 0.6471 | +| spearman_max | 0.6346 | + +#### Binary Classification +* Dataset: `allNLI-dev` +* Evaluated with [BinaryClassificationEvaluator](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.BinaryClassificationEvaluator) + +| Metric | Value | +|:-----------------------------|:----------| +| cosine_accuracy | 0.707 | +| cosine_accuracy_threshold | 0.9086 | +| cosine_f1 | 0.5522 | +| cosine_f1_threshold | 0.8629 | +| cosine_precision | 0.4847 | +| cosine_recall | 0.6416 | +| cosine_ap | 0.543 | +| dot_accuracy | 0.709 | +| dot_accuracy_threshold | 698.3016 | +| dot_f1 | 0.5522 | +| dot_f1_threshold | 663.3273 | +| dot_precision | 0.4847 | +| dot_recall | 0.6416 | +| dot_ap | 0.543 | +| manhattan_accuracy | 0.7051 | +| manhattan_accuracy_threshold | 263.4868 | +| manhattan_f1 | 0.5521 | +| manhattan_f1_threshold | 326.1763 | +| manhattan_precision | 0.475 | +| manhattan_recall | 0.659 | +| manhattan_ap | 0.5414 | +| euclidean_accuracy | 0.707 | +| euclidean_accuracy_threshold | 11.854 | +| euclidean_f1 | 0.5522 | +| euclidean_f1_threshold | 14.517 | +| euclidean_precision | 0.4847 | +| euclidean_recall | 0.6416 | +| euclidean_ap | 0.543 | +| max_accuracy | 0.709 | +| max_accuracy_threshold | 698.3016 | +| max_f1 | 0.5522 | +| max_f1_threshold | 663.3273 | +| max_precision | 0.4847 | +| max_recall | 0.659 | +| **max_ap** | **0.543** | + +#### Binary Classification +* Dataset: `Qnli-dev` +* Evaluated with [BinaryClassificationEvaluator](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.BinaryClassificationEvaluator) + +| Metric | Value | +|:-----------------------------|:-----------| +| cosine_accuracy | 0.6836 | +| cosine_accuracy_threshold | 0.8038 | +| cosine_f1 | 0.6764 | +| cosine_f1_threshold | 0.7277 | +| cosine_precision | 0.5488 | +| cosine_recall | 0.8814 | +| cosine_ap | 0.6916 | +| dot_accuracy | 0.6836 | +| dot_accuracy_threshold | 617.9758 | +| dot_f1 | 0.6775 | +| dot_f1_threshold | 559.7401 | +| dot_precision | 0.5503 | +| dot_recall | 0.8814 | +| dot_ap | 0.6915 | +| manhattan_accuracy | 0.6816 | +| manhattan_accuracy_threshold | 384.6437 | +| manhattan_f1 | 0.6743 | +| manhattan_f1_threshold | 451.5675 | +| manhattan_precision | 0.5493 | +| manhattan_recall | 0.8729 | +| manhattan_ap | 0.6912 | +| euclidean_accuracy | 0.6836 | +| euclidean_accuracy_threshold | 17.3682 | +| euclidean_f1 | 0.6764 | +| euclidean_f1_threshold | 20.4617 | +| euclidean_precision | 0.5488 | +| euclidean_recall | 0.8814 | +| euclidean_ap | 0.6916 | +| max_accuracy | 0.6836 | +| max_accuracy_threshold | 617.9758 | +| max_f1 | 0.6775 | +| max_f1_threshold | 559.7401 | +| max_precision | 0.5503 | +| max_recall | 0.8814 | +| **max_ap** | **0.6916** | + + + + + +## Training Details + +### Training Dataset + +#### Unnamed Dataset + + +* Size: 32,500 training samples +* Columns: sentence1 and sentence2 +* Approximate statistics based on the first 1000 samples: + | | sentence1 | sentence2 | + |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | + | details | | | +* Samples: + | sentence1 | sentence2 | + |:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| + | The song ‘Fashion for His Love’ by Lady Gaga is a tribute to which late fashion designer? | Fashion Of His Love by Lady Gaga Songfacts Fashion Of His Love by Lady Gaga Songfacts Songfacts Gaga explained in a tweet that this track from her Born This Way Special Edition album is about the late Alexander McQueen. The fashion designer committed suicide by hanging on February 11, 2010 and Gaga was deeply affected by the tragic death of McQueen, who was a close personal friend. That same month, she performed at the 2010 Brit Awards wearing one of his couture creations and she also paid tribute to her late friend by setting the date on the prison security cameras in her Telephone video as the same day that McQueen's body was discovered in his London home. | + | e. in solids the atoms are closely locked in position and can only vibrate, in liquids the atoms and molecules are more loosely connected and can collide with and move past one another, while in gases the atoms or molecules are free to move independently, colliding frequently. | Within a substance, atoms that collide frequently and move independently of one another are most likely in a gas | + | Helen Lederer is an English comedian . | Helen Lederer ( born 24 September 1954 ) is an English : //www.scotsman.com/news/now-or-never-1-1396369 comedian , writer and actress who emerged as part of the alternative comedy boom at the beginning of the 1980s . | +* Loss: [GISTEmbedLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#gistembedloss) with these parameters: + ```json + {'guide': SentenceTransformer( + (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel + (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True}) + (2): Normalize() + ), 'temperature': 0.025} + ``` + +### Evaluation Dataset + +#### Unnamed Dataset + + +* Size: 1,664 evaluation samples +* Columns: sentence1 and sentence2 +* Approximate statistics based on the first 1000 samples: + | | sentence1 | sentence2 | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | + | details | | | +* Samples: + | sentence1 | sentence2 | + |:--------------------------------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| + | What planet did the voyager 1 spacecraft visit in 1980? | The Voyager 1 spacecraft visited Saturn in 1980. Voyager 2 followed in 1981. These probes sent back detailed pictures of Saturn, its rings, and some of its moons ( Figure below ). From the Voyager data, we learned what Saturn’s rings are made of. They are particles of water and ice with a little bit of dust. There are several gaps in the rings. These gaps were cleared out by moons within the rings. Gravity attracts dust and gas to the moon from the ring. This leaves a gap in the rings. Other gaps in the rings are caused by the competing forces of Saturn and its moons outside the rings. | + | Diffusion Diffusion is a process where atoms or molecules move from areas of high concentration to areas of low concentration. | Diffusion is the process in which a substance naturally moves from an area of higher to lower concentration. | + | Who had an 80s No 1 with Don't You Want Me? | The Human League - Don't You Want Me - YouTube The Human League - Don't You Want Me Want to watch this again later? Sign in to add this video to a playlist. Need to report the video? Sign in to report inappropriate content. Rating is available when the video has been rented. This feature is not available right now. Please try again later. Uploaded on Feb 27, 2009 Music video by The Human League performing Don't You Want Me (2003 Digital Remaster). Category | +* Loss: [GISTEmbedLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#gistembedloss) with these parameters: + ```json + {'guide': SentenceTransformer( + (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel + (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True}) + (2): Normalize() + ), 'temperature': 0.025} + ``` + +### Training Hyperparameters +#### Non-Default Hyperparameters + +- `eval_strategy`: steps +- `per_device_train_batch_size`: 32 +- `per_device_eval_batch_size`: 256 +- `lr_scheduler_type`: cosine_with_min_lr +- `lr_scheduler_kwargs`: {'num_cycles': 0.5, 'min_lr': 3.3333333333333337e-06} +- `warmup_ratio`: 0.33 +- `save_safetensors`: False +- `fp16`: True +- `push_to_hub`: True +- `hub_model_id`: bobox/DeBERTa3-s-CustomPoolin-toytest3-step1-checkpoints-tmp +- `hub_strategy`: all_checkpoints +- `batch_sampler`: no_duplicates + +#### All Hyperparameters +
Click to expand + +- `overwrite_output_dir`: False +- `do_predict`: False +- `eval_strategy`: steps +- `prediction_loss_only`: True +- `per_device_train_batch_size`: 32 +- `per_device_eval_batch_size`: 256 +- `per_gpu_train_batch_size`: None +- `per_gpu_eval_batch_size`: None +- `gradient_accumulation_steps`: 1 +- `eval_accumulation_steps`: None +- `torch_empty_cache_steps`: None +- `learning_rate`: 5e-05 +- `weight_decay`: 0.0 +- `adam_beta1`: 0.9 +- `adam_beta2`: 0.999 +- `adam_epsilon`: 1e-08 +- `max_grad_norm`: 1.0 +- `num_train_epochs`: 3 +- `max_steps`: -1 +- `lr_scheduler_type`: cosine_with_min_lr +- `lr_scheduler_kwargs`: {'num_cycles': 0.5, 'min_lr': 3.3333333333333337e-06} +- `warmup_ratio`: 0.33 +- `warmup_steps`: 0 +- `log_level`: passive +- `log_level_replica`: warning +- `log_on_each_node`: True +- `logging_nan_inf_filter`: True +- `save_safetensors`: False +- `save_on_each_node`: False +- `save_only_model`: False +- `restore_callback_states_from_checkpoint`: False +- `no_cuda`: False +- `use_cpu`: False +- `use_mps_device`: False +- `seed`: 42 +- `data_seed`: None +- `jit_mode_eval`: False +- `use_ipex`: False +- `bf16`: False +- `fp16`: True +- `fp16_opt_level`: O1 +- `half_precision_backend`: auto +- `bf16_full_eval`: False +- `fp16_full_eval`: False +- `tf32`: None +- `local_rank`: 0 +- `ddp_backend`: None +- `tpu_num_cores`: None +- `tpu_metrics_debug`: False +- `debug`: [] +- `dataloader_drop_last`: False +- `dataloader_num_workers`: 0 +- `dataloader_prefetch_factor`: None +- `past_index`: -1 +- `disable_tqdm`: False +- `remove_unused_columns`: True +- `label_names`: None +- `load_best_model_at_end`: False +- `ignore_data_skip`: False +- `fsdp`: [] +- `fsdp_min_num_params`: 0 +- `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False} +- `fsdp_transformer_layer_cls_to_wrap`: None +- `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None} +- `deepspeed`: None +- `label_smoothing_factor`: 0.0 +- `optim`: adamw_torch +- `optim_args`: None +- `adafactor`: False +- `group_by_length`: False +- `length_column_name`: length +- `ddp_find_unused_parameters`: None +- `ddp_bucket_cap_mb`: None +- `ddp_broadcast_buffers`: False +- `dataloader_pin_memory`: True +- `dataloader_persistent_workers`: False +- `skip_memory_metrics`: True +- `use_legacy_prediction_loop`: False +- `push_to_hub`: True +- `resume_from_checkpoint`: None +- `hub_model_id`: bobox/DeBERTa3-s-CustomPoolin-toytest3-step1-checkpoints-tmp +- `hub_strategy`: all_checkpoints +- `hub_private_repo`: False +- `hub_always_push`: False +- `gradient_checkpointing`: False +- `gradient_checkpointing_kwargs`: None +- `include_inputs_for_metrics`: False +- `eval_do_concat_batches`: True +- `fp16_backend`: auto +- `push_to_hub_model_id`: None +- `push_to_hub_organization`: None +- `mp_parameters`: +- `auto_find_batch_size`: False +- `full_determinism`: False +- `torchdynamo`: None +- `ray_scope`: last +- `ddp_timeout`: 1800 +- `torch_compile`: False +- `torch_compile_backend`: None +- `torch_compile_mode`: None +- `dispatch_batches`: None +- `split_batches`: None +- `include_tokens_per_second`: False +- `include_num_input_tokens_seen`: False +- `neftune_noise_alpha`: None +- `optim_target_modules`: None +- `batch_eval_metrics`: False +- `eval_on_start`: False +- `eval_use_gather_object`: False +- `batch_sampler`: no_duplicates +- `multi_dataset_batch_sampler`: proportional + +
+ +### Training Logs +
Click to expand + +| Epoch | Step | Training Loss | Validation Loss | sts-test_spearman_cosine | allNLI-dev_max_ap | Qnli-dev_max_ap | +|:------:|:----:|:-------------:|:---------------:|:------------------------:|:-----------------:|:---------------:| +| 0.0010 | 1 | 10.4072 | - | - | - | - | +| 0.0020 | 2 | 11.0865 | - | - | - | - | +| 0.0030 | 3 | 9.5114 | - | - | - | - | +| 0.0039 | 4 | 9.9584 | - | - | - | - | +| 0.0049 | 5 | 10.068 | - | - | - | - | +| 0.0059 | 6 | 11.0224 | - | - | - | - | +| 0.0069 | 7 | 9.7703 | - | - | - | - | +| 0.0079 | 8 | 10.5005 | - | - | - | - | +| 0.0089 | 9 | 10.1987 | - | - | - | - | +| 0.0098 | 10 | 10.0277 | - | - | - | - | +| 0.0108 | 11 | 10.6965 | - | - | - | - | +| 0.0118 | 12 | 10.0609 | - | - | - | - | +| 0.0128 | 13 | 11.6214 | - | - | - | - | +| 0.0138 | 14 | 9.4053 | - | - | - | - | +| 0.0148 | 15 | 10.4014 | - | - | - | - | +| 0.0157 | 16 | 10.4119 | - | - | - | - | +| 0.0167 | 17 | 9.4658 | - | - | - | - | +| 0.0177 | 18 | 9.2169 | - | - | - | - | +| 0.0187 | 19 | 11.2337 | - | - | - | - | +| 0.0197 | 20 | 11.0572 | - | - | - | - | +| 0.0207 | 21 | 11.0452 | - | - | - | - | +| 0.0217 | 22 | 10.31 | - | - | - | - | +| 0.0226 | 23 | 9.1395 | - | - | - | - | +| 0.0236 | 24 | 8.4201 | - | - | - | - | +| 0.0246 | 25 | 8.6036 | - | - | - | - | +| 0.0256 | 26 | 11.7579 | - | - | - | - | +| 0.0266 | 27 | 10.1307 | - | - | - | - | +| 0.0276 | 28 | 9.2915 | - | - | - | - | +| 0.0285 | 29 | 9.0208 | - | - | - | - | +| 0.0295 | 30 | 8.6867 | - | - | - | - | +| 0.0305 | 31 | 8.0925 | - | - | - | - | +| 0.0315 | 32 | 8.6617 | - | - | - | - | +| 0.0325 | 33 | 8.3374 | - | - | - | - | +| 0.0335 | 34 | 7.8566 | - | - | - | - | +| 0.0344 | 35 | 9.0698 | - | - | - | - | +| 0.0354 | 36 | 7.7727 | - | - | - | - | +| 0.0364 | 37 | 7.6128 | - | - | - | - | +| 0.0374 | 38 | 7.8762 | - | - | - | - | +| 0.0384 | 39 | 7.5191 | - | - | - | - | +| 0.0394 | 40 | 7.5638 | - | - | - | - | +| 0.0404 | 41 | 7.1878 | - | - | - | - | +| 0.0413 | 42 | 6.8878 | - | - | - | - | +| 0.0423 | 43 | 7.5775 | - | - | - | - | +| 0.0433 | 44 | 7.1076 | - | - | - | - | +| 0.0443 | 45 | 6.5589 | - | - | - | - | +| 0.0453 | 46 | 7.4456 | - | - | - | - | +| 0.0463 | 47 | 6.8233 | - | - | - | - | +| 0.0472 | 48 | 6.7633 | - | - | - | - | +| 0.0482 | 49 | 6.6024 | - | - | - | - | +| 0.0492 | 50 | 6.2778 | - | - | - | - | +| 0.0502 | 51 | 6.1026 | - | - | - | - | +| 0.0512 | 52 | 6.632 | - | - | - | - | +| 0.0522 | 53 | 6.6962 | - | - | - | - | +| 0.0531 | 54 | 5.8514 | - | - | - | - | +| 0.0541 | 55 | 5.9951 | - | - | - | - | +| 0.0551 | 56 | 5.4554 | - | - | - | - | +| 0.0561 | 57 | 6.0147 | - | - | - | - | +| 0.0571 | 58 | 5.215 | - | - | - | - | +| 0.0581 | 59 | 6.4525 | - | - | - | - | +| 0.0591 | 60 | 5.4048 | - | - | - | - | +| 0.0600 | 61 | 5.0424 | - | - | - | - | +| 0.0610 | 62 | 6.2646 | - | - | - | - | +| 0.0620 | 63 | 5.0847 | - | - | - | - | +| 0.0630 | 64 | 5.4415 | - | - | - | - | +| 0.0640 | 65 | 5.2469 | - | - | - | - | +| 0.0650 | 66 | 5.1378 | - | - | - | - | +| 0.0659 | 67 | 5.1636 | - | - | - | - | +| 0.0669 | 68 | 5.5596 | - | - | - | - | +| 0.0679 | 69 | 4.9508 | - | - | - | - | +| 0.0689 | 70 | 5.2355 | - | - | - | - | +| 0.0699 | 71 | 4.7359 | - | - | - | - | +| 0.0709 | 72 | 4.8947 | - | - | - | - | +| 0.0719 | 73 | 4.6269 | - | - | - | - | +| 0.0728 | 74 | 4.6072 | - | - | - | - | +| 0.0738 | 75 | 4.9125 | - | - | - | - | +| 0.0748 | 76 | 4.5856 | - | - | - | - | +| 0.0758 | 77 | 4.7879 | - | - | - | - | +| 0.0768 | 78 | 4.5348 | - | - | - | - | +| 0.0778 | 79 | 4.3554 | - | - | - | - | +| 0.0787 | 80 | 4.2984 | - | - | - | - | +| 0.0797 | 81 | 4.5505 | - | - | - | - | +| 0.0807 | 82 | 4.5325 | - | - | - | - | +| 0.0817 | 83 | 4.2725 | - | - | - | - | +| 0.0827 | 84 | 4.3054 | - | - | - | - | +| 0.0837 | 85 | 4.5536 | - | - | - | - | +| 0.0846 | 86 | 4.0265 | - | - | - | - | +| 0.0856 | 87 | 4.7453 | - | - | - | - | +| 0.0866 | 88 | 4.071 | - | - | - | - | +| 0.0876 | 89 | 4.1582 | - | - | - | - | +| 0.0886 | 90 | 4.1131 | - | - | - | - | +| 0.0896 | 91 | 3.6582 | - | - | - | - | +| 0.0906 | 92 | 4.143 | - | - | - | - | +| 0.0915 | 93 | 4.2273 | - | - | - | - | +| 0.0925 | 94 | 3.9321 | - | - | - | - | +| 0.0935 | 95 | 4.2061 | - | - | - | - | +| 0.0945 | 96 | 4.1042 | - | - | - | - | +| 0.0955 | 97 | 3.9513 | - | - | - | - | +| 0.0965 | 98 | 3.8627 | - | - | - | - | +| 0.0974 | 99 | 4.3613 | - | - | - | - | +| 0.0984 | 100 | 3.8513 | - | - | - | - | +| 0.0994 | 101 | 3.5866 | - | - | - | - | +| 0.1004 | 102 | 3.5239 | - | - | - | - | +| 0.1014 | 103 | 3.5921 | - | - | - | - | +| 0.1024 | 104 | 3.5962 | - | - | - | - | +| 0.1033 | 105 | 4.0001 | - | - | - | - | +| 0.1043 | 106 | 4.1374 | - | - | - | - | +| 0.1053 | 107 | 3.9049 | - | - | - | - | +| 0.1063 | 108 | 3.2511 | - | - | - | - | +| 0.1073 | 109 | 3.2479 | - | - | - | - | +| 0.1083 | 110 | 3.6414 | - | - | - | - | +| 0.1093 | 111 | 3.6429 | - | - | - | - | +| 0.1102 | 112 | 3.423 | - | - | - | - | +| 0.1112 | 113 | 3.4967 | - | - | - | - | +| 0.1122 | 114 | 3.7649 | - | - | - | - | +| 0.1132 | 115 | 3.2845 | - | - | - | - | +| 0.1142 | 116 | 3.356 | - | - | - | - | +| 0.1152 | 117 | 3.2086 | - | - | - | - | +| 0.1161 | 118 | 3.5561 | - | - | - | - | +| 0.1171 | 119 | 3.7353 | - | - | - | - | +| 0.1181 | 120 | 3.403 | - | - | - | - | +| 0.1191 | 121 | 3.1009 | - | - | - | - | +| 0.1201 | 122 | 3.2139 | - | - | - | - | +| 0.1211 | 123 | 3.3339 | - | - | - | - | +| 0.1220 | 124 | 2.9464 | - | - | - | - | +| 0.1230 | 125 | 3.3366 | - | - | - | - | +| 0.1240 | 126 | 3.0618 | - | - | - | - | +| 0.125 | 127 | 3.0092 | - | - | - | - | +| 0.1260 | 128 | 2.7152 | - | - | - | - | +| 0.1270 | 129 | 2.9423 | - | - | - | - | +| 0.1280 | 130 | 2.6569 | - | - | - | - | +| 0.1289 | 131 | 2.8469 | - | - | - | - | +| 0.1299 | 132 | 2.9089 | - | - | - | - | +| 0.1309 | 133 | 2.5809 | - | - | - | - | +| 0.1319 | 134 | 2.6987 | - | - | - | - | +| 0.1329 | 135 | 3.2518 | - | - | - | - | +| 0.1339 | 136 | 2.9145 | - | - | - | - | +| 0.1348 | 137 | 2.4809 | - | - | - | - | +| 0.1358 | 138 | 2.8264 | - | - | - | - | +| 0.1368 | 139 | 2.5724 | - | - | - | - | +| 0.1378 | 140 | 2.6949 | - | - | - | - | +| 0.1388 | 141 | 2.6925 | - | - | - | - | +| 0.1398 | 142 | 2.9311 | - | - | - | - | +| 0.1407 | 143 | 2.5667 | - | - | - | - | +| 0.1417 | 144 | 3.2471 | - | - | - | - | +| 0.1427 | 145 | 2.2441 | - | - | - | - | +| 0.1437 | 146 | 2.75 | - | - | - | - | +| 0.1447 | 147 | 2.9669 | - | - | - | - | +| 0.1457 | 148 | 2.736 | - | - | - | - | +| 0.1467 | 149 | 3.104 | - | - | - | - | +| 0.1476 | 150 | 2.2175 | - | - | - | - | +| 0.1486 | 151 | 2.7415 | - | - | - | - | +| 0.1496 | 152 | 1.8707 | - | - | - | - | +| 0.1506 | 153 | 2.5961 | 2.2653 | 0.3116 | 0.4265 | 0.6462 | +| 0.1516 | 154 | 3.1149 | - | - | - | - | +| 0.1526 | 155 | 2.2976 | - | - | - | - | +| 0.1535 | 156 | 2.4436 | - | - | - | - | +| 0.1545 | 157 | 2.8826 | - | - | - | - | +| 0.1555 | 158 | 2.3664 | - | - | - | - | +| 0.1565 | 159 | 2.2485 | - | - | - | - | +| 0.1575 | 160 | 2.5167 | - | - | - | - | +| 0.1585 | 161 | 1.7183 | - | - | - | - | +| 0.1594 | 162 | 2.1003 | - | - | - | - | +| 0.1604 | 163 | 2.5785 | - | - | - | - | +| 0.1614 | 164 | 2.8789 | - | - | - | - | +| 0.1624 | 165 | 2.3425 | - | - | - | - | +| 0.1634 | 166 | 2.0966 | - | - | - | - | +| 0.1644 | 167 | 2.1126 | - | - | - | - | +| 0.1654 | 168 | 2.1824 | - | - | - | - | +| 0.1663 | 169 | 2.2009 | - | - | - | - | +| 0.1673 | 170 | 2.3796 | - | - | - | - | +| 0.1683 | 171 | 2.3096 | - | - | - | - | +| 0.1693 | 172 | 2.7897 | - | - | - | - | +| 0.1703 | 173 | 2.2097 | - | - | - | - | +| 0.1713 | 174 | 1.7508 | - | - | - | - | +| 0.1722 | 175 | 2.353 | - | - | - | - | +| 0.1732 | 176 | 2.4276 | - | - | - | - | +| 0.1742 | 177 | 2.1016 | - | - | - | - | +| 0.1752 | 178 | 1.8461 | - | - | - | - | +| 0.1762 | 179 | 1.8104 | - | - | - | - | +| 0.1772 | 180 | 2.6023 | - | - | - | - | +| 0.1781 | 181 | 2.5261 | - | - | - | - | +| 0.1791 | 182 | 2.1053 | - | - | - | - | +| 0.1801 | 183 | 1.9712 | - | - | - | - | +| 0.1811 | 184 | 2.4693 | - | - | - | - | +| 0.1821 | 185 | 2.1119 | - | - | - | - | +| 0.1831 | 186 | 2.4797 | - | - | - | - | +| 0.1841 | 187 | 2.1587 | - | - | - | - | +| 0.1850 | 188 | 1.9578 | - | - | - | - | +| 0.1860 | 189 | 2.1368 | - | - | - | - | +| 0.1870 | 190 | 2.4212 | - | - | - | - | +| 0.1880 | 191 | 1.9591 | - | - | - | - | +| 0.1890 | 192 | 1.5816 | - | - | - | - | +| 0.1900 | 193 | 1.4029 | - | - | - | - | +| 0.1909 | 194 | 1.9385 | - | - | - | - | +| 0.1919 | 195 | 1.5596 | - | - | - | - | +| 0.1929 | 196 | 1.6663 | - | - | - | - | +| 0.1939 | 197 | 2.0026 | - | - | - | - | +| 0.1949 | 198 | 2.0046 | - | - | - | - | +| 0.1959 | 199 | 1.5016 | - | - | - | - | +| 0.1969 | 200 | 2.184 | - | - | - | - | +| 0.1978 | 201 | 2.3442 | - | - | - | - | +| 0.1988 | 202 | 2.6981 | - | - | - | - | +| 0.1998 | 203 | 2.5481 | - | - | - | - | +| 0.2008 | 204 | 2.9798 | - | - | - | - | +| 0.2018 | 205 | 2.287 | - | - | - | - | +| 0.2028 | 206 | 1.9393 | - | - | - | - | +| 0.2037 | 207 | 2.892 | - | - | - | - | +| 0.2047 | 208 | 2.26 | - | - | - | - | +| 0.2057 | 209 | 2.5911 | - | - | - | - | +| 0.2067 | 210 | 2.1239 | - | - | - | - | +| 0.2077 | 211 | 2.0683 | - | - | - | - | +| 0.2087 | 212 | 1.768 | - | - | - | - | +| 0.2096 | 213 | 2.5468 | - | - | - | - | +| 0.2106 | 214 | 1.8956 | - | - | - | - | +| 0.2116 | 215 | 2.044 | - | - | - | - | +| 0.2126 | 216 | 1.5721 | - | - | - | - | +| 0.2136 | 217 | 1.6278 | - | - | - | - | +| 0.2146 | 218 | 1.7754 | - | - | - | - | +| 0.2156 | 219 | 1.8594 | - | - | - | - | +| 0.2165 | 220 | 1.8309 | - | - | - | - | +| 0.2175 | 221 | 2.0619 | - | - | - | - | +| 0.2185 | 222 | 2.3335 | - | - | - | - | +| 0.2195 | 223 | 2.023 | - | - | - | - | +| 0.2205 | 224 | 2.1975 | - | - | - | - | +| 0.2215 | 225 | 1.9228 | - | - | - | - | +| 0.2224 | 226 | 2.3565 | - | - | - | - | +| 0.2234 | 227 | 1.896 | - | - | - | - | +| 0.2244 | 228 | 2.0912 | - | - | - | - | +| 0.2254 | 229 | 2.7703 | - | - | - | - | +| 0.2264 | 230 | 1.6988 | - | - | - | - | +| 0.2274 | 231 | 2.0406 | - | - | - | - | +| 0.2283 | 232 | 1.9288 | - | - | - | - | +| 0.2293 | 233 | 2.0457 | - | - | - | - | +| 0.2303 | 234 | 1.7061 | - | - | - | - | +| 0.2313 | 235 | 1.6244 | - | - | - | - | +| 0.2323 | 236 | 2.0241 | - | - | - | - | +| 0.2333 | 237 | 1.567 | - | - | - | - | +| 0.2343 | 238 | 1.8084 | - | - | - | - | +| 0.2352 | 239 | 2.4363 | - | - | - | - | +| 0.2362 | 240 | 1.7532 | - | - | - | - | +| 0.2372 | 241 | 2.0797 | - | - | - | - | +| 0.2382 | 242 | 1.9562 | - | - | - | - | +| 0.2392 | 243 | 1.6751 | - | - | - | - | +| 0.2402 | 244 | 2.0265 | - | - | - | - | +| 0.2411 | 245 | 1.6065 | - | - | - | - | +| 0.2421 | 246 | 1.7439 | - | - | - | - | +| 0.2431 | 247 | 2.0237 | - | - | - | - | +| 0.2441 | 248 | 1.6128 | - | - | - | - | +| 0.2451 | 249 | 1.6581 | - | - | - | - | +| 0.2461 | 250 | 2.1538 | - | - | - | - | +| 0.2470 | 251 | 2.049 | - | - | - | - | +| 0.2480 | 252 | 1.2573 | - | - | - | - | +| 0.2490 | 253 | 1.5619 | - | - | - | - | +| 0.25 | 254 | 1.2611 | - | - | - | - | +| 0.2510 | 255 | 1.3443 | - | - | - | - | +| 0.2520 | 256 | 1.3436 | - | - | - | - | +| 0.2530 | 257 | 2.8117 | - | - | - | - | +| 0.2539 | 258 | 1.7563 | - | - | - | - | +| 0.2549 | 259 | 1.3148 | - | - | - | - | +| 0.2559 | 260 | 2.0278 | - | - | - | - | +| 0.2569 | 261 | 1.2403 | - | - | - | - | +| 0.2579 | 262 | 1.588 | - | - | - | - | +| 0.2589 | 263 | 2.0071 | - | - | - | - | +| 0.2598 | 264 | 1.5312 | - | - | - | - | +| 0.2608 | 265 | 1.8641 | - | - | - | - | +| 0.2618 | 266 | 1.2933 | - | - | - | - | +| 0.2628 | 267 | 1.6262 | - | - | - | - | +| 0.2638 | 268 | 1.721 | - | - | - | - | +| 0.2648 | 269 | 1.4713 | - | - | - | - | +| 0.2657 | 270 | 1.4625 | - | - | - | - | +| 0.2667 | 271 | 1.7254 | - | - | - | - | +| 0.2677 | 272 | 1.5108 | - | - | - | - | +| 0.2687 | 273 | 2.1126 | - | - | - | - | +| 0.2697 | 274 | 1.3967 | - | - | - | - | +| 0.2707 | 275 | 1.7067 | - | - | - | - | +| 0.2717 | 276 | 1.4847 | - | - | - | - | +| 0.2726 | 277 | 1.6515 | - | - | - | - | +| 0.2736 | 278 | 0.9367 | - | - | - | - | +| 0.2746 | 279 | 2.0267 | - | - | - | - | +| 0.2756 | 280 | 1.5023 | - | - | - | - | +| 0.2766 | 281 | 1.1248 | - | - | - | - | +| 0.2776 | 282 | 1.6224 | - | - | - | - | +| 0.2785 | 283 | 1.7969 | - | - | - | - | +| 0.2795 | 284 | 2.2498 | - | - | - | - | +| 0.2805 | 285 | 1.7477 | - | - | - | - | +| 0.2815 | 286 | 1.6261 | - | - | - | - | +| 0.2825 | 287 | 2.0911 | - | - | - | - | +| 0.2835 | 288 | 1.9519 | - | - | - | - | +| 0.2844 | 289 | 1.3132 | - | - | - | - | +| 0.2854 | 290 | 2.3292 | - | - | - | - | +| 0.2864 | 291 | 1.3781 | - | - | - | - | +| 0.2874 | 292 | 1.5753 | - | - | - | - | +| 0.2884 | 293 | 1.4158 | - | - | - | - | +| 0.2894 | 294 | 2.1661 | - | - | - | - | +| 0.2904 | 295 | 1.4928 | - | - | - | - | +| 0.2913 | 296 | 2.2825 | - | - | - | - | +| 0.2923 | 297 | 1.7261 | - | - | - | - | +| 0.2933 | 298 | 1.8635 | - | - | - | - | +| 0.2943 | 299 | 0.974 | - | - | - | - | +| 0.2953 | 300 | 1.53 | - | - | - | - | +| 0.2963 | 301 | 1.5985 | - | - | - | - | +| 0.2972 | 302 | 1.2169 | - | - | - | - | +| 0.2982 | 303 | 1.771 | - | - | - | - | +| 0.2992 | 304 | 1.4506 | - | - | - | - | +| 0.3002 | 305 | 1.9496 | - | - | - | - | +| 0.3012 | 306 | 1.2436 | 1.5213 | 0.4673 | 0.4808 | 0.6993 | +| 0.3022 | 307 | 2.2057 | - | - | - | - | +| 0.3031 | 308 | 1.6786 | - | - | - | - | +| 0.3041 | 309 | 1.748 | - | - | - | - | +| 0.3051 | 310 | 1.5541 | - | - | - | - | +| 0.3061 | 311 | 2.2968 | - | - | - | - | +| 0.3071 | 312 | 1.585 | - | - | - | - | +| 0.3081 | 313 | 1.8371 | - | - | - | - | +| 0.3091 | 314 | 1.1129 | - | - | - | - | +| 0.3100 | 315 | 1.5495 | - | - | - | - | +| 0.3110 | 316 | 1.4327 | - | - | - | - | +| 0.3120 | 317 | 1.4801 | - | - | - | - | +| 0.3130 | 318 | 1.7096 | - | - | - | - | +| 0.3140 | 319 | 1.6717 | - | - | - | - | +| 0.3150 | 320 | 1.7151 | - | - | - | - | +| 0.3159 | 321 | 1.7081 | - | - | - | - | +| 0.3169 | 322 | 1.431 | - | - | - | - | +| 0.3179 | 323 | 1.5734 | - | - | - | - | +| 0.3189 | 324 | 1.7307 | - | - | - | - | +| 0.3199 | 325 | 1.0644 | - | - | - | - | +| 0.3209 | 326 | 1.0651 | - | - | - | - | +| 0.3219 | 327 | 1.4805 | - | - | - | - | +| 0.3228 | 328 | 0.839 | - | - | - | - | +| 0.3238 | 329 | 1.1801 | - | - | - | - | +| 0.3248 | 330 | 1.36 | - | - | - | - | +| 0.3258 | 331 | 1.3371 | - | - | - | - | +| 0.3268 | 332 | 1.1707 | - | - | - | - | +| 0.3278 | 333 | 1.2572 | - | - | - | - | +| 0.3287 | 334 | 1.3537 | - | - | - | - | +| 0.3297 | 335 | 1.7096 | - | - | - | - | +| 0.3307 | 336 | 1.5137 | - | - | - | - | +| 0.3317 | 337 | 1.1989 | - | - | - | - | +| 0.3327 | 338 | 1.3559 | - | - | - | - | +| 0.3337 | 339 | 1.3643 | - | - | - | - | +| 0.3346 | 340 | 1.2283 | - | - | - | - | +| 0.3356 | 341 | 1.5829 | - | - | - | - | +| 0.3366 | 342 | 1.1866 | - | - | - | - | +| 0.3376 | 343 | 1.531 | - | - | - | - | +| 0.3386 | 344 | 1.5581 | - | - | - | - | +| 0.3396 | 345 | 1.5587 | - | - | - | - | +| 0.3406 | 346 | 1.1403 | - | - | - | - | +| 0.3415 | 347 | 1.9728 | - | - | - | - | +| 0.3425 | 348 | 1.0818 | - | - | - | - | +| 0.3435 | 349 | 1.2993 | - | - | - | - | +| 0.3445 | 350 | 1.7779 | - | - | - | - | +| 0.3455 | 351 | 1.319 | - | - | - | - | +| 0.3465 | 352 | 1.9236 | - | - | - | - | +| 0.3474 | 353 | 1.3085 | - | - | - | - | +| 0.3484 | 354 | 2.2049 | - | - | - | - | +| 0.3494 | 355 | 1.3697 | - | - | - | - | +| 0.3504 | 356 | 1.5367 | - | - | - | - | +| 0.3514 | 357 | 1.2516 | - | - | - | - | +| 0.3524 | 358 | 1.6497 | - | - | - | - | +| 0.3533 | 359 | 1.2457 | - | - | - | - | +| 0.3543 | 360 | 1.2733 | - | - | - | - | +| 0.3553 | 361 | 1.4768 | - | - | - | - | +| 0.3563 | 362 | 1.1363 | - | - | - | - | +| 0.3573 | 363 | 1.5731 | - | - | - | - | +| 0.3583 | 364 | 1.0821 | - | - | - | - | +| 0.3593 | 365 | 1.1563 | - | - | - | - | +| 0.3602 | 366 | 1.8843 | - | - | - | - | +| 0.3612 | 367 | 1.2239 | - | - | - | - | +| 0.3622 | 368 | 1.4411 | - | - | - | - | +| 0.3632 | 369 | 2.1003 | - | - | - | - | +| 0.3642 | 370 | 1.6558 | - | - | - | - | +| 0.3652 | 371 | 1.6502 | - | - | - | - | +| 0.3661 | 372 | 1.7204 | - | - | - | - | +| 0.3671 | 373 | 1.7422 | - | - | - | - | +| 0.3681 | 374 | 1.3859 | - | - | - | - | +| 0.3691 | 375 | 0.8876 | - | - | - | - | +| 0.3701 | 376 | 1.2399 | - | - | - | - | +| 0.3711 | 377 | 1.1039 | - | - | - | - | +| 0.3720 | 378 | 1.733 | - | - | - | - | +| 0.3730 | 379 | 1.6897 | - | - | - | - | +| 0.3740 | 380 | 2.0532 | - | - | - | - | +| 0.375 | 381 | 1.0156 | - | - | - | - | +| 0.3760 | 382 | 0.8888 | - | - | - | - | +| 0.3770 | 383 | 1.322 | - | - | - | - | +| 0.3780 | 384 | 1.6828 | - | - | - | - | +| 0.3789 | 385 | 1.1567 | - | - | - | - | +| 0.3799 | 386 | 1.6117 | - | - | - | - | +| 0.3809 | 387 | 1.1776 | - | - | - | - | +| 0.3819 | 388 | 1.641 | - | - | - | - | +| 0.3829 | 389 | 1.3454 | - | - | - | - | +| 0.3839 | 390 | 1.4292 | - | - | - | - | +| 0.3848 | 391 | 1.2256 | - | - | - | - | +| 0.3858 | 392 | 1.08 | - | - | - | - | +| 0.3868 | 393 | 0.7436 | - | - | - | - | +| 0.3878 | 394 | 1.4112 | - | - | - | - | +| 0.3888 | 395 | 0.8917 | - | - | - | - | +| 0.3898 | 396 | 0.9955 | - | - | - | - | +| 0.3907 | 397 | 1.2867 | - | - | - | - | +| 0.3917 | 398 | 1.0683 | - | - | - | - | +| 0.3927 | 399 | 0.9355 | - | - | - | - | +| 0.3937 | 400 | 1.1153 | - | - | - | - | +| 0.3947 | 401 | 1.1724 | - | - | - | - | +| 0.3957 | 402 | 1.4069 | - | - | - | - | +| 0.3967 | 403 | 1.2546 | - | - | - | - | +| 0.3976 | 404 | 2.2862 | - | - | - | - | +| 0.3986 | 405 | 1.2316 | - | - | - | - | +| 0.3996 | 406 | 1.7876 | - | - | - | - | +| 0.4006 | 407 | 0.6936 | - | - | - | - | +| 0.4016 | 408 | 1.3852 | - | - | - | - | +| 0.4026 | 409 | 1.9046 | - | - | - | - | +| 0.4035 | 410 | 1.4972 | - | - | - | - | +| 0.4045 | 411 | 0.5531 | - | - | - | - | +| 0.4055 | 412 | 1.3685 | - | - | - | - | +| 0.4065 | 413 | 1.1367 | - | - | - | - | +| 0.4075 | 414 | 1.1304 | - | - | - | - | +| 0.4085 | 415 | 1.5953 | - | - | - | - | +| 0.4094 | 416 | 2.0308 | - | - | - | - | +| 0.4104 | 417 | 1.7275 | - | - | - | - | +| 0.4114 | 418 | 0.9921 | - | - | - | - | +| 0.4124 | 419 | 1.3418 | - | - | - | - | +| 0.4134 | 420 | 1.108 | - | - | - | - | +| 0.4144 | 421 | 1.4359 | - | - | - | - | +| 0.4154 | 422 | 1.4537 | - | - | - | - | +| 0.4163 | 423 | 0.8416 | - | - | - | - | +| 0.4173 | 424 | 0.8904 | - | - | - | - | +| 0.4183 | 425 | 0.7937 | - | - | - | - | +| 0.4193 | 426 | 0.9105 | - | - | - | - | +| 0.4203 | 427 | 1.1661 | - | - | - | - | +| 0.4213 | 428 | 0.7751 | - | - | - | - | +| 0.4222 | 429 | 0.9039 | - | - | - | - | +| 0.4232 | 430 | 1.2651 | - | - | - | - | +| 0.4242 | 431 | 1.44 | - | - | - | - | +| 0.4252 | 432 | 0.9795 | - | - | - | - | +| 0.4262 | 433 | 2.1892 | - | - | - | - | +| 0.4272 | 434 | 1.214 | - | - | - | - | +| 0.4281 | 435 | 1.185 | - | - | - | - | +| 0.4291 | 436 | 1.2501 | - | - | - | - | +| 0.4301 | 437 | 1.6432 | - | - | - | - | +| 0.4311 | 438 | 1.0203 | - | - | - | - | +| 0.4321 | 439 | 1.5179 | - | - | - | - | +| 0.4331 | 440 | 1.1445 | - | - | - | - | +| 0.4341 | 441 | 1.3099 | - | - | - | - | +| 0.4350 | 442 | 0.8856 | - | - | - | - | +| 0.4360 | 443 | 0.5869 | - | - | - | - | +| 0.4370 | 444 | 1.6335 | - | - | - | - | +| 0.4380 | 445 | 1.4134 | - | - | - | - | +| 0.4390 | 446 | 1.0244 | - | - | - | - | +| 0.4400 | 447 | 1.103 | - | - | - | - | +| 0.4409 | 448 | 0.9848 | - | - | - | - | +| 0.4419 | 449 | 1.5089 | - | - | - | - | +| 0.4429 | 450 | 1.0422 | - | - | - | - | +| 0.4439 | 451 | 1.0462 | - | - | - | - | +| 0.4449 | 452 | 1.2857 | - | - | - | - | +| 0.4459 | 453 | 1.4132 | - | - | - | - | +| 0.4469 | 454 | 1.3061 | - | - | - | - | +| 0.4478 | 455 | 1.3977 | - | - | - | - | +| 0.4488 | 456 | 1.3557 | - | - | - | - | +| 0.4498 | 457 | 1.3595 | - | - | - | - | +| 0.4508 | 458 | 0.8647 | - | - | - | - | +| 0.4518 | 459 | 1.3905 | 1.2969 | 0.5433 | 0.4937 | 0.7094 | +| 0.4528 | 460 | 0.9467 | - | - | - | - | +| 0.4537 | 461 | 1.9372 | - | - | - | - | +| 0.4547 | 462 | 0.871 | - | - | - | - | +| 0.4557 | 463 | 1.2282 | - | - | - | - | +| 0.4567 | 464 | 1.3845 | - | - | - | - | +| 0.4577 | 465 | 1.2571 | - | - | - | - | +| 0.4587 | 466 | 1.2288 | - | - | - | - | +| 0.4596 | 467 | 1.1165 | - | - | - | - | +| 0.4606 | 468 | 1.8463 | - | - | - | - | +| 0.4616 | 469 | 0.9158 | - | - | - | - | +| 0.4626 | 470 | 0.8711 | - | - | - | - | +| 0.4636 | 471 | 1.4741 | - | - | - | - | +| 0.4646 | 472 | 0.914 | - | - | - | - | +| 0.4656 | 473 | 0.9435 | - | - | - | - | +| 0.4665 | 474 | 1.0876 | - | - | - | - | +| 0.4675 | 475 | 1.2365 | - | - | - | - | +| 0.4685 | 476 | 1.1237 | - | - | - | - | +| 0.4695 | 477 | 1.0097 | - | - | - | - | +| 0.4705 | 478 | 1.1548 | - | - | - | - | +| 0.4715 | 479 | 1.3203 | - | - | - | - | +| 0.4724 | 480 | 1.2533 | - | - | - | - | +| 0.4734 | 481 | 1.093 | - | - | - | - | +| 0.4744 | 482 | 1.2591 | - | - | - | - | +| 0.4754 | 483 | 0.6764 | - | - | - | - | +| 0.4764 | 484 | 0.8922 | - | - | - | - | +| 0.4774 | 485 | 0.8524 | - | - | - | - | +| 0.4783 | 486 | 1.2777 | - | - | - | - | +| 0.4793 | 487 | 1.1682 | - | - | - | - | +| 0.4803 | 488 | 0.8617 | - | - | - | - | +| 0.4813 | 489 | 1.0303 | - | - | - | - | +| 0.4823 | 490 | 0.9843 | - | - | - | - | +| 0.4833 | 491 | 1.2951 | - | - | - | - | +| 0.4843 | 492 | 1.7889 | - | - | - | - | +| 0.4852 | 493 | 1.118 | - | - | - | - | +| 0.4862 | 494 | 0.6772 | - | - | - | - | +| 0.4872 | 495 | 1.5058 | - | - | - | - | +| 0.4882 | 496 | 1.0068 | - | - | - | - | +| 0.4892 | 497 | 0.9024 | - | - | - | - | +| 0.4902 | 498 | 1.4816 | - | - | - | - | +| 0.4911 | 499 | 0.894 | - | - | - | - | +| 0.4921 | 500 | 1.1582 | - | - | - | - | +| 0.4931 | 501 | 1.4804 | - | - | - | - | +| 0.4941 | 502 | 1.2636 | - | - | - | - | +| 0.4951 | 503 | 1.0094 | - | - | - | - | +| 0.4961 | 504 | 0.7594 | - | - | - | - | +| 0.4970 | 505 | 1.2898 | - | - | - | - | +| 0.4980 | 506 | 1.3565 | - | - | - | - | +| 0.4990 | 507 | 1.0325 | - | - | - | - | +| 0.5 | 508 | 1.0519 | - | - | - | - | +| 0.5010 | 509 | 0.9802 | - | - | - | - | +| 0.5020 | 510 | 1.1117 | - | - | - | - | +| 0.5030 | 511 | 1.3585 | - | - | - | - | +| 0.5039 | 512 | 1.0381 | - | - | - | - | +| 0.5049 | 513 | 1.0171 | - | - | - | - | +| 0.5059 | 514 | 0.5678 | - | - | - | - | +| 0.5069 | 515 | 0.9347 | - | - | - | - | +| 0.5079 | 516 | 0.6305 | - | - | - | - | +| 0.5089 | 517 | 0.7072 | - | - | - | - | +| 0.5098 | 518 | 0.9746 | - | - | - | - | +| 0.5108 | 519 | 1.1782 | - | - | - | - | +| 0.5118 | 520 | 1.1354 | - | - | - | - | +| 0.5128 | 521 | 1.5752 | - | - | - | - | +| 0.5138 | 522 | 0.5952 | - | - | - | - | +| 0.5148 | 523 | 1.1171 | - | - | - | - | +| 0.5157 | 524 | 0.8234 | - | - | - | - | +| 0.5167 | 525 | 1.6701 | - | - | - | - | +| 0.5177 | 526 | 1.2111 | - | - | - | - | +| 0.5187 | 527 | 0.8299 | - | - | - | - | +| 0.5197 | 528 | 1.5734 | - | - | - | - | +| 0.5207 | 529 | 0.9172 | - | - | - | - | +| 0.5217 | 530 | 0.8025 | - | - | - | - | +| 0.5226 | 531 | 1.1499 | - | - | - | - | +| 0.5236 | 532 | 1.0328 | - | - | - | - | +| 0.5246 | 533 | 1.1305 | - | - | - | - | +| 0.5256 | 534 | 0.6715 | - | - | - | - | +| 0.5266 | 535 | 1.1361 | - | - | - | - | +| 0.5276 | 536 | 0.9132 | - | - | - | - | +| 0.5285 | 537 | 1.2195 | - | - | - | - | +| 0.5295 | 538 | 0.3731 | - | - | - | - | +| 0.5305 | 539 | 1.0005 | - | - | - | - | +| 0.5315 | 540 | 0.5519 | - | - | - | - | +| 0.5325 | 541 | 0.7529 | - | - | - | - | +| 0.5335 | 542 | 1.7004 | - | - | - | - | +| 0.5344 | 543 | 1.4667 | - | - | - | - | +| 0.5354 | 544 | 0.8349 | - | - | - | - | +| 0.5364 | 545 | 1.5575 | - | - | - | - | +| 0.5374 | 546 | 1.1703 | - | - | - | - | +| 0.5384 | 547 | 1.01 | - | - | - | - | +| 0.5394 | 548 | 1.1114 | - | - | - | - | +| 0.5404 | 549 | 0.516 | - | - | - | - | +| 0.5413 | 550 | 1.0422 | - | - | - | - | +| 0.5423 | 551 | 1.078 | - | - | - | - | +| 0.5433 | 552 | 1.0573 | - | - | - | - | +| 0.5443 | 553 | 0.9754 | - | - | - | - | +| 0.5453 | 554 | 0.9227 | - | - | - | - | +| 0.5463 | 555 | 1.5012 | - | - | - | - | +| 0.5472 | 556 | 1.0697 | - | - | - | - | +| 0.5482 | 557 | 1.4437 | - | - | - | - | +| 0.5492 | 558 | 1.0697 | - | - | - | - | +| 0.5502 | 559 | 0.8346 | - | - | - | - | +| 0.5512 | 560 | 0.6421 | - | - | - | - | +| 0.5522 | 561 | 0.6687 | - | - | - | - | +| 0.5531 | 562 | 0.982 | - | - | - | - | +| 0.5541 | 563 | 0.9299 | - | - | - | - | +| 0.5551 | 564 | 1.5852 | - | - | - | - | +| 0.5561 | 565 | 1.2132 | - | - | - | - | +| 0.5571 | 566 | 0.8426 | - | - | - | - | +| 0.5581 | 567 | 1.0496 | - | - | - | - | +| 0.5591 | 568 | 1.0436 | - | - | - | - | +| 0.5600 | 569 | 0.806 | - | - | - | - | +| 0.5610 | 570 | 0.6396 | - | - | - | - | +| 0.5620 | 571 | 1.6315 | - | - | - | - | +| 0.5630 | 572 | 1.3286 | - | - | - | - | +| 0.5640 | 573 | 0.7682 | - | - | - | - | +| 0.5650 | 574 | 0.7861 | - | - | - | - | +| 0.5659 | 575 | 1.0368 | - | - | - | - | +| 0.5669 | 576 | 1.1497 | - | - | - | - | +| 0.5679 | 577 | 0.9691 | - | - | - | - | +| 0.5689 | 578 | 0.7447 | - | - | - | - | +| 0.5699 | 579 | 1.3933 | - | - | - | - | +| 0.5709 | 580 | 1.0668 | - | - | - | - | +| 0.5719 | 581 | 0.6065 | - | - | - | - | +| 0.5728 | 582 | 0.9566 | - | - | - | - | +| 0.5738 | 583 | 0.7957 | - | - | - | - | +| 0.5748 | 584 | 1.0232 | - | - | - | - | +| 0.5758 | 585 | 1.4559 | - | - | - | - | +| 0.5768 | 586 | 0.8003 | - | - | - | - | +| 0.5778 | 587 | 0.9504 | - | - | - | - | +| 0.5787 | 588 | 1.5257 | - | - | - | - | +| 0.5797 | 589 | 0.5798 | - | - | - | - | +| 0.5807 | 590 | 0.8169 | - | - | - | - | +| 0.5817 | 591 | 1.1131 | - | - | - | - | +| 0.5827 | 592 | 1.2498 | - | - | - | - | +| 0.5837 | 593 | 0.8541 | - | - | - | - | +| 0.5846 | 594 | 1.0848 | - | - | - | - | +| 0.5856 | 595 | 0.8909 | - | - | - | - | +| 0.5866 | 596 | 0.7572 | - | - | - | - | +| 0.5876 | 597 | 1.3636 | - | - | - | - | +| 0.5886 | 598 | 0.8493 | - | - | - | - | +| 0.5896 | 599 | 0.9594 | - | - | - | - | +| 0.5906 | 600 | 1.1143 | - | - | - | - | +| 0.5915 | 601 | 0.7093 | - | - | - | - | +| 0.5925 | 602 | 1.0542 | - | - | - | - | +| 0.5935 | 603 | 1.0621 | - | - | - | - | +| 0.5945 | 604 | 0.6916 | - | - | - | - | +| 0.5955 | 605 | 1.0125 | - | - | - | - | +| 0.5965 | 606 | 0.8425 | - | - | - | - | +| 0.5974 | 607 | 1.2868 | - | - | - | - | +| 0.5984 | 608 | 1.3505 | - | - | - | - | +| 0.5994 | 609 | 1.2699 | - | - | - | - | +| 0.6004 | 610 | 1.1798 | - | - | - | - | +| 0.6014 | 611 | 1.3607 | - | - | - | - | +| 0.6024 | 612 | 1.0807 | 1.2167 | 0.5879 | 0.5143 | 0.7076 | +| 0.6033 | 613 | 1.4339 | - | - | - | - | +| 0.6043 | 614 | 1.1194 | - | - | - | - | +| 0.6053 | 615 | 1.0682 | - | - | - | - | +| 0.6063 | 616 | 1.0429 | - | - | - | - | +| 0.6073 | 617 | 1.2554 | - | - | - | - | +| 0.6083 | 618 | 1.2466 | - | - | - | - | +| 0.6093 | 619 | 1.1207 | - | - | - | - | +| 0.6102 | 620 | 0.9822 | - | - | - | - | +| 0.6112 | 621 | 1.7369 | - | - | - | - | +| 0.6122 | 622 | 1.3305 | - | - | - | - | +| 0.6132 | 623 | 0.9064 | - | - | - | - | +| 0.6142 | 624 | 0.7123 | - | - | - | - | +| 0.6152 | 625 | 0.7461 | - | - | - | - | +| 0.6161 | 626 | 0.8082 | - | - | - | - | +| 0.6171 | 627 | 1.0113 | - | - | - | - | +| 0.6181 | 628 | 0.9483 | - | - | - | - | +| 0.6191 | 629 | 0.9269 | - | - | - | - | +| 0.6201 | 630 | 1.3134 | - | - | - | - | +| 0.6211 | 631 | 0.7253 | - | - | - | - | +| 0.6220 | 632 | 0.809 | - | - | - | - | +| 0.6230 | 633 | 1.2514 | - | - | - | - | +| 0.6240 | 634 | 0.6718 | - | - | - | - | +| 0.625 | 635 | 0.6658 | - | - | - | - | +| 0.6260 | 636 | 1.3988 | - | - | - | - | +| 0.6270 | 637 | 0.7358 | - | - | - | - | +| 0.6280 | 638 | 0.7797 | - | - | - | - | +| 0.6289 | 639 | 1.048 | - | - | - | - | +| 0.6299 | 640 | 0.9559 | - | - | - | - | +| 0.6309 | 641 | 0.4561 | - | - | - | - | +| 0.6319 | 642 | 1.1078 | - | - | - | - | +| 0.6329 | 643 | 0.9724 | - | - | - | - | +| 0.6339 | 644 | 1.0702 | - | - | - | - | +| 0.6348 | 645 | 1.0911 | - | - | - | - | +| 0.6358 | 646 | 1.1584 | - | - | - | - | +| 0.6368 | 647 | 0.9063 | - | - | - | - | +| 0.6378 | 648 | 0.5036 | - | - | - | - | +| 0.6388 | 649 | 0.8331 | - | - | - | - | +| 0.6398 | 650 | 1.0772 | - | - | - | - | +| 0.6407 | 651 | 0.7466 | - | - | - | - | +| 0.6417 | 652 | 1.1614 | - | - | - | - | +| 0.6427 | 653 | 0.6319 | - | - | - | - | +| 0.6437 | 654 | 0.7519 | - | - | - | - | +| 0.6447 | 655 | 1.1067 | - | - | - | - | +| 0.6457 | 656 | 1.2561 | - | - | - | - | +| 0.6467 | 657 | 0.6509 | - | - | - | - | +| 0.6476 | 658 | 1.0201 | - | - | - | - | +| 0.6486 | 659 | 1.6782 | - | - | - | - | +| 0.6496 | 660 | 1.3718 | - | - | - | - | +| 0.6506 | 661 | 0.6883 | - | - | - | - | +| 0.6516 | 662 | 1.0951 | - | - | - | - | +| 0.6526 | 663 | 1.2543 | - | - | - | - | +| 0.6535 | 664 | 1.2208 | - | - | - | - | +| 0.6545 | 665 | 0.6009 | - | - | - | - | +| 0.6555 | 666 | 1.1146 | - | - | - | - | +| 0.6565 | 667 | 1.0411 | - | - | - | - | +| 0.6575 | 668 | 0.6938 | - | - | - | - | +| 0.6585 | 669 | 1.0415 | - | - | - | - | +| 0.6594 | 670 | 0.4991 | - | - | - | - | +| 0.6604 | 671 | 1.4716 | - | - | - | - | +| 0.6614 | 672 | 0.745 | - | - | - | - | +| 0.6624 | 673 | 1.5687 | - | - | - | - | +| 0.6634 | 674 | 0.7606 | - | - | - | - | +| 0.6644 | 675 | 0.2446 | - | - | - | - | +| 0.6654 | 676 | 0.4829 | - | - | - | - | +| 0.6663 | 677 | 1.0112 | - | - | - | - | +| 0.6673 | 678 | 1.3718 | - | - | - | - | +| 0.6683 | 679 | 1.3441 | - | - | - | - | +| 0.6693 | 680 | 0.5089 | - | - | - | - | +| 0.6703 | 681 | 0.9052 | - | - | - | - | +| 0.6713 | 682 | 0.7006 | - | - | - | - | +| 0.6722 | 683 | 1.2755 | - | - | - | - | +| 0.6732 | 684 | 0.8308 | - | - | - | - | +| 0.6742 | 685 | 0.797 | - | - | - | - | +| 0.6752 | 686 | 0.5807 | - | - | - | - | +| 0.6762 | 687 | 0.9666 | - | - | - | - | +| 0.6772 | 688 | 1.0587 | - | - | - | - | +| 0.6781 | 689 | 1.1675 | - | - | - | - | +| 0.6791 | 690 | 0.725 | - | - | - | - | +| 0.6801 | 691 | 0.9958 | - | - | - | - | +| 0.6811 | 692 | 1.13 | - | - | - | - | +| 0.6821 | 693 | 1.6021 | - | - | - | - | +| 0.6831 | 694 | 0.8968 | - | - | - | - | +| 0.6841 | 695 | 0.9741 | - | - | - | - | +| 0.6850 | 696 | 1.1929 | - | - | - | - | +| 0.6860 | 697 | 0.6117 | - | - | - | - | +| 0.6870 | 698 | 0.9741 | - | - | - | - | +| 0.6880 | 699 | 0.9963 | - | - | - | - | +| 0.6890 | 700 | 0.6098 | - | - | - | - | +| 0.6900 | 701 | 0.9233 | - | - | - | - | +| 0.6909 | 702 | 1.4652 | - | - | - | - | +| 0.6919 | 703 | 1.3325 | - | - | - | - | +| 0.6929 | 704 | 1.1559 | - | - | - | - | +| 0.6939 | 705 | 1.021 | - | - | - | - | +| 0.6949 | 706 | 1.1437 | - | - | - | - | +| 0.6959 | 707 | 1.5533 | - | - | - | - | +| 0.6969 | 708 | 0.4733 | - | - | - | - | +| 0.6978 | 709 | 1.4539 | - | - | - | - | +| 0.6988 | 710 | 1.132 | - | - | - | - | +| 0.6998 | 711 | 1.315 | - | - | - | - | +| 0.7008 | 712 | 0.6671 | - | - | - | - | +| 0.7018 | 713 | 1.0689 | - | - | - | - | +| 0.7028 | 714 | 1.2344 | - | - | - | - | +| 0.7037 | 715 | 0.9918 | - | - | - | - | +| 0.7047 | 716 | 0.6537 | - | - | - | - | +| 0.7057 | 717 | 1.4362 | - | - | - | - | +| 0.7067 | 718 | 1.2486 | - | - | - | - | +| 0.7077 | 719 | 0.6777 | - | - | - | - | +| 0.7087 | 720 | 0.965 | - | - | - | - | +| 0.7096 | 721 | 1.1881 | - | - | - | - | +| 0.7106 | 722 | 1.2064 | - | - | - | - | +| 0.7116 | 723 | 0.5049 | - | - | - | - | +| 0.7126 | 724 | 0.7258 | - | - | - | - | +| 0.7136 | 725 | 0.458 | - | - | - | - | +| 0.7146 | 726 | 1.0756 | - | - | - | - | +| 0.7156 | 727 | 0.8171 | - | - | - | - | +| 0.7165 | 728 | 0.786 | - | - | - | - | +| 0.7175 | 729 | 1.3556 | - | - | - | - | +| 0.7185 | 730 | 1.181 | - | - | - | - | +| 0.7195 | 731 | 1.0563 | - | - | - | - | +| 0.7205 | 732 | 0.5951 | - | - | - | - | +| 0.7215 | 733 | 0.8533 | - | - | - | - | +| 0.7224 | 734 | 0.6561 | - | - | - | - | +| 0.7234 | 735 | 1.1081 | - | - | - | - | +| 0.7244 | 736 | 0.6016 | - | - | - | - | +| 0.7254 | 737 | 0.6155 | - | - | - | - | +| 0.7264 | 738 | 0.2202 | - | - | - | - | +| 0.7274 | 739 | 1.1072 | - | - | - | - | +| 0.7283 | 740 | 1.0147 | - | - | - | - | +| 0.7293 | 741 | 0.2117 | - | - | - | - | +| 0.7303 | 742 | 1.3508 | - | - | - | - | +| 0.7313 | 743 | 0.7085 | - | - | - | - | +| 0.7323 | 744 | 0.7357 | - | - | - | - | +| 0.7333 | 745 | 1.0121 | - | - | - | - | +| 0.7343 | 746 | 1.2527 | - | - | - | - | +| 0.7352 | 747 | 1.5227 | - | - | - | - | +| 0.7362 | 748 | 1.2253 | - | - | - | - | +| 0.7372 | 749 | 0.8419 | - | - | - | - | +| 0.7382 | 750 | 0.5649 | - | - | - | - | +| 0.7392 | 751 | 1.3501 | - | - | - | - | +| 0.7402 | 752 | 1.042 | - | - | - | - | +| 0.7411 | 753 | 1.1964 | - | - | - | - | +| 0.7421 | 754 | 1.1352 | - | - | - | - | +| 0.7431 | 755 | 0.8928 | - | - | - | - | +| 0.7441 | 756 | 0.7438 | - | - | - | - | +| 0.7451 | 757 | 1.4773 | - | - | - | - | +| 0.7461 | 758 | 1.196 | - | - | - | - | +| 0.7470 | 759 | 1.1562 | - | - | - | - | +| 0.7480 | 760 | 0.8362 | - | - | - | - | +| 0.7490 | 761 | 0.904 | - | - | - | - | +| 0.75 | 762 | 0.855 | - | - | - | - | +| 0.7510 | 763 | 0.748 | - | - | - | - | +| 0.7520 | 764 | 0.6261 | - | - | - | - | +| 0.7530 | 765 | 1.1903 | 1.1807 | 0.5774 | 0.5204 | 0.7123 | +| 0.7539 | 766 | 0.8415 | - | - | - | - | +| 0.7549 | 767 | 0.712 | - | - | - | - | +| 0.7559 | 768 | 1.4149 | - | - | - | - | +| 0.7569 | 769 | 0.844 | - | - | - | - | +| 0.7579 | 770 | 0.9184 | - | - | - | - | +| 0.7589 | 771 | 0.9229 | - | - | - | - | +| 0.7598 | 772 | 1.3872 | - | - | - | - | +| 0.7608 | 773 | 0.7914 | - | - | - | - | +| 0.7618 | 774 | 0.8064 | - | - | - | - | +| 0.7628 | 775 | 1.0489 | - | - | - | - | +| 0.7638 | 776 | 1.0517 | - | - | - | - | +| 0.7648 | 777 | 0.9025 | - | - | - | - | +| 0.7657 | 778 | 0.7241 | - | - | - | - | +| 0.7667 | 779 | 1.0115 | - | - | - | - | +| 0.7677 | 780 | 1.1583 | - | - | - | - | +| 0.7687 | 781 | 1.0957 | - | - | - | - | +| 0.7697 | 782 | 0.8654 | - | - | - | - | +| 0.7707 | 783 | 1.1943 | - | - | - | - | +| 0.7717 | 784 | 0.9565 | - | - | - | - | +| 0.7726 | 785 | 1.0079 | - | - | - | - | +| 0.7736 | 786 | 1.3196 | - | - | - | - | +| 0.7746 | 787 | 0.8066 | - | - | - | - | +| 0.7756 | 788 | 1.1875 | - | - | - | - | +| 0.7766 | 789 | 0.9068 | - | - | - | - | +| 0.7776 | 790 | 0.9388 | - | - | - | - | +| 0.7785 | 791 | 1.5462 | - | - | - | - | +| 0.7795 | 792 | 0.9369 | - | - | - | - | +| 0.7805 | 793 | 1.6793 | - | - | - | - | +| 0.7815 | 794 | 1.0793 | - | - | - | - | +| 0.7825 | 795 | 0.7758 | - | - | - | - | +| 0.7835 | 796 | 0.6 | - | - | - | - | +| 0.7844 | 797 | 0.7136 | - | - | - | - | +| 0.7854 | 798 | 0.813 | - | - | - | - | +| 0.7864 | 799 | 0.8777 | - | - | - | - | +| 0.7874 | 800 | 1.119 | - | - | - | - | +| 0.7884 | 801 | 0.5711 | - | - | - | - | +| 0.7894 | 802 | 0.6798 | - | - | - | - | +| 0.7904 | 803 | 0.8154 | - | - | - | - | +| 0.7913 | 804 | 0.3272 | - | - | - | - | +| 0.7923 | 805 | 0.9906 | - | - | - | - | +| 0.7933 | 806 | 1.0634 | - | - | - | - | +| 0.7943 | 807 | 0.9913 | - | - | - | - | +| 0.7953 | 808 | 1.0392 | - | - | - | - | +| 0.7963 | 809 | 0.7832 | - | - | - | - | +| 0.7972 | 810 | 0.4475 | - | - | - | - | +| 0.7982 | 811 | 0.708 | - | - | - | - | +| 0.7992 | 812 | 0.8815 | - | - | - | - | +| 0.8002 | 813 | 1.3039 | - | - | - | - | +| 0.8012 | 814 | 1.3863 | - | - | - | - | +| 0.8022 | 815 | 1.0562 | - | - | - | - | +| 0.8031 | 816 | 0.7251 | - | - | - | - | +| 0.8041 | 817 | 0.6901 | - | - | - | - | +| 0.8051 | 818 | 0.7074 | - | - | - | - | +| 0.8061 | 819 | 0.5985 | - | - | - | - | +| 0.8071 | 820 | 0.674 | - | - | - | - | +| 0.8081 | 821 | 0.6977 | - | - | - | - | +| 0.8091 | 822 | 0.6939 | - | - | - | - | +| 0.8100 | 823 | 0.7825 | - | - | - | - | +| 0.8110 | 824 | 0.9403 | - | - | - | - | +| 0.8120 | 825 | 0.5739 | - | - | - | - | +| 0.8130 | 826 | 1.2775 | - | - | - | - | +| 0.8140 | 827 | 0.7558 | - | - | - | - | +| 0.8150 | 828 | 0.9289 | - | - | - | - | +| 0.8159 | 829 | 0.7306 | - | - | - | - | +| 0.8169 | 830 | 0.8876 | - | - | - | - | +| 0.8179 | 831 | 0.9344 | - | - | - | - | +| 0.8189 | 832 | 0.8379 | - | - | - | - | +| 0.8199 | 833 | 0.3775 | - | - | - | - | +| 0.8209 | 834 | 0.4071 | - | - | - | - | +| 0.8219 | 835 | 0.5419 | - | - | - | - | +| 0.8228 | 836 | 0.7428 | - | - | - | - | +| 0.8238 | 837 | 0.905 | - | - | - | - | +| 0.8248 | 838 | 0.605 | - | - | - | - | +| 0.8258 | 839 | 1.6087 | - | - | - | - | +| 0.8268 | 840 | 0.5758 | - | - | - | - | +| 0.8278 | 841 | 0.9991 | - | - | - | - | +| 0.8287 | 842 | 1.3015 | - | - | - | - | +| 0.8297 | 843 | 0.8529 | - | - | - | - | +| 0.8307 | 844 | 0.8257 | - | - | - | - | +| 0.8317 | 845 | 0.8513 | - | - | - | - | +| 0.8327 | 846 | 0.9995 | - | - | - | - | +| 0.8337 | 847 | 1.0182 | - | - | - | - | +| 0.8346 | 848 | 0.6523 | - | - | - | - | +| 0.8356 | 849 | 0.8436 | - | - | - | - | +| 0.8366 | 850 | 1.4555 | - | - | - | - | +| 0.8376 | 851 | 0.6176 | - | - | - | - | +| 0.8386 | 852 | 1.1224 | - | - | - | - | +| 0.8396 | 853 | 0.5743 | - | - | - | - | +| 0.8406 | 854 | 0.6488 | - | - | - | - | +| 0.8415 | 855 | 0.6553 | - | - | - | - | +| 0.8425 | 856 | 1.0901 | - | - | - | - | +| 0.8435 | 857 | 1.2568 | - | - | - | - | +| 0.8445 | 858 | 0.7643 | - | - | - | - | +| 0.8455 | 859 | 0.3966 | - | - | - | - | +| 0.8465 | 860 | 0.6586 | - | - | - | - | +| 0.8474 | 861 | 0.8597 | - | - | - | - | +| 0.8484 | 862 | 1.237 | - | - | - | - | +| 0.8494 | 863 | 0.9306 | - | - | - | - | +| 0.8504 | 864 | 0.7643 | - | - | - | - | +| 0.8514 | 865 | 0.7402 | - | - | - | - | +| 0.8524 | 866 | 0.9191 | - | - | - | - | +| 0.8533 | 867 | 0.9644 | - | - | - | - | +| 0.8543 | 868 | 0.7933 | - | - | - | - | +| 0.8553 | 869 | 1.5964 | - | - | - | - | +| 0.8563 | 870 | 0.8953 | - | - | - | - | +| 0.8573 | 871 | 1.0073 | - | - | - | - | +| 0.8583 | 872 | 0.517 | - | - | - | - | +| 0.8593 | 873 | 0.8879 | - | - | - | - | +| 0.8602 | 874 | 1.5371 | - | - | - | - | +| 0.8612 | 875 | 0.9743 | - | - | - | - | +| 0.8622 | 876 | 1.0717 | - | - | - | - | +| 0.8632 | 877 | 0.6625 | - | - | - | - | +| 0.8642 | 878 | 0.8521 | - | - | - | - | +| 0.8652 | 879 | 0.7955 | - | - | - | - | +| 0.8661 | 880 | 0.9416 | - | - | - | - | +| 0.8671 | 881 | 0.8257 | - | - | - | - | +| 0.8681 | 882 | 1.3879 | - | - | - | - | +| 0.8691 | 883 | 0.9457 | - | - | - | - | +| 0.8701 | 884 | 0.891 | - | - | - | - | +| 0.8711 | 885 | 0.9427 | - | - | - | - | +| 0.8720 | 886 | 0.8526 | - | - | - | - | +| 0.8730 | 887 | 1.2298 | - | - | - | - | +| 0.8740 | 888 | 0.6241 | - | - | - | - | +| 0.875 | 889 | 0.7055 | - | - | - | - | +| 0.8760 | 890 | 0.9713 | - | - | - | - | +| 0.8770 | 891 | 1.0591 | - | - | - | - | +| 0.8780 | 892 | 1.0597 | - | - | - | - | +| 0.8789 | 893 | 1.1631 | - | - | - | - | +| 0.8799 | 894 | 0.6098 | - | - | - | - | +| 0.8809 | 895 | 1.1498 | - | - | - | - | +| 0.8819 | 896 | 0.5379 | - | - | - | - | +| 0.8829 | 897 | 0.7921 | - | - | - | - | +| 0.8839 | 898 | 0.9092 | - | - | - | - | +| 0.8848 | 899 | 1.0348 | - | - | - | - | +| 0.8858 | 900 | 0.9087 | - | - | - | - | +| 0.8868 | 901 | 1.5328 | - | - | - | - | +| 0.8878 | 902 | 0.8664 | - | - | - | - | +| 0.8888 | 903 | 0.6873 | - | - | - | - | +| 0.8898 | 904 | 1.1763 | - | - | - | - | +| 0.8907 | 905 | 1.2853 | - | - | - | - | +| 0.8917 | 906 | 0.8163 | - | - | - | - | +| 0.8927 | 907 | 0.7383 | - | - | - | - | +| 0.8937 | 908 | 0.7833 | - | - | - | - | +| 0.8947 | 909 | 1.078 | - | - | - | - | +| 0.8957 | 910 | 0.6647 | - | - | - | - | +| 0.8967 | 911 | 1.0016 | - | - | - | - | +| 0.8976 | 912 | 0.8432 | - | - | - | - | +| 0.8986 | 913 | 0.9927 | - | - | - | - | +| 0.8996 | 914 | 0.4985 | - | - | - | - | +| 0.9006 | 915 | 0.1726 | - | - | - | - | +| 0.9016 | 916 | 0.9437 | - | - | - | - | +| 0.9026 | 917 | 0.2565 | - | - | - | - | +| 0.9035 | 918 | 0.6238 | 1.1640 | 0.6064 | 0.5481 | 0.6999 | +| 0.9045 | 919 | 0.9311 | - | - | - | - | +| 0.9055 | 920 | 1.6868 | - | - | - | - | +| 0.9065 | 921 | 0.8606 | - | - | - | - | +| 0.9075 | 922 | 0.4508 | - | - | - | - | +| 0.9085 | 923 | 0.6556 | - | - | - | - | +| 0.9094 | 924 | 0.5244 | - | - | - | - | +| 0.9104 | 925 | 0.6512 | - | - | - | - | +| 0.9114 | 926 | 0.6594 | - | - | - | - | +| 0.9124 | 927 | 1.0091 | - | - | - | - | +| 0.9134 | 928 | 0.6465 | - | - | - | - | +| 0.9144 | 929 | 0.798 | - | - | - | - | +| 0.9154 | 930 | 1.0069 | - | - | - | - | +| 0.9163 | 931 | 0.7407 | - | - | - | - | +| 0.9173 | 932 | 0.7979 | - | - | - | - | +| 0.9183 | 933 | 1.688 | - | - | - | - | +| 0.9193 | 934 | 0.8505 | - | - | - | - | +| 0.9203 | 935 | 0.8101 | - | - | - | - | +| 0.9213 | 936 | 0.8542 | - | - | - | - | +| 0.9222 | 937 | 0.956 | - | - | - | - | +| 0.9232 | 938 | 0.7072 | - | - | - | - | +| 0.9242 | 939 | 0.6316 | - | - | - | - | +| 0.9252 | 940 | 1.3868 | - | - | - | - | +| 0.9262 | 941 | 0.3854 | - | - | - | - | +| 0.9272 | 942 | 1.3662 | - | - | - | - | +| 0.9281 | 943 | 1.0133 | - | - | - | - | +| 0.9291 | 944 | 0.8543 | - | - | - | - | +| 0.9301 | 945 | 0.9697 | - | - | - | - | +| 0.9311 | 946 | 0.7126 | - | - | - | - | +| 0.9321 | 947 | 1.0513 | - | - | - | - | +| 0.9331 | 948 | 0.6503 | - | - | - | - | +| 0.9341 | 949 | 0.894 | - | - | - | - | +| 0.9350 | 950 | 0.4608 | - | - | - | - | +| 0.9360 | 951 | 0.8966 | - | - | - | - | +| 0.9370 | 952 | 0.7412 | - | - | - | - | +| 0.9380 | 953 | 0.9692 | - | - | - | - | +| 0.9390 | 954 | 0.5728 | - | - | - | - | +| 0.9400 | 955 | 0.6812 | - | - | - | - | +| 0.9409 | 956 | 0.5401 | - | - | - | - | +| 0.9419 | 957 | 0.9315 | - | - | - | - | +| 0.9429 | 958 | 0.6438 | - | - | - | - | +| 0.9439 | 959 | 1.0856 | - | - | - | - | +| 0.9449 | 960 | 0.7523 | - | - | - | - | +| 0.9459 | 961 | 0.5826 | - | - | - | - | +| 0.9469 | 962 | 0.9469 | - | - | - | - | +| 0.9478 | 963 | 0.647 | - | - | - | - | +| 0.9488 | 964 | 1.8012 | - | - | - | - | +| 0.9498 | 965 | 0.6264 | - | - | - | - | +| 0.9508 | 966 | 1.2779 | - | - | - | - | +| 0.9518 | 967 | 0.8312 | - | - | - | - | +| 0.9528 | 968 | 0.6442 | - | - | - | - | +| 0.9537 | 969 | 0.5953 | - | - | - | - | +| 0.9547 | 970 | 0.4099 | - | - | - | - | +| 0.9557 | 971 | 0.9008 | - | - | - | - | +| 0.9567 | 972 | 1.4286 | - | - | - | - | +| 0.9577 | 973 | 0.9222 | - | - | - | - | +| 0.9587 | 974 | 0.8414 | - | - | - | - | +| 0.9596 | 975 | 0.7063 | - | - | - | - | +| 0.9606 | 976 | 0.6207 | - | - | - | - | +| 0.9616 | 977 | 0.8273 | - | - | - | - | +| 0.9626 | 978 | 0.8464 | - | - | - | - | +| 0.9636 | 979 | 0.8247 | - | - | - | - | +| 0.9646 | 980 | 0.7133 | - | - | - | - | +| 0.9656 | 981 | 0.7903 | - | - | - | - | +| 0.9665 | 982 | 1.0951 | - | - | - | - | +| 0.9675 | 983 | 1.099 | - | - | - | - | +| 0.9685 | 984 | 0.5865 | - | - | - | - | +| 0.9695 | 985 | 0.8717 | - | - | - | - | +| 0.9705 | 986 | 0.9441 | - | - | - | - | +| 0.9715 | 987 | 0.9627 | - | - | - | - | +| 0.9724 | 988 | 1.039 | - | - | - | - | +| 0.9734 | 989 | 1.2139 | - | - | - | - | +| 0.9744 | 990 | 0.6284 | - | - | - | - | +| 0.9754 | 991 | 0.914 | - | - | - | - | +| 0.9764 | 992 | 0.4021 | - | - | - | - | +| 0.9774 | 993 | 0.5822 | - | - | - | - | +| 0.9783 | 994 | 0.5666 | - | - | - | - | +| 0.9793 | 995 | 1.3606 | - | - | - | - | +| 0.9803 | 996 | 0.7811 | - | - | - | - | +| 0.9813 | 997 | 0.3886 | - | - | - | - | +| 0.9823 | 998 | 1.5192 | - | - | - | - | +| 0.9833 | 999 | 0.3559 | - | - | - | - | +| 0.9843 | 1000 | 1.1638 | - | - | - | - | +| 0.9852 | 1001 | 0.5369 | - | - | - | - | +| 0.9862 | 1002 | 0.6804 | - | - | - | - | +| 0.9872 | 1003 | 0.5193 | - | - | - | - | +| 0.9882 | 1004 | 1.1559 | - | - | - | - | +| 0.9892 | 1005 | 0.7566 | - | - | - | - | +| 0.9902 | 1006 | 0.6815 | - | - | - | - | +| 0.9911 | 1007 | 1.0367 | - | - | - | - | +| 0.9921 | 1008 | 0.5514 | - | - | - | - | +| 0.9931 | 1009 | 0.7299 | - | - | - | - | +| 0.9941 | 1010 | 1.1474 | - | - | - | - | +| 0.9951 | 1011 | 0.5969 | - | - | - | - | +| 0.9961 | 1012 | 0.575 | - | - | - | - | +| 0.9970 | 1013 | 0.6357 | - | - | - | - | +| 0.9980 | 1014 | 0.8671 | - | - | - | - | +| 0.9990 | 1015 | 0.7021 | - | - | - | - | +| 1.0 | 1016 | 0.6608 | - | - | - | - | +| 1.0010 | 1017 | 1.0985 | - | - | - | - | +| 1.0020 | 1018 | 1.7193 | - | - | - | - | +| 1.0030 | 1019 | 0.3447 | - | - | - | - | +| 1.0039 | 1020 | 0.6574 | - | - | - | - | +| 1.0049 | 1021 | 0.7307 | - | - | - | - | +| 1.0059 | 1022 | 0.605 | - | - | - | - | +| 1.0069 | 1023 | 0.9642 | - | - | - | - | +| 1.0079 | 1024 | 0.7899 | - | - | - | - | +| 1.0089 | 1025 | 0.6818 | - | - | - | - | +| 1.0098 | 1026 | 0.9047 | - | - | - | - | +| 1.0108 | 1027 | 1.0251 | - | - | - | - | +| 1.0118 | 1028 | 0.635 | - | - | - | - | +| 1.0128 | 1029 | 0.9896 | - | - | - | - | +| 1.0138 | 1030 | 0.8988 | - | - | - | - | +| 1.0148 | 1031 | 0.6257 | - | - | - | - | +| 1.0157 | 1032 | 0.4369 | - | - | - | - | +| 1.0167 | 1033 | 0.7827 | - | - | - | - | +| 1.0177 | 1034 | 0.9601 | - | - | - | - | +| 1.0187 | 1035 | 0.9565 | - | - | - | - | +| 1.0197 | 1036 | 0.6667 | - | - | - | - | +| 1.0207 | 1037 | 0.4217 | - | - | - | - | +| 1.0217 | 1038 | 0.7592 | - | - | - | - | +| 1.0226 | 1039 | 0.8667 | - | - | - | - | +| 1.0236 | 1040 | 0.7705 | - | - | - | - | +| 1.0246 | 1041 | 0.9951 | - | - | - | - | +| 1.0256 | 1042 | 1.1144 | - | - | - | - | +| 1.0266 | 1043 | 1.0319 | - | - | - | - | +| 1.0276 | 1044 | 1.1595 | - | - | - | - | +| 1.0285 | 1045 | 0.6343 | - | - | - | - | +| 1.0295 | 1046 | 1.2074 | - | - | - | - | +| 1.0305 | 1047 | 0.8404 | - | - | - | - | +| 1.0315 | 1048 | 1.5037 | - | - | - | - | +| 1.0325 | 1049 | 0.4995 | - | - | - | - | +| 1.0335 | 1050 | 0.568 | - | - | - | - | +| 1.0344 | 1051 | 0.7489 | - | - | - | - | +| 1.0354 | 1052 | 0.7327 | - | - | - | - | +| 1.0364 | 1053 | 1.3957 | - | - | - | - | +| 1.0374 | 1054 | 1.0428 | - | - | - | - | +| 1.0384 | 1055 | 0.7656 | - | - | - | - | +| 1.0394 | 1056 | 1.1611 | - | - | - | - | +| 1.0404 | 1057 | 0.4786 | - | - | - | - | +| 1.0413 | 1058 | 0.5765 | - | - | - | - | +| 1.0423 | 1059 | 0.9421 | - | - | - | - | +| 1.0433 | 1060 | 0.7738 | - | - | - | - | +| 1.0443 | 1061 | 0.7882 | - | - | - | - | +| 1.0453 | 1062 | 0.9898 | - | - | - | - | +| 1.0463 | 1063 | 0.7618 | - | - | - | - | +| 1.0472 | 1064 | 0.5399 | - | - | - | - | +| 1.0482 | 1065 | 0.8189 | - | - | - | - | +| 1.0492 | 1066 | 0.4776 | - | - | - | - | +| 1.0502 | 1067 | 0.4333 | - | - | - | - | +| 1.0512 | 1068 | 0.4207 | - | - | - | - | +| 1.0522 | 1069 | 1.0206 | - | - | - | - | +| 1.0531 | 1070 | 0.4865 | - | - | - | - | +| 1.0541 | 1071 | 0.897 | 1.0710 | 0.6346 | 0.5430 | 0.6916 | +| 1.0551 | 1072 | 0.8402 | - | - | - | - | +| 1.0561 | 1073 | 0.7688 | - | - | - | - | +| 1.0571 | 1074 | 0.2184 | - | - | - | - | +| 1.0581 | 1075 | 0.863 | - | - | - | - | +| 1.0591 | 1076 | 0.63 | - | - | - | - | +| 1.0600 | 1077 | 0.6715 | - | - | - | - | +| 1.0610 | 1078 | 0.5824 | - | - | - | - | +| 1.0620 | 1079 | 0.4253 | - | - | - | - | +| 1.0630 | 1080 | 0.7626 | - | - | - | - | +| 1.0640 | 1081 | 0.6314 | - | - | - | - | +| 1.0650 | 1082 | 0.6581 | - | - | - | - | +| 1.0659 | 1083 | 0.4651 | - | - | - | - | +| 1.0669 | 1084 | 1.3387 | - | - | - | - | +| 1.0679 | 1085 | 0.8808 | - | - | - | - | +| 1.0689 | 1086 | 0.7236 | - | - | - | - | +| 1.0699 | 1087 | 0.7806 | - | - | - | - | +| 1.0709 | 1088 | 1.3413 | - | - | - | - | +| 1.0719 | 1089 | 0.4676 | - | - | - | - | +| 1.0728 | 1090 | 0.3322 | - | - | - | - | +| 1.0738 | 1091 | 0.3032 | - | - | - | - | +| 1.0748 | 1092 | 0.7566 | - | - | - | - | +| 1.0758 | 1093 | 1.2515 | - | - | - | - | +| 1.0768 | 1094 | 1.1035 | - | - | - | - | +| 1.0778 | 1095 | 0.5504 | - | - | - | - | +| 1.0787 | 1096 | 1.2568 | - | - | - | - | +| 1.0797 | 1097 | 1.0059 | - | - | - | - | +| 1.0807 | 1098 | 0.9695 | - | - | - | - | +| 1.0817 | 1099 | 0.5669 | - | - | - | - | +| 1.0827 | 1100 | 0.6268 | - | - | - | - | +| 1.0837 | 1101 | 1.013 | - | - | - | - | +| 1.0846 | 1102 | 1.5633 | - | - | - | - | +| 1.0856 | 1103 | 1.3625 | - | - | - | - | +| 1.0866 | 1104 | 0.7289 | - | - | - | - | +| 1.0876 | 1105 | 1.0045 | - | - | - | - | +| 1.0886 | 1106 | 1.2376 | - | - | - | - | +| 1.0896 | 1107 | 0.4695 | - | - | - | - | +| 1.0906 | 1108 | 1.1059 | - | - | - | - | +| 1.0915 | 1109 | 0.6343 | - | - | - | - | +| 1.0925 | 1110 | 0.7101 | - | - | - | - | +| 1.0935 | 1111 | 0.6253 | - | - | - | - | +| 1.0945 | 1112 | 1.1293 | - | - | - | - | +| 1.0955 | 1113 | 0.5038 | - | - | - | - | +| 1.0965 | 1114 | 0.8907 | - | - | - | - | +| 1.0974 | 1115 | 0.553 | - | - | - | - | +| 1.0984 | 1116 | 0.8102 | - | - | - | - | +| 1.0994 | 1117 | 0.904 | - | - | - | - | +| 1.1004 | 1118 | 0.5524 | - | - | - | - | +| 1.1014 | 1119 | 1.1347 | - | - | - | - | +| 1.1024 | 1120 | 0.4371 | - | - | - | - | +| 1.1033 | 1121 | 0.875 | - | - | - | - | +| 1.1043 | 1122 | 1.3085 | - | - | - | - | +| 1.1053 | 1123 | 0.7923 | - | - | - | - | +| 1.1063 | 1124 | 0.5889 | - | - | - | - | +| 1.1073 | 1125 | 0.5114 | - | - | - | - | +| 1.1083 | 1126 | 0.6616 | - | - | - | - | +| 1.1093 | 1127 | 0.9752 | - | - | - | - | +| 1.1102 | 1128 | 0.6389 | - | - | - | - | +| 1.1112 | 1129 | 0.9866 | - | - | - | - | +| 1.1122 | 1130 | 0.365 | - | - | - | - | +| 1.1132 | 1131 | 0.6243 | - | - | - | - | +| 1.1142 | 1132 | 0.5302 | - | - | - | - | +| 1.1152 | 1133 | 0.5457 | - | - | - | - | +| 1.1161 | 1134 | 0.7722 | - | - | - | - | +| 1.1171 | 1135 | 1.2737 | - | - | - | - | +| 1.1181 | 1136 | 0.7274 | - | - | - | - | +| 1.1191 | 1137 | 0.9102 | - | - | - | - | +| 1.1201 | 1138 | 0.673 | - | - | - | - | +| 1.1211 | 1139 | 0.5895 | - | - | - | - | +| 1.1220 | 1140 | 0.6718 | - | - | - | - | +| 1.1230 | 1141 | 0.9109 | - | - | - | - | +| 1.1240 | 1142 | 0.4086 | - | - | - | - | +| 1.125 | 1143 | 0.8096 | - | - | - | - | +| 1.1260 | 1144 | 0.6197 | - | - | - | - | +| 1.1270 | 1145 | 0.6548 | - | - | - | - | +| 1.1280 | 1146 | 0.7632 | - | - | - | - | +| 1.1289 | 1147 | 0.3288 | - | - | - | - | +| 1.1299 | 1148 | 0.6114 | - | - | - | - | +| 1.1309 | 1149 | 0.4121 | - | - | - | - | +| 1.1319 | 1150 | 0.6772 | - | - | - | - | +| 1.1329 | 1151 | 0.9555 | - | - | - | - | +| 1.1339 | 1152 | 0.5712 | - | - | - | - | +| 1.1348 | 1153 | 0.8391 | - | - | - | - | +| 1.1358 | 1154 | 0.6745 | - | - | - | - | +| 1.1368 | 1155 | 0.5267 | - | - | - | - | +| 1.1378 | 1156 | 1.0252 | - | - | - | - | +| 1.1388 | 1157 | 0.4004 | - | - | - | - | +| 1.1398 | 1158 | 0.925 | - | - | - | - | +| 1.1407 | 1159 | 0.6741 | - | - | - | - | +| 1.1417 | 1160 | 0.5167 | - | - | - | - | +| 1.1427 | 1161 | 0.6953 | - | - | - | - | +| 1.1437 | 1162 | 0.5611 | - | - | - | - | +| 1.1447 | 1163 | 1.0161 | - | - | - | - | +| 1.1457 | 1164 | 1.3154 | - | - | - | - | +| 1.1467 | 1165 | 0.6765 | - | - | - | - | +| 1.1476 | 1166 | 0.8017 | - | - | - | - | +| 1.1486 | 1167 | 0.8971 | - | - | - | - | +| 1.1496 | 1168 | 0.4928 | - | - | - | - | +| 1.1506 | 1169 | 0.6463 | - | - | - | - | +| 1.1516 | 1170 | 1.1188 | - | - | - | - | +| 1.1526 | 1171 | 0.7682 | - | - | - | - | +| 1.1535 | 1172 | 0.4076 | - | - | - | - | +| 1.1545 | 1173 | 0.6429 | - | - | - | - | +| 1.1555 | 1174 | 1.1348 | - | - | - | - | +| 1.1565 | 1175 | 0.4246 | - | - | - | - | +| 1.1575 | 1176 | 0.8091 | - | - | - | - | +| 1.1585 | 1177 | 0.3452 | - | - | - | - | +| 1.1594 | 1178 | 0.7898 | - | - | - | - | +| 1.1604 | 1179 | 0.5909 | - | - | - | - | +| 1.1614 | 1180 | 1.0561 | - | - | - | - | +| 1.1624 | 1181 | 1.0296 | - | - | - | - | +| 1.1634 | 1182 | 0.5792 | - | - | - | - | +| 1.1644 | 1183 | 0.5314 | - | - | - | - | +| 1.1654 | 1184 | 0.8981 | - | - | - | - | +| 1.1663 | 1185 | 0.8561 | - | - | - | - | +| 1.1673 | 1186 | 0.6095 | - | - | - | - | +| 1.1683 | 1187 | 0.9399 | - | - | - | - | +| 1.1693 | 1188 | 1.1345 | - | - | - | - | +| 1.1703 | 1189 | 0.4627 | - | - | - | - | +| 1.1713 | 1190 | 0.6207 | - | - | - | - | +| 1.1722 | 1191 | 0.6967 | - | - | - | - | +| 1.1732 | 1192 | 0.498 | - | - | - | - | +| 1.1742 | 1193 | 0.7233 | - | - | - | - | +| 1.1752 | 1194 | 0.443 | - | - | - | - | +| 1.1762 | 1195 | 0.6022 | - | - | - | - | +| 1.1772 | 1196 | 0.5702 | - | - | - | - | +| 1.1781 | 1197 | 0.8733 | - | - | - | - | +| 1.1791 | 1198 | 0.432 | - | - | - | - | +| 1.1801 | 1199 | 0.6508 | - | - | - | - | +| 1.1811 | 1200 | 0.8595 | - | - | - | - | +| 1.1821 | 1201 | 0.6948 | - | - | - | - | +| 1.1831 | 1202 | 0.6306 | - | - | - | - | +| 1.1841 | 1203 | 0.9615 | - | - | - | - | +| 1.1850 | 1204 | 0.5652 | - | - | - | - | +| 1.1860 | 1205 | 0.4482 | - | - | - | - | +| 1.1870 | 1206 | 0.8112 | - | - | - | - | +| 1.1880 | 1207 | 0.6432 | - | - | - | - | +| 1.1890 | 1208 | 0.6797 | - | - | - | - | +| 1.1900 | 1209 | 0.4737 | - | - | - | - | +| 1.1909 | 1210 | 0.5752 | - | - | - | - | +| 1.1919 | 1211 | 0.4858 | - | - | - | - | +| 1.1929 | 1212 | 0.4213 | - | - | - | - | +| 1.1939 | 1213 | 0.3251 | - | - | - | - | +| 1.1949 | 1214 | 0.8442 | - | - | - | - | +| 1.1959 | 1215 | 0.4813 | - | - | - | - | +| 1.1969 | 1216 | 0.4635 | - | - | - | - | +| 1.1978 | 1217 | 0.4121 | - | - | - | - | +| 1.1988 | 1218 | 0.8145 | - | - | - | - | +| 1.1998 | 1219 | 1.7243 | - | - | - | - | +| 1.2008 | 1220 | 1.0789 | - | - | - | - | + +
+ +### Framework Versions +- Python: 3.10.12 +- Sentence Transformers: 3.2.1 +- Transformers: 4.44.2 +- PyTorch: 2.5.0+cu121 +- Accelerate: 0.34.2 +- Datasets: 3.0.2 +- Tokenizers: 0.19.1 + +## Citation + +### BibTeX + +#### Sentence Transformers +```bibtex +@inproceedings{reimers-2019-sentence-bert, + title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks", + author = "Reimers, Nils and Gurevych, Iryna", + booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing", + month = "11", + year = "2019", + publisher = "Association for Computational Linguistics", + url = "https://arxiv.org/abs/1908.10084", +} +``` + +#### GISTEmbedLoss +```bibtex +@misc{solatorio2024gistembed, + title={GISTEmbed: Guided In-sample Selection of Training Negatives for Text Embedding Fine-tuning}, + author={Aivin V. Solatorio}, + year={2024}, + eprint={2402.16829}, + archivePrefix={arXiv}, + primaryClass={cs.LG} +} +``` + + + + + + \ No newline at end of file