diff --git a/.gitattributes b/.gitattributes
index 36ecd5f0b77b62f0e1dc51fbde0fc412cb53ad1b..b6328ef3d5fb0d3a80b92094037342d0defbf35e 100644
--- a/.gitattributes
+++ b/.gitattributes
@@ -57,3 +57,11 @@ gdn_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn
gdn_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_z0tiim1_/attempt_0/5/stderr.log filter=lfs diff=lfs merge=lfs -text
gdn_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_z0tiim1_/attempt_0/6/stderr.log filter=lfs diff=lfs merge=lfs -text
gdn_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_z0tiim1_/attempt_0/7/stderr.log filter=lfs diff=lfs merge=lfs -text
+mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/0/stderr.log filter=lfs diff=lfs merge=lfs -text
+mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/1/stderr.log filter=lfs diff=lfs merge=lfs -text
+mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/2/stderr.log filter=lfs diff=lfs merge=lfs -text
+mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/3/stderr.log filter=lfs diff=lfs merge=lfs -text
+mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/4/stderr.log filter=lfs diff=lfs merge=lfs -text
+mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/5/stderr.log filter=lfs diff=lfs merge=lfs -text
+mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/6/stderr.log filter=lfs diff=lfs merge=lfs -text
+mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/7/stderr.log filter=lfs diff=lfs merge=lfs -text
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/8k-100.sh b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/8k-100.sh
new file mode 100644
index 0000000000000000000000000000000000000000..abbb1e7b6e8a4a839978ca4106f1cf4bcc85a9e7
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/8k-100.sh
@@ -0,0 +1,65 @@
+FLAME_PATH=/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame
+DATASET_ROOT=/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset
+TOKENIZER=/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer
+
+cd $FLAME_PATH
+source .venv/bin/activate
+
+# =========== train config ===========
+CONFIG=${1:-transformer_340M.json}
+SEQ_LEN=8192
+WARMUP_STEPS=100
+STEPS=95366
+LR=3e-4
+BATCH_SIZE=8
+GAS=2
+DECAY_TYPE=linear
+DECAY_RATIO=1
+NNODE=1
+NGPU=8
+LOG_RANK=0
+# ====================================
+
+# if jq command is not found, install it
+if ! command -v jq &> /dev/null; then
+ echo "jq could not be found, installing it..."
+ sudo yum install -y jq
+fi
+
+EXP_NAME=$(basename $CONFIG | sed 's/\.config//')-ctx${SEQ_LEN}-steps${STEPS}-lr${LR}-decay_type${DECAY_TYPE}-decay_ratio${DECAY_RATIO}-bs${BATCH_SIZE}-nn${NNODE}-gas${GAS}
+
+bash train.sh \
+ --job.config_file flame/models/fla.toml \
+ --job.dump_folder $FLAME_PATH/exp/$EXP_NAME \
+ --model.config $FLAME_PATH/configs/$CONFIG \
+ --model.tokenizer_path $TOKENIZER \
+ --optimizer.name AdamW \
+ --optimizer.eps 1e-8 \
+ --optimizer.lr $LR \
+ --lr_scheduler.warmup_steps $WARMUP_STEPS \
+ --lr_scheduler.lr_min 0.01 \
+ --lr_scheduler.decay_type $DECAY_TYPE \
+ --lr_scheduler.decay_ratio $DECAY_RATIO \
+ --training.batch_size $BATCH_SIZE \
+ --training.seq_len $SEQ_LEN \
+ --training.context_len $SEQ_LEN \
+ --training.gradient_accumulation_steps $GAS \
+ --training.steps $STEPS \
+ --training.max_norm 1.0 \
+ --training.skip_nan_inf \
+ --training.dataset $DATASET_ROOT/fineweb-edu-sample,$DATASET_ROOT/small_repos_20B_sample_merged,$DATASET_ROOT/megamath-web-pro \
+ --training.data_probs 0.55,0.3,0.15 \
+ --training.dataset_split train,train,train \
+ --training.dataset_name default,default,default \
+ --training.streaming \
+ --training.num_workers 32 \
+ --training.prefetch_factor 2 \
+ --training.seed 42 \
+ --training.compile \
+ --checkpoint.interval 8192 \
+ --checkpoint.load_step -1 \
+ --checkpoint.keep_latest_k 100 \
+ --metrics.log_freq 1 \
+ --metrics.enable_tensorboard \
+ --training.streaming
+
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/config.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..e7ffa1364dca6d44752e60c47c942e30226c60c2
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/config.json
@@ -0,0 +1,57 @@
+{
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "torch_dtype": "float32",
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/delta_net_1B.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/delta_net_1B.json
new file mode 100644
index 0000000000000000000000000000000000000000..97726ff8c42fe0c32cad027fa3bd419e195bc68a
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/delta_net_1B.json
@@ -0,0 +1,29 @@
+{
+ "attn": null,
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "conv_size": 4,
+ "eos_token_id": 2,
+ "expand_k": 1,
+ "expand_v": 1,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "hidden_act": "swish",
+ "hidden_ratio": 4,
+ "hidden_size": 2048,
+ "initializer_range": 0.02,
+ "intermediate_size": null,
+ "model_type": "delta_net",
+ "norm_eps": 1e-06,
+ "num_heads": 16,
+ "num_hidden_layers": 24,
+ "pad_token_id": 2,
+ "qk_activation": "silu",
+ "qk_norm": "l2",
+ "tie_word_embeddings": false,
+ "use_beta": true,
+ "use_cache": true,
+ "use_gate": false,
+ "use_output_norm": true,
+ "use_short_conv": true
+}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/delta_net_340M.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/delta_net_340M.json
new file mode 100644
index 0000000000000000000000000000000000000000..2eb566c873b179893174edfd816706a22cf4d53a
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/delta_net_340M.json
@@ -0,0 +1,26 @@
+{
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "conv_size": 4,
+ "eos_token_id": 2,
+ "expand_k": 1,
+ "expand_v": 1,
+ "fuse_cross_entropy": true,
+ "hidden_act": "swish",
+ "hidden_ratio": 4,
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "intermediate_size": null,
+ "model_type": "delta_net",
+ "norm_eps": 1e-06,
+ "num_heads": 8,
+ "num_hidden_layers": 24,
+ "qk_activation": "silu",
+ "qk_norm": "l2",
+ "tie_word_embeddings": false,
+ "use_beta": true,
+ "use_cache": true,
+ "use_gate": false,
+ "use_output_norm": true,
+ "use_short_conv": true
+}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/gated_deltanet_1B.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/gated_deltanet_1B.json
new file mode 100644
index 0000000000000000000000000000000000000000..f10793ea778aa039138a8821c2a6eda1c3f76cef
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/gated_deltanet_1B.json
@@ -0,0 +1,22 @@
+{
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "conv_size": 4,
+ "eos_token_id": 2,
+ "expand_v": 2,
+ "fuse_cross_entropy": true,
+ "head_dim": 256,
+ "hidden_act": "swish",
+ "hidden_ratio": 4,
+ "hidden_size": 2048,
+ "initializer_range": 0.02,
+ "intermediate_size": null,
+ "model_type": "gated_deltanet",
+ "norm_eps": 1e-06,
+ "num_heads": 6,
+ "num_hidden_layers": 21,
+ "tie_word_embeddings": false,
+ "use_cache": true,
+ "use_gate": true,
+ "use_short_conv": true
+}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/gated_deltanet_340M.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/gated_deltanet_340M.json
new file mode 100644
index 0000000000000000000000000000000000000000..0c592213c2013465f36b394eaae49f54f484936a
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/gated_deltanet_340M.json
@@ -0,0 +1,22 @@
+{
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "conv_size": 4,
+ "eos_token_id": 2,
+ "expand_v": 2,
+ "fuse_cross_entropy": true,
+ "head_dim": 256,
+ "hidden_act": "swish",
+ "hidden_ratio": 4,
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "intermediate_size": null,
+ "model_type": "gated_deltanet",
+ "norm_eps": 1e-06,
+ "num_heads": 6,
+ "num_hidden_layers": 21,
+ "tie_word_embeddings": false,
+ "use_cache": true,
+ "use_gate": true,
+ "use_short_conv": true
+}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/gdn_6_1_340M.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/gdn_6_1_340M.json
new file mode 100644
index 0000000000000000000000000000000000000000..8b17fd826dd82b406a0d575a42edb5e8efbe4bf7
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/gdn_6_1_340M.json
@@ -0,0 +1,50 @@
+{
+ "architectures": [
+ "GatedDeltaNetForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "conv_size": 4,
+ "eos_token_id": 2,
+ "expand_k": 1,
+ "expand_v": 1,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 256,
+ "hidden_act": "swish",
+ "hidden_ratio": 4,
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "intermediate_size": null,
+ "max_position_embeddings": 8192,
+ "model_type": "gated_deltanet",
+ "norm_eps": 1e-06,
+ "norm_first": false,
+ "num_heads": 4,
+ "num_hidden_layers": 24,
+ "qk_activation": "silu",
+ "qk_norm": "l2",
+ "tie_word_embeddings": false,
+ "torch_dtype": "float32",
+ "transformers_version": "4.51.3",
+ "use_beta": true,
+ "use_cache": true,
+ "use_gate": true,
+ "use_output_norm": true,
+ "use_short_conv": true,
+ "vocab_size": 32000
+}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/gla_340M.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/gla_340M.json
new file mode 100644
index 0000000000000000000000000000000000000000..7190f4fd6b1abb5ee0af83f108bd4100cc6cd038
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/gla_340M.json
@@ -0,0 +1,24 @@
+{
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "clamp_min": null,
+ "eos_token_id": 2,
+ "expand_k": 0.5,
+ "expand_v": 1,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "hidden_act": "swish",
+ "hidden_ratio": 4,
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "intermediate_size": null,
+ "model_type": "gla",
+ "num_heads": 4,
+ "num_hidden_layers": 24,
+ "norm_eps": 1e-06,
+ "tie_word_embeddings": false,
+ "use_cache": true,
+ "use_gk": true,
+ "use_gv": false,
+ "vocab_size": 32000
+}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/gla_7B.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/gla_7B.json
new file mode 100644
index 0000000000000000000000000000000000000000..16d373247ce3c3f96a804f8e41014e162c17b9d8
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/gla_7B.json
@@ -0,0 +1,25 @@
+{
+ "attn": null,
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "eos_token_id": 2,
+ "expand_k": 0.5,
+ "expand_v": 1,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "hidden_act": "swish",
+ "hidden_ratio": 4,
+ "hidden_size": 4096,
+ "initializer_range": 0.02,
+ "intermediate_size": 11008,
+ "model_type": "gla",
+ "norm_eps": 1e-06,
+ "num_heads": 16,
+ "num_hidden_layers": 32,
+ "tie_word_embeddings": false,
+ "use_cache": true,
+ "use_gk": true,
+ "use_gv": false,
+ "use_output_gate": true,
+ "use_short_conv": false
+}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/gsa_340M.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/gsa_340M.json
new file mode 100644
index 0000000000000000000000000000000000000000..8b4f4ec77ac1b3978dac071e91e26ced4ae58342
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/gsa_340M.json
@@ -0,0 +1,29 @@
+{
+ "bos_token_id": 1,
+ "conv_size": 4,
+ "eos_token_id": 2,
+ "expand_k": 1,
+ "expand_v": 1,
+ "elementwise_affine": false,
+ "feature_map": "swish",
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "gate_logit_normalizer": 4,
+ "hidden_act": "swish",
+ "hidden_ratio": 4,
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "intermediate_size": null,
+ "model_type": "gsa",
+ "num_heads": 4,
+ "num_hidden_layers": 24,
+ "num_slots": 64,
+ "norm_eps": 1e-06,
+ "share_conv_kernel": true,
+ "tie_word_embeddings": false,
+ "use_cache": true,
+ "use_norm": true,
+ "use_output_gate": true,
+ "use_rope": false,
+ "use_short_conv": false
+}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/hgrn2_340M.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/hgrn2_340M.json
new file mode 100644
index 0000000000000000000000000000000000000000..9f8ab2e06998cb0c2c42bee50e6a0b43833cb440
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/hgrn2_340M.json
@@ -0,0 +1,20 @@
+{
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "eos_token_id": 2,
+ "expand_ratio": 128,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "hidden_act": "swish",
+ "hidden_ratio": 4,
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "intermediate_size": null,
+ "model_type": "hgrn2",
+ "num_heads": 8,
+ "num_hidden_layers": 24,
+ "norm_eps": 1e-06,
+ "tie_word_embeddings": false,
+ "use_cache": true,
+ "vocab_size": 32000
+}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/mamba2_1B.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/mamba2_1B.json
new file mode 100644
index 0000000000000000000000000000000000000000..7469f2819a2bc60796e9a76d5e1ead16f4039d68
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/mamba2_1B.json
@@ -0,0 +1,32 @@
+{
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 2048,
+ "initializer_range": 0.02,
+ "norm_eps": 1e-05,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.50.1",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "vocab_size": 32000
+}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/mamba2_340M.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/mamba2_340M.json
new file mode 100644
index 0000000000000000000000000000000000000000..0f9c34d4ef6280cb1b3f813f7bac9de907a7ebed
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/mamba2_340M.json
@@ -0,0 +1,32 @@
+{
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "norm_eps": 1e-05,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.50.1",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "vocab_size": 32000
+}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/mamba2_6_1_340M.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/mamba2_6_1_340M.json
new file mode 100644
index 0000000000000000000000000000000000000000..34203fa36cc4c6cdbcc724497e486b19a922d531
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/mamba2_6_1_340M.json
@@ -0,0 +1,50 @@
+{
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "norm_eps": 1e-05,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.50.1",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "vocab_size": 32000
+}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/mamba_1B.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/mamba_1B.json
new file mode 100644
index 0000000000000000000000000000000000000000..8ab111e041711a082aef0e9cac5ca750a2f59c91
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/mamba_1B.json
@@ -0,0 +1,30 @@
+{
+ "bos_token_id": 1,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "hidden_act": "silu",
+ "hidden_size": 2048,
+ "initializer_range": 0.02,
+ "model_type": "mamba",
+ "norm_eps": 1e-05,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": false,
+ "residual_in_fp32": false,
+ "state_size": 16,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_init_scheme": "random",
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "time_step_scale": 1.0,
+ "transformers_version": "4.50.1",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "vocab_size": 32000
+}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/mamba_340M.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/mamba_340M.json
new file mode 100644
index 0000000000000000000000000000000000000000..a316b792619d6a871a3f561f477fc2ab044cf5bd
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/mamba_340M.json
@@ -0,0 +1,30 @@
+{
+ "bos_token_id": 1,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba",
+ "norm_eps": 1e-05,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": false,
+ "residual_in_fp32": false,
+ "state_size": 16,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_init_scheme": "random",
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "time_step_scale": 1.0,
+ "transformers_version": "4.50.1",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "vocab_size": 32000
+}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/samba_1B.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/samba_1B.json
new file mode 100644
index 0000000000000000000000000000000000000000..58f91099417bf59574da7188624fe51c49bfe441
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/samba_1B.json
@@ -0,0 +1,52 @@
+{
+ "attn": {
+ "layers": [
+ 1,
+ 3,
+ 5,
+ 7,
+ 9,
+ 11,
+ 13,
+ 15,
+ 17
+ ],
+ "num_heads": 18,
+ "num_kv_heads": 18,
+ "qkv_bias": false,
+ "rope_theta": 10000.0,
+ "window_size": 2048
+ },
+ "bos_token_id": 1,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "hidden_act": "swish",
+ "hidden_ratio": 4,
+ "hidden_size": 2304,
+ "initializer_range": 0.02,
+ "intermediate_size": 4608,
+ "max_position_embeddings": 2048,
+ "model_type": "samba",
+ "norm_eps": 1e-05,
+ "num_hidden_layers": 18,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": false,
+ "residual_in_fp32": false,
+ "state_size": 16,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_init_scheme": "random",
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 144,
+ "time_step_scale": 1.0,
+ "transformers_version": "4.50.1",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "vocab_size": 32000
+}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/sba_340m.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/sba_340m.json
new file mode 100644
index 0000000000000000000000000000000000000000..97c98de03c9c1f8ea8062d1f30147856cd036df6
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/sba_340m.json
@@ -0,0 +1,18 @@
+{
+ "attention_bias": false,
+ "bos_token_id": 1,
+ "eos_token_id": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "hidden_act": "swish",
+ "hidden_size": 1024,
+ "initializer_range": 0.006,
+ "max_position_embeddings": 8192,
+ "model_type": "sba",
+ "num_heads": 16,
+ "num_hidden_layers": 24,
+ "norm_eps": 1e-06,
+ "tie_word_embeddings": false,
+ "use_cache": true,
+ "vocab_size": 32000
+}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/transformer_1B.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/transformer_1B.json
new file mode 100644
index 0000000000000000000000000000000000000000..042dd38334d708c6d0e21aba0e791f0e995af605
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/transformer_1B.json
@@ -0,0 +1,22 @@
+{
+ "bos_token_id": 1,
+ "elementwise_affine": true,
+ "eos_token_id": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "hidden_act": "swish",
+ "hidden_ratio": 4,
+ "hidden_size": 2048,
+ "initializer_range": 0.02,
+ "intermediate_size": null,
+ "max_position_embeddings": 8192,
+ "model_type": "transformer",
+ "norm_eps": 1e-06,
+ "num_heads": 32,
+ "num_hidden_layers": 24,
+ "num_kv_heads": null,
+ "pad_token_id": 2,
+ "rope_theta": 10000.0,
+ "tie_word_embeddings": false
+}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/transformer_340M.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/transformer_340M.json
new file mode 100644
index 0000000000000000000000000000000000000000..696bc5cbeee41ec89f6c1cb06399b2189b627a93
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/transformer_340M.json
@@ -0,0 +1,18 @@
+{
+ "attention_bias": false,
+ "bos_token_id": 1,
+ "eos_token_id": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "hidden_act": "swish",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "max_position_embeddings": 8192,
+ "model_type": "transformer",
+ "num_heads": 16,
+ "num_hidden_layers": 24,
+ "norm_eps": 1e-06,
+ "tie_word_embeddings": false,
+ "use_cache": true,
+ "vocab_size": 32000
+}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/transformer_7B.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/transformer_7B.json
new file mode 100644
index 0000000000000000000000000000000000000000..a2a3c2c81d0aecf83e4ccdd97851192f0e16939e
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/configs/transformer_7B.json
@@ -0,0 +1,21 @@
+{
+ "attention_bias": false,
+ "bos_token_id": 1,
+ "eos_token_id": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "hidden_act": "swish",
+ "hidden_ratio": 4,
+ "hidden_size": 4096,
+ "initializer_range": 0.02,
+ "intermediate_size": 14336,
+ "model_type": "transformer",
+ "norm_eps": 1e-06,
+ "num_heads": 32,
+ "num_hidden_layers": 32,
+ "num_kv_heads": 8,
+ "rope_theta": 10000.0,
+ "tie_word_embeddings": false,
+ "use_cache": true,
+ "window_size": null
+}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/generation_config.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..00505bc1d8228a6333e98b247832d82817229e8c
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/generation_config.json
@@ -0,0 +1,7 @@
+{
+ "_from_model_config": true,
+ "bos_token_id": 1,
+ "eos_token_id": 2,
+ "pad_token_id": 0,
+ "transformers_version": "4.53.3"
+}
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/0/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/0/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..766fbfe21987830bd5e5273294b4586c37ddd15d
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/0/error.json
@@ -0,0 +1 @@
+{"message": {"message": "OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 0 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003896 has 316.00 MiB memory in use. Process 696027 has 316.00 MiB memory in use. Process 1114693 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 601, in forward\n return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 528, in torch_forward\n G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]\n ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~\ntorch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 0 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003896 has 316.00 MiB memory in use. Process 696027 has 316.00 MiB memory in use. Process 1114693 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)\n", "timestamp": "1753252283"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/0/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/0/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..b9b1be51a8510012cb8fe08cd79cdb9984f0b67e
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/0/stderr.log
@@ -0,0 +1,463 @@
+[titan] 2025-07-23 14:27:46,756 - root - INFO - Starting job: default job
+[titan] 2025-07-23 14:27:46,756 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 14:27:46,756 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 14:27:46,757 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 14:27:46,772 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 14:27:46,951 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 14:27:46,951 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 14:27:46,951 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 14:27:47,501 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 14:27:47,997 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 14:27:47,998 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:47,998 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:48,644 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 14:27:48,644 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:27:48,645 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:48,645 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:39,750 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:39,881 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 14:28:39,881 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:28:39,882 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:39,882 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,150 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,316 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 14:28:40,316 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:28:40,316 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,316 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:46,507 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 14:28:47,196 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 14:28:47,310 - root - INFO - Building dataloader...
+[titan] 2025-07-23 14:28:47,312 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 14:28:47,314 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 14:28:50,147 - fla.layers.mamba2 - WARNING - The fast path is not available because one of `(selective_state_update)` is None. Falling back to the naive implementation. To install follow https://github.com/state-spaces/mamba/#installation
+[titan] 2025-07-23 14:28:50,147 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 14:28:50,265 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 14:28:50,317 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 14:28:50,318 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 14:28:50,318 - root - WARNING - No norm found in model
+[titan] 2025-07-23 14:28:50,318 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 14:28:50,540 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 14:28:50,884 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 14:28:51,042 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 14:28:51,272 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 14:28:51,273 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 14:28:51,297 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+wandb: Network error (InvalidURL), entering retry loop.
+wandb: W&B API key is configured. Use `wandb login --relogin` to force relogin
+wandb: Network error (InvalidURL), entering retry loop.
+[titan] 2025-07-23 14:30:44,436 - root - ERROR - Failed to create WandB logger: Run initialization has timed out after 90.0 sec. Please try increasing the timeout with the `init_timeout` setting: `wandb.init(settings=wandb.Settings(init_timeout=120))`.
+[titan] 2025-07-23 14:30:44,442 - root - INFO - TensorBoard logging enabled. Logs will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1428
+[titan] 2025-07-23 14:30:44,442 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 14:30:44,527 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 14:30:50,781 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 14:30:50,784 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 14:30:50,784 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 14:30:50,784 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 14:30:50,785 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 14:30:50,785 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 14:30:50,785 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 14:30:50,785 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 14:30:50,785 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 14:30:50,785 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+Traceback (most recent call last):
+ File "", line 198, in _run_module_as_main
+ File "", line 88, in _run_code
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+ main(config)
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+ return f(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+ output = model(
+ ^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+ return self._call_impl(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+ return inner()
+ ^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+ result = forward_call(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+ return func(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+ outputs = self.backbone(
+ ^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+ return self._call_impl(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+ return forward_call(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+ hidden_states = mixer_block(
+ ^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+ return self._call_impl(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+ return inner()
+ ^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+ result = forward_call(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+ return fn(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+ return self._call_impl(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+ return forward_call(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+ hidden_states = self.norm(hidden_states)
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+ hidden_states = self.mixer(
+ ^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+ return self._call_impl(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+ return forward_call(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 601, in forward
+ return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 528, in torch_forward
+ G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]
+ ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~
+torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 0 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003896 has 316.00 MiB memory in use. Process 696027 has 316.00 MiB memory in use. Process 1114693 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
+[rank0]: Traceback (most recent call last):
+[rank0]: File "", line 198, in _run_module_as_main
+[rank0]: File "", line 88, in _run_code
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank0]: main(config)
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank0]: return f(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank0]: output = model(
+[rank0]: ^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank0]: return self._call_impl(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank0]: return inner()
+[rank0]: ^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank0]: result = forward_call(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank0]: return func(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank0]: outputs = self.backbone(
+[rank0]: ^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank0]: return self._call_impl(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank0]: return forward_call(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank0]: hidden_states = mixer_block(
+[rank0]: ^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank0]: return self._call_impl(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank0]: return inner()
+[rank0]: ^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank0]: result = forward_call(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank0]: return fn(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank0]: return self._call_impl(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank0]: return forward_call(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank0]: hidden_states = self.norm(hidden_states)
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank0]: hidden_states = self.mixer(
+[rank0]: ^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank0]: return self._call_impl(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank0]: return forward_call(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 601, in forward
+[rank0]: return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 528, in torch_forward
+[rank0]: G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]
+[rank0]: ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~
+[rank0]: torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 0 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003896 has 316.00 MiB memory in use. Process 696027 has 316.00 MiB memory in use. Process 1114693 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/0/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/0/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/1/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/1/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..a256487d56600171cf254aaee97c9eb1f5503417
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/1/error.json
@@ -0,0 +1 @@
+{"message": {"message": "OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 1 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003901 has 316.00 MiB memory in use. Process 696029 has 316.00 MiB memory in use. Process 1114694 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 601, in forward\n return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 528, in torch_forward\n G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]\n ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~\ntorch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 1 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003901 has 316.00 MiB memory in use. Process 696029 has 316.00 MiB memory in use. Process 1114694 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)\n", "timestamp": "1753252283"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/1/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/1/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..d2a6a202089712e1931006935465601d3d3adc18
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/1/stderr.log
@@ -0,0 +1,387 @@
+[titan] 2025-07-23 14:27:46,323 - root - INFO - Starting job: default job
+[titan] 2025-07-23 14:27:46,323 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 14:27:46,324 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 14:27:47,255 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 14:27:47,258 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 14:27:47,324 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 14:27:47,324 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 14:27:47,324 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 14:27:47,411 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 14:27:47,997 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 14:27:47,998 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:47,998 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:48,493 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 14:27:48,493 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:27:48,493 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:48,493 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:41,064 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:41,096 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 14:28:41,096 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:28:41,096 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:41,097 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:41,357 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:41,441 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 14:28:41,441 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:28:41,441 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:41,441 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:47,757 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 14:28:48,445 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 14:28:48,560 - root - INFO - Building dataloader...
+[titan] 2025-07-23 14:28:48,562 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 14:28:48,564 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 14:28:50,147 - fla.layers.mamba2 - WARNING - The fast path is not available because one of `(selective_state_update)` is None. Falling back to the naive implementation. To install follow https://github.com/state-spaces/mamba/#installation
+[titan] 2025-07-23 14:28:50,147 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 14:28:50,264 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 14:28:50,316 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 14:28:50,316 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 14:28:50,317 - root - WARNING - No norm found in model
+[titan] 2025-07-23 14:28:50,317 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 14:28:50,541 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 14:28:50,886 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 14:28:51,042 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 14:28:51,271 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 14:28:51,273 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 14:28:51,297 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 14:28:51,302 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 14:28:51,429 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 14:28:58,662 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 14:28:58,667 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 14:28:58,670 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 14:28:58,670 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 14:28:58,670 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 14:28:58,670 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 14:28:58,670 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 14:28:58,670 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 14:28:58,671 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 14:28:58,671 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank1]: Traceback (most recent call last):
+[rank1]: File "", line 198, in _run_module_as_main
+[rank1]: File "", line 88, in _run_code
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank1]: main(config)
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank1]: return f(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank1]: output = model(
+[rank1]: ^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank1]: return self._call_impl(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank1]: return inner()
+[rank1]: ^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank1]: result = forward_call(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank1]: return func(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank1]: outputs = self.backbone(
+[rank1]: ^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank1]: return self._call_impl(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank1]: return forward_call(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank1]: hidden_states = mixer_block(
+[rank1]: ^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank1]: return self._call_impl(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank1]: return inner()
+[rank1]: ^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank1]: result = forward_call(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank1]: return fn(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank1]: return self._call_impl(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank1]: return forward_call(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank1]: hidden_states = self.norm(hidden_states)
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank1]: hidden_states = self.mixer(
+[rank1]: ^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank1]: return self._call_impl(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank1]: return forward_call(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 601, in forward
+[rank1]: return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 528, in torch_forward
+[rank1]: G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]
+[rank1]: ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~
+[rank1]: torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 1 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003901 has 316.00 MiB memory in use. Process 696029 has 316.00 MiB memory in use. Process 1114694 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/1/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/1/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/2/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/2/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..cd0a72452a8e65a4195597949d114c785c14fef1
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/2/error.json
@@ -0,0 +1 @@
+{"message": {"message": "OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 2 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003900 has 316.00 MiB memory in use. Process 696030 has 316.00 MiB memory in use. Process 1114695 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 601, in forward\n return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 528, in torch_forward\n G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]\n ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~\ntorch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 2 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003900 has 316.00 MiB memory in use. Process 696030 has 316.00 MiB memory in use. Process 1114695 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)\n", "timestamp": "1753252283"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/2/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/2/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..993e4a166978641eeb4e9dbc866f7ed74c1cf0f6
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/2/stderr.log
@@ -0,0 +1,387 @@
+[titan] 2025-07-23 14:27:46,385 - root - INFO - Starting job: default job
+[titan] 2025-07-23 14:27:46,386 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 14:27:46,386 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 14:27:47,325 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 14:27:47,327 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 14:27:47,375 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 14:27:47,375 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 14:27:47,376 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 14:27:47,418 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 14:27:47,997 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 14:27:47,998 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:47,998 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:48,543 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 14:27:48,543 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:27:48,544 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:48,544 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:39,999 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,032 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 14:28:40,032 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:28:40,032 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,032 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,289 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,382 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 14:28:40,382 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:28:40,382 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,382 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:46,791 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 14:28:47,494 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 14:28:47,612 - root - INFO - Building dataloader...
+[titan] 2025-07-23 14:28:47,614 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 14:28:47,616 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 14:28:50,147 - fla.layers.mamba2 - WARNING - The fast path is not available because one of `(selective_state_update)` is None. Falling back to the naive implementation. To install follow https://github.com/state-spaces/mamba/#installation
+[titan] 2025-07-23 14:28:50,148 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 14:28:50,264 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 14:28:50,315 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 14:28:50,316 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 14:28:50,316 - root - WARNING - No norm found in model
+[titan] 2025-07-23 14:28:50,316 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 14:28:50,539 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 14:28:50,882 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 14:28:51,042 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 14:28:51,274 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 14:28:51,275 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 14:28:51,299 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 14:28:51,307 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 14:28:51,458 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 14:28:58,658 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 14:28:58,660 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 14:28:58,664 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 14:28:58,664 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 14:28:58,665 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 14:28:58,665 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 14:28:58,665 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 14:28:58,665 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 14:28:58,666 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 14:28:58,666 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank2]: Traceback (most recent call last):
+[rank2]: File "", line 198, in _run_module_as_main
+[rank2]: File "", line 88, in _run_code
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank2]: main(config)
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank2]: return f(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank2]: output = model(
+[rank2]: ^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank2]: return self._call_impl(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank2]: return inner()
+[rank2]: ^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank2]: result = forward_call(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank2]: return func(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank2]: outputs = self.backbone(
+[rank2]: ^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank2]: return self._call_impl(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank2]: return forward_call(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank2]: hidden_states = mixer_block(
+[rank2]: ^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank2]: return self._call_impl(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank2]: return inner()
+[rank2]: ^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank2]: result = forward_call(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank2]: return fn(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank2]: return self._call_impl(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank2]: return forward_call(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank2]: hidden_states = self.norm(hidden_states)
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank2]: hidden_states = self.mixer(
+[rank2]: ^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank2]: return self._call_impl(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank2]: return forward_call(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 601, in forward
+[rank2]: return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 528, in torch_forward
+[rank2]: G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]
+[rank2]: ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~
+[rank2]: torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 2 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003900 has 316.00 MiB memory in use. Process 696030 has 316.00 MiB memory in use. Process 1114695 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/2/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/2/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/3/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/3/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..508be882dbee45d2e901769069b6f851878a1a11
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/3/error.json
@@ -0,0 +1 @@
+{"message": {"message": "OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 3 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003902 has 316.00 MiB memory in use. Process 696032 has 316.00 MiB memory in use. Process 1114696 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 601, in forward\n return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 528, in torch_forward\n G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]\n ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~\ntorch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 3 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003902 has 316.00 MiB memory in use. Process 696032 has 316.00 MiB memory in use. Process 1114696 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)\n", "timestamp": "1753252283"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/3/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/3/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..0c056c5fcbf3aecc691b23a6f3262803ee1c792d
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/3/stderr.log
@@ -0,0 +1,387 @@
+[titan] 2025-07-23 14:27:46,466 - root - INFO - Starting job: default job
+[titan] 2025-07-23 14:27:46,466 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 14:27:46,466 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 14:27:47,421 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 14:27:47,423 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 14:27:47,484 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 14:27:47,485 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 14:27:47,485 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 14:27:47,493 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 14:27:47,997 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 14:27:47,998 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:47,999 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:48,494 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 14:27:48,494 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:27:48,494 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:48,494 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:39,997 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,028 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 14:28:40,029 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:28:40,029 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,029 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,298 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,391 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 14:28:40,392 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:28:40,392 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,392 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:46,760 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 14:28:47,557 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 14:28:47,677 - root - INFO - Building dataloader...
+[titan] 2025-07-23 14:28:47,680 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 14:28:47,681 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 14:28:50,147 - fla.layers.mamba2 - WARNING - The fast path is not available because one of `(selective_state_update)` is None. Falling back to the naive implementation. To install follow https://github.com/state-spaces/mamba/#installation
+[titan] 2025-07-23 14:28:50,148 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 14:28:50,265 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 14:28:50,322 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 14:28:50,322 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 14:28:50,322 - root - WARNING - No norm found in model
+[titan] 2025-07-23 14:28:50,323 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 14:28:50,540 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 14:28:50,882 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 14:28:51,042 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 14:28:51,272 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 14:28:51,273 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 14:28:51,297 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 14:28:51,324 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 14:28:51,477 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 14:28:58,659 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 14:28:58,661 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 14:28:58,663 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 14:28:58,663 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 14:28:58,664 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 14:28:58,664 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 14:28:58,664 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 14:28:58,664 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 14:28:58,664 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 14:28:58,665 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank3]: Traceback (most recent call last):
+[rank3]: File "", line 198, in _run_module_as_main
+[rank3]: File "", line 88, in _run_code
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank3]: main(config)
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank3]: return f(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank3]: output = model(
+[rank3]: ^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank3]: return self._call_impl(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank3]: return inner()
+[rank3]: ^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank3]: result = forward_call(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank3]: return func(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank3]: outputs = self.backbone(
+[rank3]: ^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank3]: return self._call_impl(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank3]: return forward_call(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank3]: hidden_states = mixer_block(
+[rank3]: ^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank3]: return self._call_impl(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank3]: return inner()
+[rank3]: ^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank3]: result = forward_call(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank3]: return fn(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank3]: return self._call_impl(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank3]: return forward_call(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank3]: hidden_states = self.norm(hidden_states)
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank3]: hidden_states = self.mixer(
+[rank3]: ^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank3]: return self._call_impl(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank3]: return forward_call(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 601, in forward
+[rank3]: return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 528, in torch_forward
+[rank3]: G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]
+[rank3]: ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~
+[rank3]: torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 3 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003902 has 316.00 MiB memory in use. Process 696032 has 316.00 MiB memory in use. Process 1114696 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/3/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/3/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/4/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/4/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..6fd081ba076e9a2cdcb4c6f67200d323f2f028c2
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/4/error.json
@@ -0,0 +1 @@
+{"message": {"message": "OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 4 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003903 has 316.00 MiB memory in use. Process 696034 has 316.00 MiB memory in use. Process 1114697 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 601, in forward\n return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 528, in torch_forward\n G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]\n ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~\ntorch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 4 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003903 has 316.00 MiB memory in use. Process 696034 has 316.00 MiB memory in use. Process 1114697 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)\n", "timestamp": "1753252283"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/4/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/4/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..9ca3d497a1f0aeeb038ef0e679c76e7c1eb5ec3c
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/4/stderr.log
@@ -0,0 +1,387 @@
+[titan] 2025-07-23 14:27:46,452 - root - INFO - Starting job: default job
+[titan] 2025-07-23 14:27:46,453 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 14:27:46,453 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 14:27:47,428 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 14:27:47,431 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 14:27:47,487 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 14:27:47,487 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 14:27:47,487 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 14:27:47,494 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 14:27:47,997 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 14:27:47,998 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:47,998 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:48,492 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 14:27:48,492 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:27:48,492 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:48,492 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:39,720 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:39,830 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 14:28:39,831 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:28:39,831 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:39,831 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,087 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,312 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 14:28:40,313 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:28:40,313 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,313 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:46,672 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 14:28:47,386 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 14:28:47,512 - root - INFO - Building dataloader...
+[titan] 2025-07-23 14:28:47,515 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 14:28:47,516 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 14:28:50,147 - fla.layers.mamba2 - WARNING - The fast path is not available because one of `(selective_state_update)` is None. Falling back to the naive implementation. To install follow https://github.com/state-spaces/mamba/#installation
+[titan] 2025-07-23 14:28:50,147 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 14:28:50,265 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 14:28:50,317 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 14:28:50,317 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 14:28:50,317 - root - WARNING - No norm found in model
+[titan] 2025-07-23 14:28:50,317 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 14:28:50,541 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 14:28:50,886 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 14:28:51,042 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 14:28:51,272 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 14:28:51,274 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 14:28:51,298 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 14:28:51,315 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 14:28:51,473 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 14:28:58,659 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 14:28:58,667 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 14:28:58,668 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 14:28:58,668 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 14:28:58,668 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 14:28:58,668 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 14:28:58,668 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 14:28:58,668 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 14:28:58,668 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 14:28:58,669 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank4]: Traceback (most recent call last):
+[rank4]: File "", line 198, in _run_module_as_main
+[rank4]: File "", line 88, in _run_code
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank4]: main(config)
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank4]: return f(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank4]: output = model(
+[rank4]: ^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank4]: return self._call_impl(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank4]: return inner()
+[rank4]: ^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank4]: result = forward_call(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank4]: return func(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank4]: outputs = self.backbone(
+[rank4]: ^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank4]: return self._call_impl(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank4]: return forward_call(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank4]: hidden_states = mixer_block(
+[rank4]: ^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank4]: return self._call_impl(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank4]: return inner()
+[rank4]: ^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank4]: result = forward_call(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank4]: return fn(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank4]: return self._call_impl(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank4]: return forward_call(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank4]: hidden_states = self.norm(hidden_states)
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank4]: hidden_states = self.mixer(
+[rank4]: ^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank4]: return self._call_impl(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank4]: return forward_call(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 601, in forward
+[rank4]: return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 528, in torch_forward
+[rank4]: G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]
+[rank4]: ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~
+[rank4]: torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 4 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003903 has 316.00 MiB memory in use. Process 696034 has 316.00 MiB memory in use. Process 1114697 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/4/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/4/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/5/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/5/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..52b26c714d5c3cfab63093b2b31a2f7cd8bba1cc
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/5/error.json
@@ -0,0 +1 @@
+{"message": {"message": "OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 5 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003905 has 316.00 MiB memory in use. Process 696036 has 316.00 MiB memory in use. Process 1114698 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 601, in forward\n return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 528, in torch_forward\n G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]\n ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~\ntorch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 5 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003905 has 316.00 MiB memory in use. Process 696036 has 316.00 MiB memory in use. Process 1114698 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)\n", "timestamp": "1753252283"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/5/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/5/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..813275ca7fdfcbd62b1c3a28c54714fd36a0fc52
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/5/stderr.log
@@ -0,0 +1,387 @@
+[titan] 2025-07-23 14:27:46,350 - root - INFO - Starting job: default job
+[titan] 2025-07-23 14:27:46,350 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 14:27:46,350 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 14:27:47,258 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 14:27:47,260 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 14:27:47,324 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 14:27:47,324 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 14:27:47,324 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 14:27:47,411 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 14:27:47,997 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 14:27:47,999 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:47,999 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:48,496 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 14:27:48,496 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:27:48,496 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:48,496 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:39,812 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:39,845 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 14:28:39,845 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:28:39,845 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:39,845 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,105 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,312 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 14:28:40,312 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:28:40,312 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,312 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:46,706 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 14:28:47,415 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 14:28:47,539 - root - INFO - Building dataloader...
+[titan] 2025-07-23 14:28:47,541 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 14:28:47,543 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 14:28:50,147 - fla.layers.mamba2 - WARNING - The fast path is not available because one of `(selective_state_update)` is None. Falling back to the naive implementation. To install follow https://github.com/state-spaces/mamba/#installation
+[titan] 2025-07-23 14:28:50,147 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 14:28:50,264 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 14:28:50,315 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 14:28:50,316 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 14:28:50,316 - root - WARNING - No norm found in model
+[titan] 2025-07-23 14:28:50,316 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 14:28:50,539 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 14:28:50,884 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 14:28:51,042 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 14:28:51,273 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 14:28:51,275 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 14:28:51,299 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 14:28:51,333 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 14:28:51,479 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 14:28:59,100 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 14:28:59,101 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 14:28:59,153 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 14:28:59,170 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 14:28:59,171 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 14:28:59,171 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 14:28:59,172 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 14:28:59,172 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 14:28:59,172 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 14:28:59,173 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank5]: Traceback (most recent call last):
+[rank5]: File "", line 198, in _run_module_as_main
+[rank5]: File "", line 88, in _run_code
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank5]: main(config)
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank5]: return f(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank5]: output = model(
+[rank5]: ^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank5]: return self._call_impl(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank5]: return inner()
+[rank5]: ^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank5]: result = forward_call(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank5]: return func(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank5]: outputs = self.backbone(
+[rank5]: ^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank5]: return self._call_impl(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank5]: return forward_call(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank5]: hidden_states = mixer_block(
+[rank5]: ^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank5]: return self._call_impl(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank5]: return inner()
+[rank5]: ^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank5]: result = forward_call(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank5]: return fn(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank5]: return self._call_impl(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank5]: return forward_call(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank5]: hidden_states = self.norm(hidden_states)
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank5]: hidden_states = self.mixer(
+[rank5]: ^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank5]: return self._call_impl(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank5]: return forward_call(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 601, in forward
+[rank5]: return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 528, in torch_forward
+[rank5]: G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]
+[rank5]: ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~
+[rank5]: torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 5 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003905 has 316.00 MiB memory in use. Process 696036 has 316.00 MiB memory in use. Process 1114698 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/5/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/5/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/6/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/6/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..4d3aaca590a877a793ddb8ef6617368826d17345
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/6/error.json
@@ -0,0 +1 @@
+{"message": {"message": "OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 6 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003904 has 316.00 MiB memory in use. Process 696035 has 316.00 MiB memory in use. Process 1114699 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 601, in forward\n return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 528, in torch_forward\n G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]\n ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~\ntorch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 6 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003904 has 316.00 MiB memory in use. Process 696035 has 316.00 MiB memory in use. Process 1114699 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)\n", "timestamp": "1753252283"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/6/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/6/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..6f31aeca60cf9436d6ca4d4c33ecf1416943ca58
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/6/stderr.log
@@ -0,0 +1,387 @@
+[titan] 2025-07-23 14:27:46,733 - root - INFO - Starting job: default job
+[titan] 2025-07-23 14:27:46,734 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 14:27:46,734 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 14:27:47,466 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 14:27:47,468 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 14:27:47,519 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 14:27:47,519 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 14:27:47,519 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 14:27:47,528 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 14:27:47,997 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 14:27:47,998 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:47,998 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:48,495 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 14:27:48,495 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:27:48,495 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:48,495 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:39,968 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,002 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 14:28:40,002 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:28:40,002 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,002 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,261 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,356 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 14:28:40,357 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:28:40,357 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,357 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:46,633 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 14:28:47,337 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 14:28:47,461 - root - INFO - Building dataloader...
+[titan] 2025-07-23 14:28:47,463 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 14:28:47,465 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 14:28:50,147 - fla.layers.mamba2 - WARNING - The fast path is not available because one of `(selective_state_update)` is None. Falling back to the naive implementation. To install follow https://github.com/state-spaces/mamba/#installation
+[titan] 2025-07-23 14:28:50,147 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 14:28:50,263 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 14:28:50,315 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 14:28:50,315 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 14:28:50,316 - root - WARNING - No norm found in model
+[titan] 2025-07-23 14:28:50,316 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 14:28:50,542 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 14:28:50,885 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 14:28:51,042 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 14:28:51,273 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 14:28:51,274 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 14:28:51,299 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 14:28:51,320 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 14:28:51,477 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 14:28:58,659 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 14:28:58,665 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 14:28:58,665 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 14:28:58,667 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 14:28:58,669 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 14:28:58,670 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 14:28:58,670 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 14:28:58,670 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 14:28:58,670 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 14:28:58,670 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank6]: Traceback (most recent call last):
+[rank6]: File "", line 198, in _run_module_as_main
+[rank6]: File "", line 88, in _run_code
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank6]: main(config)
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank6]: return f(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank6]: output = model(
+[rank6]: ^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank6]: return self._call_impl(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank6]: return inner()
+[rank6]: ^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank6]: result = forward_call(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank6]: return func(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank6]: outputs = self.backbone(
+[rank6]: ^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank6]: return self._call_impl(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank6]: return forward_call(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank6]: hidden_states = mixer_block(
+[rank6]: ^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank6]: return self._call_impl(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank6]: return inner()
+[rank6]: ^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank6]: result = forward_call(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank6]: return fn(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank6]: return self._call_impl(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank6]: return forward_call(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank6]: hidden_states = self.norm(hidden_states)
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank6]: hidden_states = self.mixer(
+[rank6]: ^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank6]: return self._call_impl(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank6]: return forward_call(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 601, in forward
+[rank6]: return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 528, in torch_forward
+[rank6]: G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]
+[rank6]: ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~
+[rank6]: torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 6 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003904 has 316.00 MiB memory in use. Process 696035 has 316.00 MiB memory in use. Process 1114699 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/6/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/6/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/7/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/7/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..e2c3a81b296aa26e97573e6baec302d738dc4ee4
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/7/error.json
@@ -0,0 +1 @@
+{"message": {"message": "OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 7 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003906 has 316.00 MiB memory in use. Process 696037 has 316.00 MiB memory in use. Process 1114702 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 601, in forward\n return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 528, in torch_forward\n G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]\n ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~\ntorch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 7 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003906 has 316.00 MiB memory in use. Process 696037 has 316.00 MiB memory in use. Process 1114702 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)\n", "timestamp": "1753252283"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/7/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/7/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..3c52333ba51b4ad850ff772d5e70208cbf04035e
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/7/stderr.log
@@ -0,0 +1,387 @@
+[titan] 2025-07-23 14:27:46,256 - root - INFO - Starting job: default job
+[titan] 2025-07-23 14:27:46,256 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 14:27:46,257 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 14:27:47,217 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 14:27:47,220 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 14:27:47,268 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 14:27:47,268 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 14:27:47,268 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 14:27:47,411 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 14:27:47,997 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 14:27:47,998 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:47,998 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:48,594 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 14:27:48,594 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:27:48,594 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:27:48,594 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,263 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,297 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 14:28:40,298 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:28:40,298 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,298 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,563 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,649 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 14:28:40,649 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 14:28:40,649 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:40,649 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 14:28:46,975 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 14:28:47,679 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 14:28:47,795 - root - INFO - Building dataloader...
+[titan] 2025-07-23 14:28:47,797 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 14:28:47,799 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 14:28:50,147 - fla.layers.mamba2 - WARNING - The fast path is not available because one of `(selective_state_update)` is None. Falling back to the naive implementation. To install follow https://github.com/state-spaces/mamba/#installation
+[titan] 2025-07-23 14:28:50,148 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 14:28:50,264 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 14:28:50,316 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 14:28:50,316 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 14:28:50,317 - root - WARNING - No norm found in model
+[titan] 2025-07-23 14:28:50,317 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 14:28:50,540 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 14:28:50,884 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 14:28:51,042 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 14:28:51,273 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 14:28:51,275 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 14:28:51,299 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 14:28:51,311 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 14:28:51,462 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 14:28:58,659 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 14:28:58,661 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 14:28:58,662 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 14:28:58,662 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 14:28:58,667 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 14:28:58,667 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 14:28:58,668 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 14:28:58,669 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 14:28:58,669 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 14:28:58,669 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank7]: Traceback (most recent call last):
+[rank7]: File "", line 198, in _run_module_as_main
+[rank7]: File "", line 88, in _run_code
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank7]: main(config)
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank7]: return f(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank7]: output = model(
+[rank7]: ^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank7]: return self._call_impl(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank7]: return inner()
+[rank7]: ^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank7]: result = forward_call(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank7]: return func(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank7]: outputs = self.backbone(
+[rank7]: ^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank7]: return self._call_impl(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank7]: return forward_call(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank7]: hidden_states = mixer_block(
+[rank7]: ^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank7]: return self._call_impl(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank7]: return inner()
+[rank7]: ^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank7]: result = forward_call(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank7]: return fn(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank7]: return self._call_impl(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank7]: return forward_call(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank7]: hidden_states = self.norm(hidden_states)
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank7]: hidden_states = self.mixer(
+[rank7]: ^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank7]: return self._call_impl(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank7]: return forward_call(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 601, in forward
+[rank7]: return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 528, in torch_forward
+[rank7]: G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]
+[rank7]: ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~
+[rank7]: torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 7 has a total capacity of 95.00 GiB of which 85.09 GiB is free. Process 2003906 has 316.00 MiB memory in use. Process 696037 has 316.00 MiB memory in use. Process 1114702 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/7/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_07su7ijp/attempt_0/7/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/0/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/0/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..5e528387aa6633d23cecd3403ff73576f80ebb4e
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/0/error.json
@@ -0,0 +1 @@
+{"message": {"message": "OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 0 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003896 has 316.00 MiB memory in use. Process 696027 has 316.00 MiB memory in use. Process 1850004 has 21.15 GiB memory in use. Process 2711975 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 601, in forward\n return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 528, in torch_forward\n G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]\n ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~\ntorch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 0 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003896 has 316.00 MiB memory in use. Process 696027 has 316.00 MiB memory in use. Process 1850004 has 21.15 GiB memory in use. Process 2711975 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)\n", "timestamp": "1753242220"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/0/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/0/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..2b0ad9866c16de76d02238f355dd0e756b95d922
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/0/stderr.log
@@ -0,0 +1,467 @@
+[titan] 2025-07-23 11:13:04,987 - root - INFO - Starting job: default job
+[titan] 2025-07-23 11:13:04,987 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 11:13:04,988 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 11:13:04,988 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 11:13:05,005 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 11:13:05,098 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 11:13:05,098 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 11:13:05,099 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 11:13:07,424 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 11:13:07,703 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 11:13:07,704 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:07,704 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:08,426 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 11:13:08,426 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:13:08,426 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:08,427 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:37,776 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:37,867 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 11:35:37,867 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:35:37,867 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:37,868 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+Setting num_proc from 32 back to 1 for the train split to disable multiprocessing as it only contains one shard.
+[titan] 2025-07-23 11:35:37,949 - datasets.builder - WARNING - Setting num_proc from 32 back to 1 for the train split to disable multiprocessing as it only contains one shard.
+
Generating train split: 0 examples [00:00, ? examples/s]
Generating train split: 80099 examples [00:26, 3042.47 examples/s]
Generating train split: 160549 examples [00:26, 7173.32 examples/s]
Generating train split: 238364 examples [00:27, 12479.54 examples/s]
Generating train split: 317719 examples [00:28, 19491.94 examples/s]
Generating train split: 398626 examples [00:28, 28530.10 examples/s]
Generating train split: 465855 examples [00:29, 37693.93 examples/s]
Generating train split: 465855 examples [00:29, 15902.79 examples/s]
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:07,557 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:08,012 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 11:36:08,012 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:36:08,013 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:08,013 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+
Generating train split: 0 examples [00:00, ? examples/s]
Generating train split: 1000 examples [00:12, 78.77 examples/s]
Generating train split: 6000 examples [00:12, 630.35 examples/s]
Generating train split: 11000 examples [00:12, 1395.85 examples/s]
Generating train split: 15000 examples [00:13, 2233.00 examples/s]
Generating train split: 23000 examples [00:13, 4559.55 examples/s]
Generating train split: 29000 examples [00:13, 6465.62 examples/s]
Generating train split: 40000 examples [00:17, 4010.80 examples/s]
Generating train split: 46000 examples [00:17, 5377.58 examples/s]
Generating train split: 62000 examples [00:17, 10285.74 examples/s]
Generating train split: 99000 examples [00:17, 26324.39 examples/s]
Generating train split: 147000 examples [00:17, 51769.19 examples/s]
Generating train split: 194000 examples [00:17, 82327.73 examples/s]
Generating train split: 258000 examples [00:18, 129555.52 examples/s]
Generating train split: 294000 examples [00:18, 130956.12 examples/s]
Generating train split: 328000 examples [00:21, 30613.32 examples/s]
Generating train split: 348000 examples [00:22, 32858.25 examples/s]
Generating train split: 375000 examples [00:22, 41754.09 examples/s]
Generating train split: 393000 examples [00:22, 48086.56 examples/s]
Generating train split: 461000 examples [00:22, 86972.76 examples/s]
Generating train split: 516000 examples [00:22, 124968.40 examples/s]
Generating train split: 550000 examples [00:23, 134948.88 examples/s]
Generating train split: 583000 examples [00:23, 141159.88 examples/s]
Generating train split: 610000 examples [00:25, 34766.06 examples/s]
Generating train split: 627000 examples [00:26, 32716.27 examples/s]
Generating train split: 646000 examples [00:26, 39421.37 examples/s]
Generating train split: 662000 examples [00:26, 46607.78 examples/s]
Generating train split: 699000 examples [00:27, 69315.45 examples/s]
Generating train split: 734000 examples [00:27, 94778.34 examples/s]
Generating train split: 765000 examples [00:27, 118851.17 examples/s]
Generating train split: 797000 examples [00:27, 143077.16 examples/s]
Generating train split: 825000 examples [00:27, 145062.43 examples/s]
Generating train split: 854000 examples [00:27, 158083.63 examples/s]
Generating train split: 882000 examples [00:28, 84920.16 examples/s]
Generating train split: 899000 examples [00:30, 26313.70 examples/s]
Generating train split: 917000 examples [00:31, 28756.88 examples/s]
Generating train split: 935000 examples [00:31, 35613.93 examples/s]
Generating train split: 949000 examples [00:31, 42208.45 examples/s]
Generating train split: 988000 examples [00:31, 70365.40 examples/s]
Generating train split: 1026000 examples [00:31, 95465.98 examples/s]
Generating train split: 1094000 examples [00:32, 152861.96 examples/s]
Generating train split: 1127000 examples [00:32, 145706.54 examples/s]
Generating train split: 1157000 examples [00:32, 110881.71 examples/s]
Generating train split: 1175000 examples [00:35, 27441.46 examples/s]
Generating train split: 1200000 examples [00:35, 34913.88 examples/s]
Generating train split: 1224000 examples [00:35, 44712.95 examples/s]
Generating train split: 1250000 examples [00:36, 58761.00 examples/s]
Generating train split: 1284000 examples [00:36, 81395.73 examples/s]
Generating train split: 1338000 examples [00:36, 128525.71 examples/s]
Generating train split: 1369000 examples [00:36, 136475.76 examples/s]
Generating train split: 1409000 examples [00:36, 173846.27 examples/s]
Generating train split: 1441000 examples [00:36, 190447.03 examples/s]
Generating train split: 1474000 examples [00:36, 193923.95 examples/s]
Generating train split: 1504000 examples [00:37, 109017.38 examples/s]
Generating train split: 1532000 examples [00:40, 29506.27 examples/s]
Generating train split: 1552000 examples [00:40, 35873.40 examples/s]
Generating train split: 1577000 examples [00:40, 46484.56 examples/s]
Generating train split: 1594000 examples [00:40, 49677.85 examples/s]
Generating train split: 1610000 examples [00:40, 55319.98 examples/s]
Generating train split: 1628000 examples [00:41, 62723.09 examples/s]
Generating train split: 1666000 examples [00:41, 95259.48 examples/s]
Generating train split: 1689000 examples [00:41, 100628.73 examples/s]
Generating train split: 1716000 examples [00:41, 109261.99 examples/s]
Generating train split: 1734000 examples [00:42, 55615.74 examples/s]
Generating train split: 1754542 examples [00:44, 28443.26 examples/s]
Generating train split: 1762542 examples [00:44, 25399.64 examples/s]
Generating train split: 1770542 examples [00:44, 28438.89 examples/s]
Generating train split: 1778542 examples [00:45, 30877.07 examples/s]
Generating train split: 1799542 examples [00:45, 46509.14 examples/s]
Generating train split: 1839542 examples [00:45, 85571.06 examples/s]
Generating train split: 1859542 examples [00:45, 91572.89 examples/s]
Generating train split: 1879542 examples [00:45, 107163.51 examples/s]
Generating train split: 1902542 examples [00:45, 124313.27 examples/s]
Generating train split: 1930542 examples [00:45, 139887.64 examples/s]
Generating train split: 1951542 examples [00:45, 140199.81 examples/s]
Generating train split: 1969542 examples [00:46, 116782.29 examples/s]
Generating train split: 1983542 examples [00:46, 108503.11 examples/s]
Generating train split: 2003542 examples [00:46, 83225.73 examples/s]
Generating train split: 2014542 examples [00:47, 59741.45 examples/s]
Generating train split: 2030887 examples [00:48, 22224.97 examples/s]
Generating train split: 2039887 examples [00:49, 23921.41 examples/s]
Generating train split: 2055887 examples [00:49, 31029.80 examples/s]
Generating train split: 2069887 examples [00:49, 37557.68 examples/s]
Generating train split: 2086887 examples [00:49, 45017.82 examples/s]
Generating train split: 2115887 examples [00:49, 68373.36 examples/s]
Generating train split: 2148887 examples [00:50, 93216.76 examples/s]
Generating train split: 2167887 examples [00:50, 93599.65 examples/s]
Generating train split: 2206887 examples [00:50, 126466.27 examples/s]
Generating train split: 2240883 examples [00:50, 156674.80 examples/s]
Generating train split: 2264381 examples [00:51, 105786.53 examples/s]
Generating train split: 2281381 examples [00:52, 42896.95 examples/s]
Generating train split: 2295381 examples [00:53, 26945.80 examples/s]
Generating train split: 2323764 examples [00:53, 39783.13 examples/s]
Generating train split: 2342764 examples [00:53, 47554.36 examples/s]
Generating train split: 2363764 examples [00:54, 60535.16 examples/s]
Generating train split: 2410764 examples [00:54, 90397.69 examples/s]
Generating train split: 2433087 examples [00:54, 89054.20 examples/s]
Generating train split: 2460087 examples [00:54, 102656.76 examples/s]
Generating train split: 2480087 examples [00:54, 108891.93 examples/s]
Generating train split: 2496087 examples [00:55, 112481.79 examples/s]
Generating train split: 2516087 examples [00:55, 124586.58 examples/s]
Generating train split: 2541087 examples [00:55, 127280.52 examples/s]
Generating train split: 2558464 examples [00:56, 33937.50 examples/s]
Generating train split: 2574464 examples [00:58, 25133.25 examples/s]
Generating train split: 2582464 examples [00:58, 27819.88 examples/s]
Generating train split: 2609034 examples [00:58, 42474.46 examples/s]
Generating train split: 2626034 examples [00:58, 50951.86 examples/s]
Generating train split: 2655034 examples [00:58, 74577.06 examples/s]
Generating train split: 2672034 examples [00:58, 74343.41 examples/s]
Generating train split: 2686034 examples [00:59, 80538.52 examples/s]
Generating train split: 2701034 examples [00:59, 75891.60 examples/s]
Generating train split: 2750034 examples [00:59, 141052.03 examples/s]
Generating train split: 2781034 examples [00:59, 146199.53 examples/s]
Generating train split: 2806034 examples [00:59, 149675.12 examples/s]
Generating train split: 2829034 examples [01:00, 114688.95 examples/s]
Generating train split: 2849034 examples [01:01, 34606.13 examples/s]
Generating train split: 2864034 examples [01:02, 29655.49 examples/s]
Generating train split: 2882034 examples [01:02, 37030.59 examples/s]
Generating train split: 2899034 examples [01:02, 46137.49 examples/s]
Generating train split: 2915034 examples [01:03, 53047.54 examples/s]
Generating train split: 2949034 examples [01:03, 75293.65 examples/s]
Generating train split: 2962034 examples [01:03, 73355.47 examples/s]
Generating train split: 2978034 examples [01:03, 72346.75 examples/s]
Generating train split: 2992034 examples [01:03, 76996.73 examples/s]
Generating train split: 3006034 examples [01:04, 68445.54 examples/s]
Generating train split: 3035034 examples [01:04, 98963.62 examples/s]
Generating train split: 3048034 examples [01:04, 95919.43 examples/s]
Generating train split: 3059939 examples [01:05, 42857.52 examples/s]
Generating train split: 3068939 examples [01:05, 37609.63 examples/s]
Generating train split: 3076939 examples [01:05, 30964.96 examples/s]
Generating train split: 3085939 examples [01:06, 35533.50 examples/s]
Generating train split: 3093939 examples [01:06, 28046.23 examples/s]
Generating train split: 3098939 examples [01:07, 18551.32 examples/s]
Generating train split: 3114939 examples [01:07, 28682.80 examples/s]
Generating train split: 3141939 examples [01:07, 52713.63 examples/s]
Generating train split: 3158939 examples [01:07, 66713.12 examples/s]
Generating train split: 3175939 examples [01:07, 68233.33 examples/s]
Generating train split: 3189939 examples [01:08, 76782.39 examples/s]
Generating train split: 3206939 examples [01:08, 89536.23 examples/s]
Generating train split: 3223939 examples [01:08, 100522.12 examples/s]
Generating train split: 3240939 examples [01:08, 104249.67 examples/s]
Generating train split: 3255939 examples [01:08, 113969.33 examples/s]
Generating train split: 3277939 examples [01:08, 99316.96 examples/s]
Generating train split: 3294939 examples [01:08, 112160.17 examples/s]
Generating train split: 3314939 examples [01:09, 100822.51 examples/s]
Generating train split: 3329988 examples [01:09, 50714.07 examples/s]
Generating train split: 3338988 examples [01:10, 49468.56 examples/s]
Generating train split: 3357988 examples [01:10, 49020.02 examples/s]
Generating train split: 3366988 examples [01:10, 47938.75 examples/s]
Generating train split: 3374988 examples [01:11, 31723.15 examples/s]
Generating train split: 3381988 examples [01:11, 27555.37 examples/s]
Generating train split: 3393525 examples [01:11, 31982.64 examples/s]
Generating train split: 3424525 examples [01:12, 59703.18 examples/s]
Generating train split: 3452525 examples [01:12, 87263.33 examples/s]
Generating train split: 3474080 examples [01:12, 95174.82 examples/s]
Generating train split: 3491080 examples [01:12, 91779.96 examples/s]
Generating train split: 3516080 examples [01:12, 112772.20 examples/s]
Generating train split: 3537080 examples [01:12, 97306.11 examples/s]
Generating train split: 3561080 examples [01:13, 118640.87 examples/s]
Generating train split: 3580080 examples [01:13, 63987.51 examples/s]
Generating train split: 3594940 examples [01:13, 68945.95 examples/s]
Generating train split: 3610584 examples [01:14, 61578.03 examples/s]
Generating train split: 3619584 examples [01:14, 40926.90 examples/s]
Generating train split: 3627584 examples [01:15, 31529.53 examples/s]
Generating train split: 3636443 examples [01:15, 29495.61 examples/s]
Generating train split: 3645773 examples [01:15, 34375.51 examples/s]
Generating train split: 3653773 examples [01:16, 32424.98 examples/s]
Generating train split: 3680773 examples [01:16, 59217.46 examples/s]
Generating train split: 3698773 examples [01:16, 75579.07 examples/s]
Generating train split: 3716773 examples [01:16, 70911.42 examples/s]
Generating train split: 3730773 examples [01:16, 79755.42 examples/s]
Generating train split: 3745773 examples [01:16, 85501.40 examples/s]
Generating train split: 3760773 examples [01:17, 77473.26 examples/s]
Generating train split: 3778773 examples [01:17, 86520.55 examples/s]
Generating train split: 3813773 examples [01:17, 94307.81 examples/s]
Generating train split: 3829773 examples [01:17, 101169.11 examples/s]
Generating train split: 3840773 examples [01:17, 99993.37 examples/s]
Generating train split: 3858773 examples [01:18, 45816.99 examples/s]
Generating train split: 3870773 examples [01:18, 50010.45 examples/s]
Generating train split: 3890773 examples [01:19, 61653.49 examples/s]
Generating train split: 3899773 examples [01:19, 49073.17 examples/s]
Generating train split: 3911773 examples [01:19, 38543.44 examples/s]
Generating train split: 3938073 examples [01:20, 49379.87 examples/s]
Generating train split: 3946073 examples [01:20, 45240.50 examples/s]
Generating train split: 3972073 examples [01:20, 69886.84 examples/s]
Generating train split: 3987073 examples [01:20, 62675.26 examples/s]
Generating train split: 4009073 examples [01:21, 82180.32 examples/s]
Generating train split: 4024073 examples [01:21, 62339.96 examples/s]
Generating train split: 4042073 examples [01:21, 74809.74 examples/s]
Generating train split: 4071073 examples [01:21, 87811.31 examples/s]
Generating train split: 4087073 examples [01:21, 95489.25 examples/s]
Generating train split: 4115073 examples [01:22, 102885.53 examples/s]
Generating train split: 4134073 examples [01:23, 50787.30 examples/s]
Generating train split: 4152073 examples [01:23, 54744.94 examples/s]
Generating train split: 4173073 examples [01:23, 57620.73 examples/s]
Generating train split: 4185073 examples [01:24, 48975.10 examples/s]
Generating train split: 4193073 examples [01:24, 51623.55 examples/s]
Generating train split: 4202073 examples [01:24, 38859.21 examples/s]
Generating train split: 4210073 examples [01:24, 42031.18 examples/s]
Generating train split: 4227073 examples [01:25, 48349.70 examples/s]
Generating train split: 4235073 examples [01:25, 44976.34 examples/s]
Generating train split: 4255073 examples [01:25, 52112.71 examples/s]
Generating train split: 4262073 examples [01:25, 50484.96 examples/s]
Generating train split: 4307073 examples [01:25, 100141.91 examples/s]
Generating train split: 4322325 examples [01:26, 85229.11 examples/s]
Generating train split: 4341325 examples [01:26, 91047.59 examples/s]
Generating train split: 4359325 examples [01:26, 74739.42 examples/s]
Generating train split: 4381325 examples [01:26, 91474.89 examples/s]
Generating train split: 4398325 examples [01:27, 51763.86 examples/s]
Generating train split: 4418325 examples [01:27, 54119.66 examples/s]
Generating train split: 4430325 examples [01:28, 49673.39 examples/s]
Generating train split: 4440325 examples [01:28, 45289.84 examples/s]
Generating train split: 4449325 examples [01:28, 41808.76 examples/s]
Generating train split: 4456325 examples [01:29, 36674.41 examples/s]
Generating train split: 4465325 examples [01:29, 31242.85 examples/s]
Generating train split: 4491325 examples [01:29, 54616.00 examples/s]
Generating train split: 4508325 examples [01:29, 61333.25 examples/s]
Generating train split: 4525325 examples [01:30, 66517.84 examples/s]
Generating train split: 4543325 examples [01:30, 78828.53 examples/s]
Generating train split: 4561325 examples [01:30, 60483.20 examples/s]
Generating train split: 4573325 examples [01:30, 62372.36 examples/s]
Generating train split: 4590325 examples [01:30, 73378.41 examples/s]
Generating train split: 4618325 examples [01:31, 91912.92 examples/s]
Generating train split: 4634325 examples [01:31, 66501.07 examples/s]
Generating train split: 4644325 examples [01:31, 68241.92 examples/s]
Generating train split: 4656325 examples [01:32, 40779.06 examples/s]
Generating train split: 4685325 examples [01:32, 47793.37 examples/s]
Generating train split: 4704325 examples [01:33, 58615.89 examples/s]
Generating train split: 4723325 examples [01:33, 43912.10 examples/s]
Generating train split: 4731325 examples [01:33, 44368.43 examples/s]
Generating train split: 4739325 examples [01:33, 48157.23 examples/s]
Generating train split: 4747325 examples [01:34, 45507.88 examples/s]
Generating train split: 4755325 examples [01:34, 46320.80 examples/s]
Generating train split: 4762325 examples [01:34, 48891.10 examples/s]
Generating train split: 4778325 examples [01:34, 51043.14 examples/s]
Generating train split: 4796325 examples [01:34, 68084.40 examples/s]
Generating train split: 4812325 examples [01:35, 74373.12 examples/s]
Generating train split: 4833325 examples [01:35, 78865.06 examples/s]
Generating train split: 4849325 examples [01:35, 81553.37 examples/s]
Generating train split: 4864325 examples [01:35, 69598.35 examples/s]
Generating train split: 4908325 examples [01:35, 111293.05 examples/s]
Generating train split: 4927325 examples [01:36, 70529.74 examples/s]
Generating train split: 4940325 examples [01:36, 77461.78 examples/s]
Generating train split: 4959325 examples [01:36, 86128.84 examples/s]
Generating train split: 4978262 examples [01:37, 75095.07 examples/s]
Generating train split: 4988262 examples [01:37, 44324.59 examples/s]
Generating train split: 4999262 examples [01:37, 49631.06 examples/s]
Generating train split: 5016262 examples [01:38, 57484.57 examples/s]
Generating train split: 5025262 examples [01:38, 43864.76 examples/s]
Generating train split: 5042262 examples [01:38, 39755.09 examples/s]
Generating train split: 5059262 examples [01:39, 50412.54 examples/s]
Generating train split: 5073262 examples [01:39, 52350.96 examples/s]
Generating train split: 5082262 examples [01:39, 56420.59 examples/s]
Generating train split: 5101262 examples [01:39, 73774.76 examples/s]
Generating train split: 5112688 examples [01:39, 68728.80 examples/s]
Generating train split: 5126688 examples [01:40, 69372.97 examples/s]
Generating train split: 5142688 examples [01:40, 81000.33 examples/s]
Generating train split: 5153688 examples [01:40, 82938.35 examples/s]
Generating train split: 5164688 examples [01:40, 67450.75 examples/s]
Generating train split: 5188688 examples [01:40, 86903.06 examples/s]
Generating train split: 5211656 examples [01:41, 63302.14 examples/s]
Generating train split: 5230656 examples [01:41, 69447.71 examples/s]
Generating train split: 5252656 examples [01:41, 80777.48 examples/s]
Generating train split: 5262656 examples [01:42, 36486.61 examples/s]
Generating train split: 5273656 examples [01:42, 41173.32 examples/s]
Generating train split: 5288656 examples [01:42, 48777.30 examples/s]
Generating train split: 5296656 examples [01:43, 35267.72 examples/s]
Generating train split: 5313656 examples [01:43, 41511.70 examples/s]
Generating train split: 5339656 examples [01:43, 60033.99 examples/s]
Generating train split: 5354656 examples [01:44, 66607.28 examples/s]
Generating train split: 5366656 examples [01:44, 73620.97 examples/s]
Generating train split: 5383656 examples [01:44, 85749.37 examples/s]
Generating train split: 5400030 examples [01:44, 70047.58 examples/s]
Generating train split: 5419030 examples [01:44, 73396.49 examples/s]
Generating train split: 5434030 examples [01:45, 74328.64 examples/s]
Generating train split: 5451030 examples [01:45, 87534.94 examples/s]
Generating train split: 5467030 examples [01:45, 57440.44 examples/s]
Generating train split: 5487030 examples [01:46, 57850.06 examples/s]
Generating train split: 5503030 examples [01:46, 61529.12 examples/s]
Generating train split: 5518030 examples [01:46, 63571.96 examples/s]
Generating train split: 5533030 examples [01:47, 40729.62 examples/s]
Generating train split: 5542030 examples [01:47, 44695.62 examples/s]
Generating train split: 5554030 examples [01:47, 43748.04 examples/s]
Generating train split: 5563030 examples [01:47, 48666.24 examples/s]
Generating train split: 5571030 examples [01:47, 48696.27 examples/s]
Generating train split: 5580030 examples [01:48, 41136.63 examples/s]
Generating train split: 5588030 examples [01:48, 42536.14 examples/s]
Generating train split: 5596030 examples [01:48, 34852.58 examples/s]
Generating train split: 5614030 examples [01:48, 53800.20 examples/s]
Generating train split: 5638030 examples [01:48, 73271.77 examples/s]
Generating train split: 5664030 examples [01:49, 100659.38 examples/s]
Generating train split: 5683030 examples [01:49, 87438.28 examples/s]
Generating train split: 5713030 examples [01:49, 92636.53 examples/s]
Generating train split: 5730030 examples [01:49, 87270.02 examples/s]
Generating train split: 5756030 examples [01:50, 110793.04 examples/s]
Generating train split: 5774030 examples [01:50, 96080.52 examples/s]
Generating train split: 5791030 examples [01:50, 73749.95 examples/s]
Generating train split: 5811030 examples [01:50, 79501.07 examples/s]
Generating train split: 5821030 examples [01:51, 33317.26 examples/s]
Generating train split: 5838030 examples [01:52, 34798.80 examples/s]
Generating train split: 5846030 examples [01:52, 37963.63 examples/s]
Generating train split: 5863030 examples [01:52, 40202.28 examples/s]
Generating train split: 5877030 examples [01:53, 45381.84 examples/s]
Generating train split: 5894030 examples [01:53, 54393.88 examples/s]
Generating train split: 5902030 examples [01:53, 57608.76 examples/s]
Generating train split: 5911030 examples [01:53, 52916.91 examples/s]
Generating train split: 5941030 examples [01:53, 70279.85 examples/s]
Generating train split: 5963861 examples [01:54, 84868.83 examples/s]
Generating train split: 5982861 examples [01:54, 93174.51 examples/s]
Generating train split: 6006951 examples [01:54, 79292.34 examples/s]
Generating train split: 6020951 examples [01:54, 84184.43 examples/s]
Generating train split: 6036951 examples [01:54, 92939.26 examples/s]
Generating train split: 6053951 examples [01:55, 64288.02 examples/s]
Generating train split: 6065951 examples [01:55, 58876.16 examples/s]
Generating train split: 6074951 examples [01:55, 55869.90 examples/s]
Generating train split: 6084951 examples [01:56, 40005.32 examples/s]
Generating train split: 6091951 examples [01:56, 35998.70 examples/s]
Generating train split: 6109951 examples [01:56, 42204.77 examples/s]
Generating train split: 6117951 examples [01:57, 36635.67 examples/s]
Generating train split: 6130951 examples [01:57, 44542.22 examples/s]
Generating train split: 6147951 examples [01:57, 56210.52 examples/s]
Generating train split: 6159951 examples [01:57, 57047.97 examples/s]
Generating train split: 6168951 examples [01:57, 61564.81 examples/s]
Generating train split: 6177951 examples [01:58, 48982.56 examples/s]
Generating train split: 6204951 examples [01:58, 76706.19 examples/s]
Generating train split: 6224951 examples [01:58, 90478.28 examples/s]
Generating train split: 6251951 examples [01:58, 84581.13 examples/s]
Generating train split: 6268951 examples [01:58, 88019.65 examples/s]
Generating train split: 6290951 examples [01:59, 104820.35 examples/s]
Generating train split: 6305951 examples [01:59, 72861.98 examples/s]
Generating train split: 6321951 examples [02:00, 53860.03 examples/s]
Generating train split: 6339951 examples [02:00, 57658.46 examples/s]
Generating train split: 6351995 examples [02:00, 60803.88 examples/s]
Generating train split: 6371995 examples [02:00, 69375.42 examples/s]
Generating train split: 6387995 examples [02:01, 44937.34 examples/s]
Generating train split: 6395995 examples [02:01, 39656.17 examples/s]
Generating train split: 6412995 examples [02:01, 44156.08 examples/s]
Generating train split: 6427995 examples [02:02, 52442.29 examples/s]
Generating train split: 6445995 examples [02:02, 57636.49 examples/s]
Generating train split: 6457995 examples [02:02, 62604.73 examples/s]
Generating train split: 6466995 examples [02:02, 52765.05 examples/s]
Generating train split: 6476995 examples [02:02, 55915.71 examples/s]
Generating train split: 6494995 examples [02:03, 68586.49 examples/s]
Generating train split: 6535995 examples [02:03, 118633.79 examples/s]
Generating train split: 6554995 examples [02:03, 66105.89 examples/s]
Generating train split: 6572995 examples [02:04, 61373.86 examples/s]
Generating train split: 6596113 examples [02:04, 80300.92 examples/s]
Generating train split: 6611074 examples [02:04, 70945.86 examples/s]
Generating train split: 6635074 examples [02:04, 91167.41 examples/s]
Generating train split: 6652469 examples [02:05, 38740.67 examples/s]
Generating train split: 6664469 examples [02:06, 32295.88 examples/s]
Generating train split: 6681469 examples [02:06, 38167.63 examples/s]
Generating train split: 6690469 examples [02:06, 42470.00 examples/s]
Generating train split: 6727469 examples [02:07, 71464.88 examples/s]
Generating train split: 6745469 examples [02:07, 79472.58 examples/s]
Generating train split: 6765469 examples [02:07, 92746.93 examples/s]
Generating train split: 6788469 examples [02:07, 83155.69 examples/s]
Generating train split: 6801469 examples [02:08, 61421.59 examples/s]
Generating train split: 6829469 examples [02:08, 88068.77 examples/s]
Generating train split: 6846469 examples [02:08, 62279.25 examples/s]
Generating train split: 6863469 examples [02:09, 57628.98 examples/s]
Generating train split: 6882469 examples [02:09, 69066.79 examples/s]
Generating train split: 6895469 examples [02:09, 61406.46 examples/s]
Generating train split: 6913469 examples [02:09, 75398.85 examples/s]
Generating train split: 6931605 examples [02:10, 31264.76 examples/s]
Generating train split: 6950141 examples [02:11, 37217.61 examples/s]
Generating train split: 6968141 examples [02:11, 45993.73 examples/s]
Generating train split: 6982141 examples [02:11, 52646.34 examples/s]
Generating train split: 7006141 examples [02:11, 74191.01 examples/s]
Generating train split: 7032141 examples [02:12, 69940.29 examples/s]
Generating train split: 7051141 examples [02:12, 82797.39 examples/s]
Generating train split: 7066141 examples [02:12, 81550.04 examples/s]
Generating train split: 7085141 examples [02:12, 56061.48 examples/s]
Generating train split: 7094141 examples [02:13, 53062.56 examples/s]
Generating train split: 7108141 examples [02:13, 44722.69 examples/s]
Generating train split: 7122141 examples [02:13, 53094.13 examples/s]
Generating train split: 7131141 examples [02:13, 55037.56 examples/s]
Generating train split: 7140141 examples [02:14, 57770.35 examples/s]
Generating train split: 7157141 examples [02:14, 62299.33 examples/s]
Generating train split: 7167141 examples [02:14, 59546.95 examples/s]
Generating train split: 7185141 examples [02:15, 43227.19 examples/s]
Generating train split: 7192141 examples [02:15, 40073.14 examples/s]
Generating train split: 7202141 examples [02:15, 41617.74 examples/s]
Generating train split: 7213141 examples [02:15, 43996.03 examples/s]
Generating train split: 7228141 examples [02:15, 57175.00 examples/s]
Generating train split: 7253141 examples [02:16, 78607.72 examples/s]
Generating train split: 7271141 examples [02:16, 89220.81 examples/s]
Generating train split: 7287141 examples [02:16, 78148.82 examples/s]
Generating train split: 7310141 examples [02:16, 100572.33 examples/s]
Generating train split: 7326141 examples [02:17, 64856.32 examples/s]
Generating train split: 7342141 examples [02:17, 69541.48 examples/s]
Generating train split: 7360141 examples [02:17, 70619.66 examples/s]
Generating train split: 7369141 examples [02:17, 49266.25 examples/s]
Generating train split: 7378141 examples [02:18, 49599.09 examples/s]
Generating train split: 7394141 examples [02:18, 61172.00 examples/s]
Generating train split: 7416141 examples [02:18, 64418.34 examples/s]
Generating train split: 7434141 examples [02:18, 67321.41 examples/s]
Generating train split: 7443141 examples [02:19, 61113.60 examples/s]
Generating train split: 7451141 examples [02:19, 48867.78 examples/s]
Generating train split: 7467141 examples [02:19, 39258.65 examples/s]
Generating train split: 7475141 examples [02:20, 38243.26 examples/s]
Generating train split: 7503141 examples [02:20, 66525.24 examples/s]
Generating train split: 7522141 examples [02:20, 68207.18 examples/s]
Generating train split: 7541141 examples [02:20, 72302.28 examples/s]
Generating train split: 7563141 examples [02:20, 79253.82 examples/s]
Generating train split: 7579146 examples [02:21, 70635.67 examples/s]
Generating train split: 7605146 examples [02:21, 80460.26 examples/s]
Generating train split: 7625146 examples [02:21, 89363.24 examples/s]
Generating train split: 7642146 examples [02:22, 57370.05 examples/s]
Generating train split: 7652146 examples [02:22, 60339.90 examples/s]
Generating train split: 7669253 examples [02:22, 69545.40 examples/s]
Generating train split: 7678253 examples [02:22, 63516.88 examples/s]
Generating train split: 7698253 examples [02:22, 83535.53 examples/s]
Generating train split: 7714253 examples [02:23, 83173.19 examples/s]
Generating train split: 7731253 examples [02:23, 81152.87 examples/s]
Generating train split: 7750253 examples [02:23, 71806.90 examples/s]
Generating train split: 7764253 examples [02:24, 51418.53 examples/s]
Generating train split: 7772253 examples [02:24, 49242.84 examples/s]
Generating train split: 7791253 examples [02:24, 55758.25 examples/s]
Generating train split: 7801105 examples [02:24, 43763.68 examples/s]
Generating train split: 7810105 examples [02:25, 49036.23 examples/s]
Generating train split: 7819105 examples [02:25, 47902.66 examples/s]
Generating train split: 7833105 examples [02:25, 56349.53 examples/s]
Generating train split: 7851105 examples [02:25, 72891.01 examples/s]
Generating train split: 7864173 examples [02:26, 49867.17 examples/s]
Generating train split: 7891173 examples [02:26, 63420.81 examples/s]
Generating train split: 7909173 examples [02:26, 68577.83 examples/s]
Generating train split: 7926044 examples [02:26, 57404.15 examples/s]
Generating train split: 7939605 examples [02:27, 66846.20 examples/s]
Generating train split: 7948605 examples [02:27, 65667.34 examples/s]
Generating train split: 7974605 examples [02:27, 92300.25 examples/s]
Generating train split: 7986605 examples [02:27, 66854.78 examples/s]
Generating train split: 8004605 examples [02:28, 54856.89 examples/s]
Generating train split: 8012605 examples [02:28, 48898.81 examples/s]
Generating train split: 8040361 examples [02:28, 69727.93 examples/s]
Generating train split: 8056361 examples [02:29, 49904.42 examples/s]
Generating train split: 8077361 examples [02:29, 55054.91 examples/s]
Generating train split: 8086361 examples [02:29, 55196.03 examples/s]
Generating train split: 8105361 examples [02:29, 72109.85 examples/s]
Generating train split: 8115361 examples [02:29, 60534.53 examples/s]
Generating train split: 8124361 examples [02:30, 61715.91 examples/s]
Generating train split: 8138361 examples [02:30, 66435.89 examples/s]
Generating train split: 8156361 examples [02:30, 84647.74 examples/s]
Generating train split: 8180361 examples [02:30, 97277.39 examples/s]
Generating train split: 8197361 examples [02:30, 83462.51 examples/s]
Generating train split: 8219361 examples [02:30, 100263.72 examples/s]
Generating train split: 8240361 examples [02:31, 82491.05 examples/s]
Generating train split: 8263361 examples [02:31, 99843.80 examples/s]
Generating train split: 8280361 examples [02:31, 81466.68 examples/s]
Generating train split: 8297361 examples [02:32, 58624.67 examples/s]
Generating train split: 8307361 examples [02:32, 62568.15 examples/s]
Generating train split: 8336361 examples [02:33, 53951.69 examples/s]
Generating train split: 8353361 examples [02:33, 61576.03 examples/s]
Generating train split: 8375840 examples [02:33, 66785.30 examples/s]
Generating train split: 8395840 examples [02:33, 72835.07 examples/s]
Generating train split: 8408840 examples [02:33, 69288.77 examples/s]
Generating train split: 8419038 examples [02:34, 72302.54 examples/s]
Generating train split: 8433584 examples [02:34, 75674.13 examples/s]
Generating train split: 8445584 examples [02:34, 54407.97 examples/s]
Generating train split: 8454584 examples [02:34, 58114.16 examples/s]
Generating train split: 8467077 examples [02:35, 36489.75 examples/s]
Generating train split: 8484077 examples [02:35, 48956.33 examples/s]
Generating train split: 8493077 examples [02:35, 52702.71 examples/s]
Generating train split: 8501077 examples [02:35, 50041.29 examples/s]
Generating train split: 8514077 examples [02:35, 59740.41 examples/s]
Generating train split: 8530077 examples [02:36, 75581.92 examples/s]
Generating train split: 8552287 examples [02:36, 41536.50 examples/s]
Generating train split: 8560287 examples [02:37, 45213.44 examples/s]
Generating train split: 8571287 examples [02:37, 52718.85 examples/s]
Generating train split: 8594436 examples [02:37, 77559.55 examples/s]
Generating train split: 8610436 examples [02:37, 91040.42 examples/s]
Generating train split: 8639355 examples [02:37, 71272.71 examples/s]
Generating train split: 8656355 examples [02:38, 67774.22 examples/s]
Generating train split: 8671355 examples [02:38, 70138.65 examples/s]
Generating train split: 8681355 examples [02:38, 51167.02 examples/s]
Generating train split: 8690355 examples [02:38, 53289.72 examples/s]
Generating train split: 8706355 examples [02:39, 41322.72 examples/s]
Generating train split: 8734355 examples [02:39, 50158.73 examples/s]
Generating train split: 8741355 examples [02:40, 46326.66 examples/s]
Generating train split: 8762355 examples [02:40, 63775.89 examples/s]
Generating train split: 8783355 examples [02:40, 66072.45 examples/s]
Generating train split: 8799355 examples [02:40, 66012.48 examples/s]
Generating train split: 8810355 examples [02:40, 65208.73 examples/s]
Generating train split: 8825355 examples [02:41, 36625.25 examples/s]
Generating train split: 8844355 examples [02:41, 47580.79 examples/s]
Generating train split: 8853355 examples [02:42, 52074.38 examples/s]
Generating train split: 8888699 examples [02:42, 85606.62 examples/s]
Generating train split: 8913699 examples [02:42, 97911.55 examples/s]
Generating train split: 8927699 examples [02:42, 64867.65 examples/s]
Generating train split: 8937699 examples [02:43, 55436.93 examples/s]
Generating train split: 8959699 examples [02:43, 56886.58 examples/s]
Generating train split: 8975699 examples [02:43, 60401.10 examples/s]
Generating train split: 8993373 examples [02:44, 66507.97 examples/s]
Generating train split: 9006926 examples [02:44, 55854.99 examples/s]
Generating train split: 9015926 examples [02:44, 56670.08 examples/s]
Generating train split: 9028926 examples [02:44, 60961.06 examples/s]
Generating train split: 9038926 examples [02:44, 61630.98 examples/s]
Generating train split: 9048926 examples [02:45, 52147.30 examples/s]
Generating train split: 9056926 examples [02:45, 31736.96 examples/s]
Generating train split: 9075926 examples [02:46, 40706.10 examples/s]
Generating train split: 9093926 examples [02:46, 53224.84 examples/s]
Generating train split: 9101926 examples [02:46, 56488.83 examples/s]
Generating train split: 9109926 examples [02:46, 54443.85 examples/s]
Generating train split: 9136926 examples [02:47, 47332.05 examples/s]
Generating train split: 9154926 examples [02:47, 61801.04 examples/s]
Generating train split: 9171926 examples [02:47, 71188.01 examples/s]
Generating train split: 9191926 examples [02:47, 86983.39 examples/s]
Generating train split: 9204936 examples [02:47, 76040.66 examples/s]
Generating train split: 9220936 examples [02:47, 82227.13 examples/s]
Generating train split: 9233936 examples [02:48, 60215.51 examples/s]
Generating train split: 9251936 examples [02:48, 43694.63 examples/s]
Generating train split: 9270936 examples [02:49, 57288.05 examples/s]
Generating train split: 9291936 examples [02:49, 61559.51 examples/s]
Generating train split: 9301936 examples [02:49, 55927.87 examples/s]
Generating train split: 9309936 examples [02:49, 50983.44 examples/s]
Generating train split: 9318936 examples [02:50, 43107.98 examples/s]
Generating train split: 9336936 examples [02:50, 37136.75 examples/s]
Generating train split: 9347347 examples [02:51, 34502.52 examples/s]
Generating train split: 9363347 examples [02:51, 45377.53 examples/s]
Generating train split: 9381347 examples [02:51, 56837.43 examples/s]
Generating train split: 9410347 examples [02:51, 81875.64 examples/s]
Generating train split: 9433347 examples [02:51, 85160.53 examples/s]
Generating train split: 9445347 examples [02:52, 74937.34 examples/s]
Generating train split: 9459347 examples [02:52, 73301.20 examples/s]
Generating train split: 9498347 examples [02:52, 115247.88 examples/s]
Generating train split: 9516347 examples [02:52, 95322.32 examples/s]
Generating train split: 9531347 examples [02:53, 47283.85 examples/s]
Generating train split: 9553497 examples [02:53, 56351.14 examples/s]
Generating train split: 9563497 examples [02:53, 60833.11 examples/s]
Generating train split: 9588497 examples [02:54, 64243.78 examples/s]
Generating train split: 9597497 examples [02:54, 63312.73 examples/s]
Generating train split: 9606497 examples [02:54, 61055.85 examples/s]
Generating train split: 9617497 examples [02:54, 48609.36 examples/s]
Generating train split: 9625497 examples [02:55, 43201.46 examples/s]
Generating train split: 9633497 examples [02:55, 37316.03 examples/s]
Generating train split: 9652497 examples [02:55, 45340.76 examples/s]
Generating train split: 9672148 examples [02:56, 55476.96 examples/s]
Generating train split: 9704148 examples [02:56, 85095.92 examples/s]
Generating train split: 9719148 examples [02:56, 93981.31 examples/s]
Generating train split: 9747148 examples [02:56, 124017.00 examples/s]
Generating train split: 9766148 examples [02:56, 93264.44 examples/s]
Generating train split: 9785148 examples [02:56, 94207.87 examples/s]
Generating train split: 9804148 examples [02:57, 104947.71 examples/s]
Generating train split: 9821148 examples [02:57, 65484.48 examples/s]
Generating train split: 9841148 examples [02:58, 42034.93 examples/s]
Generating train split: 9858148 examples [02:58, 51870.40 examples/s]
Generating train split: 9877148 examples [02:58, 55406.48 examples/s]
Generating train split: 9897148 examples [02:58, 68985.15 examples/s]
Generating train split: 9911857 examples [02:59, 43660.20 examples/s]
Generating train split: 9938857 examples [03:00, 52892.20 examples/s]
Generating train split: 9951404 examples [03:00, 59646.23 examples/s]
Generating train split: 9977404 examples [03:00, 59686.85 examples/s]
Generating train split: 9996404 examples [03:00, 74082.57 examples/s]
Generating train split: 10013404 examples [03:00, 83208.53 examples/s]
Generating train split: 10042404 examples [03:00, 111330.83 examples/s]
Generating train split: 10060404 examples [03:01, 111686.93 examples/s]
Generating train split: 10080404 examples [03:01, 91918.44 examples/s]
Generating train split: 10096404 examples [03:02, 42643.50 examples/s]
Generating train split: 10125934 examples [03:02, 59119.49 examples/s]
Generating train split: 10141934 examples [03:03, 47687.99 examples/s]
Generating train split: 10164934 examples [03:03, 56038.12 examples/s]
Generating train split: 10174934 examples [03:04, 33069.80 examples/s]
Generating train split: 10200934 examples [03:04, 44146.38 examples/s]
Generating train split: 10208934 examples [03:04, 45631.90 examples/s]
Generating train split: 10246934 examples [03:04, 76921.04 examples/s]
Generating train split: 10293063 examples [03:05, 118760.62 examples/s]
Generating train split: 10315303 examples [03:05, 115681.70 examples/s]
Generating train split: 10342303 examples [03:05, 124617.25 examples/s]
Generating train split: 10359303 examples [03:05, 108035.65 examples/s]
Generating train split: 10377303 examples [03:06, 43930.74 examples/s]
Generating train split: 10395303 examples [03:07, 48492.39 examples/s]
Generating train split: 10410303 examples [03:07, 56992.46 examples/s]
Generating train split: 10429303 examples [03:07, 60740.48 examples/s]
Generating train split: 10448948 examples [03:07, 75672.93 examples/s]
Generating train split: 10476544 examples [03:07, 95335.95 examples/s]
Generating train split: 10495544 examples [03:08, 67655.21 examples/s]
Generating train split: 10513544 examples [03:08, 62666.16 examples/s]
Generating train split: 10532544 examples [03:08, 57900.93 examples/s]
Generating train split: 10556544 examples [03:09, 77881.65 examples/s]
Generating train split: 10573544 examples [03:09, 84600.48 examples/s]
Generating train split: 10602544 examples [03:09, 94903.56 examples/s]
Generating train split: 10620544 examples [03:09, 104396.85 examples/s]
Generating train split: 10639544 examples [03:09, 86498.48 examples/s]
Generating train split: 10657544 examples [03:10, 55875.23 examples/s]
Generating train split: 10686544 examples [03:10, 77564.79 examples/s]
Generating train split: 10703544 examples [03:11, 55032.87 examples/s]
Generating train split: 10719544 examples [03:11, 52746.32 examples/s]
Generating train split: 10733544 examples [03:11, 56499.97 examples/s]
Generating train split: 10755835 examples [03:12, 56426.29 examples/s]
Generating train split: 10780835 examples [03:12, 64736.71 examples/s]
Generating train split: 10790835 examples [03:12, 68117.15 examples/s]
Generating train split: 10809835 examples [03:12, 85021.55 examples/s]
Generating train split: 10822835 examples [03:12, 86590.38 examples/s]
Generating train split: 10839835 examples [03:12, 94411.35 examples/s]
Generating train split: 10852835 examples [03:13, 97918.03 examples/s]
Generating train split: 10871835 examples [03:13, 73007.66 examples/s]
Generating train split: 10881835 examples [03:13, 74383.06 examples/s]
Generating train split: 10903306 examples [03:13, 71725.16 examples/s]
Generating train split: 10919306 examples [03:14, 82806.02 examples/s]
Generating train split: 10937306 examples [03:14, 44468.85 examples/s]
Generating train split: 10951306 examples [03:15, 35832.65 examples/s]
Generating train split: 10960306 examples [03:15, 38712.75 examples/s]
Generating train split: 10990987 examples [03:15, 62505.46 examples/s]
Generating train split: 11007987 examples [03:15, 70945.25 examples/s]
Generating train split: 11023987 examples [03:16, 57838.38 examples/s]
Generating train split: 11067987 examples [03:16, 93962.18 examples/s]
Generating train split: 11087987 examples [03:16, 93076.51 examples/s]
Generating train split: 11104987 examples [03:17, 60264.88 examples/s]
Generating train split: 11114987 examples [03:17, 54009.69 examples/s]
Generating train split: 11123987 examples [03:17, 51378.45 examples/s]
Generating train split: 11140987 examples [03:18, 55324.13 examples/s]
Generating train split: 11154987 examples [03:18, 50972.43 examples/s]
Generating train split: 11178915 examples [03:18, 70759.58 examples/s]
Generating train split: 11215244 examples [03:19, 64333.40 examples/s]
Generating train split: 11225244 examples [03:19, 66911.78 examples/s]
Generating train split: 11242244 examples [03:19, 70583.40 examples/s]
Generating train split: 11277244 examples [03:19, 108136.97 examples/s]
Generating train split: 11298244 examples [03:20, 83627.93 examples/s]
Generating train split: 11316244 examples [03:20, 82940.70 examples/s]
Generating train split: 11335244 examples [03:20, 80077.78 examples/s]
Generating train split: 11345244 examples [03:20, 76404.48 examples/s]
Generating train split: 11366244 examples [03:21, 55457.87 examples/s]
Generating train split: 11376244 examples [03:21, 51318.27 examples/s]
Generating train split: 11398244 examples [03:21, 61374.56 examples/s]
Generating train split: 11409244 examples [03:22, 55721.03 examples/s]
Generating train split: 11425459 examples [03:22, 67896.80 examples/s]
Generating train split: 11446459 examples [03:22, 74029.34 examples/s]
Generating train split: 11456459 examples [03:22, 63841.95 examples/s]
Generating train split: 11465459 examples [03:22, 67061.39 examples/s]
Generating train split: 11479459 examples [03:23, 60386.19 examples/s]
Generating train split: 11512318 examples [03:23, 99568.49 examples/s]
Generating train split: 11541318 examples [03:23, 122666.54 examples/s]
Generating train split: 11558318 examples [03:23, 120486.19 examples/s]
Generating train split: 11579318 examples [03:24, 70194.49 examples/s]
Generating train split: 11593848 examples [03:24, 68714.96 examples/s]
Generating train split: 11613848 examples [03:24, 59434.11 examples/s]
Generating train split: 11631645 examples [03:24, 61535.75 examples/s]
Generating train split: 11641645 examples [03:25, 47538.80 examples/s]
Generating train split: 11649645 examples [03:25, 47276.61 examples/s]
Generating train split: 11660645 examples [03:25, 48893.63 examples/s]
Generating train split: 11678645 examples [03:25, 64302.55 examples/s]
Generating train split: 11692645 examples [03:26, 74410.44 examples/s]
Generating train split: 11708645 examples [03:26, 80560.89 examples/s]
Generating train split: 11722521 examples [03:26, 86952.15 examples/s]
Generating train split: 11755521 examples [03:26, 85983.69 examples/s]
Generating train split: 11775521 examples [03:27, 72358.13 examples/s]
Generating train split: 11787521 examples [03:27, 66560.79 examples/s]
Generating train split: 11795521 examples [03:27, 51664.73 examples/s]
Generating train split: 11807521 examples [03:27, 51110.66 examples/s]
Generating train split: 11833521 examples [03:28, 66752.78 examples/s]
Generating train split: 11851521 examples [03:28, 59809.13 examples/s]
Generating train split: 11864521 examples [03:28, 64616.32 examples/s]
Generating train split: 11883521 examples [03:28, 80209.96 examples/s]
Generating train split: 11898521 examples [03:28, 80645.39 examples/s]
Generating train split: 11916521 examples [03:29, 54023.00 examples/s]
Generating train split: 11939521 examples [03:29, 66999.48 examples/s]
Generating train split: 11948114 examples [03:29, 64461.32 examples/s]
Generating train split: 11956235 examples [03:30, 66027.97 examples/s]
Generating train split: 11971235 examples [03:30, 65410.84 examples/s]
Generating train split: 11985235 examples [03:30, 71120.72 examples/s]
Generating train split: 11997983 examples [03:30, 51648.59 examples/s]
Generating train split: 12013983 examples [03:31, 45022.01 examples/s]
Generating train split: 12028983 examples [03:31, 56303.85 examples/s]
Generating train split: 12050983 examples [03:31, 77751.81 examples/s]
Generating train split: 12066983 examples [03:32, 50459.42 examples/s]
Generating train split: 12083983 examples [03:32, 56058.19 examples/s]
Generating train split: 12100983 examples [03:32, 62097.78 examples/s]
Generating train split: 12117983 examples [03:32, 64492.01 examples/s]
Generating train split: 12127983 examples [03:33, 48295.72 examples/s]
Generating train split: 12148999 examples [03:33, 67624.75 examples/s]
Generating train split: 12164999 examples [03:33, 55845.46 examples/s]
Generating train split: 12173999 examples [03:33, 59128.66 examples/s]
Generating train split: 12191999 examples [03:34, 67288.29 examples/s]
Generating train split: 12209999 examples [03:34, 69086.63 examples/s]
Generating train split: 12236058 examples [03:34, 58473.12 examples/s]
Generating train split: 12271058 examples [03:35, 60921.73 examples/s]
Generating train split: 12285058 examples [03:35, 60957.49 examples/s]
Generating train split: 12304058 examples [03:35, 71288.90 examples/s]
Generating train split: 12321058 examples [03:35, 78396.14 examples/s]
Generating train split: 12337107 examples [03:36, 65126.40 examples/s]
Generating train split: 12354107 examples [03:36, 67091.24 examples/s]
Generating train split: 12365107 examples [03:36, 71844.49 examples/s]
Generating train split: 12392107 examples [03:36, 99436.73 examples/s]
Generating train split: 12413107 examples [03:37, 54987.85 examples/s]
Generating train split: 12430107 examples [03:37, 58497.18 examples/s]
Generating train split: 12439107 examples [03:37, 56666.44 examples/s]
Generating train split: 12448107 examples [03:38, 59468.76 examples/s]
Generating train split: 12458107 examples [03:38, 51665.62 examples/s]
Generating train split: 12486107 examples [03:38, 55372.87 examples/s]
Generating train split: 12501107 examples [03:38, 62554.78 examples/s]
Generating train split: 12522107 examples [03:39, 69677.89 examples/s]
Generating train split: 12543107 examples [03:39, 74498.47 examples/s]
Generating train split: 12551107 examples [03:39, 51212.44 examples/s]
Generating train split: 12561107 examples [03:40, 49094.33 examples/s]
Generating train split: 12570107 examples [03:40, 48314.70 examples/s]
Generating train split: 12588107 examples [03:40, 59546.54 examples/s]
Generating train split: 12596107 examples [03:40, 60835.04 examples/s]
Generating train split: 12606107 examples [03:40, 56920.95 examples/s]
Generating train split: 12614107 examples [03:40, 59601.99 examples/s]
Generating train split: 12641839 examples [03:41, 66576.28 examples/s]
Generating train split: 12668910 examples [03:41, 92345.30 examples/s]
Generating train split: 12693910 examples [03:41, 97331.96 examples/s]
Generating train split: 12705910 examples [03:41, 81519.12 examples/s]
Generating train split: 12715910 examples [03:42, 66755.08 examples/s]
Generating train split: 12733910 examples [03:42, 64502.32 examples/s]
Generating train split: 12742910 examples [03:42, 64756.87 examples/s]
Generating train split: 12758059 examples [03:42, 73315.67 examples/s]
Generating train split: 12771059 examples [03:42, 83209.57 examples/s]
Generating train split: 12787059 examples [03:43, 57730.61 examples/s]
Generating train split: 12796059 examples [03:43, 53158.94 examples/s]
Generating train split: 12816059 examples [03:43, 69042.56 examples/s]
Generating train split: 12829059 examples [03:43, 71750.42 examples/s]
Generating train split: 12848059 examples [03:44, 48947.78 examples/s]
Generating train split: 12876779 examples [03:44, 65856.52 examples/s]
Generating train split: 12886779 examples [03:44, 69011.54 examples/s]
Generating train split: 12912779 examples [03:45, 74321.20 examples/s]
Generating train split: 12927779 examples [03:45, 61225.50 examples/s]
Generating train split: 12935779 examples [03:45, 63322.04 examples/s]
Generating train split: 12943779 examples [03:45, 64415.97 examples/s]
Generating train split: 12952779 examples [03:45, 67768.07 examples/s]
Generating train split: 12964779 examples [03:46, 43168.69 examples/s]
Generating train split: 12976779 examples [03:46, 41642.18 examples/s]
Generating train split: 12999779 examples [03:46, 62841.19 examples/s]
Generating train split: 13013779 examples [03:46, 65282.83 examples/s]
Generating train split: 13028779 examples [03:47, 45715.12 examples/s]
Generating train split: 13046779 examples [03:47, 61125.86 examples/s]
Generating train split: 13072612 examples [03:47, 74654.39 examples/s]
Generating train split: 13083612 examples [03:48, 48164.86 examples/s]
Generating train split: 13102612 examples [03:48, 53906.96 examples/s]
Generating train split: 13110612 examples [03:48, 54423.72 examples/s]
Generating train split: 13127612 examples [03:48, 63468.23 examples/s]
Generating train split: 13144612 examples [03:49, 51272.63 examples/s]
Generating train split: 13160612 examples [03:49, 61863.02 examples/s]
Generating train split: 13169612 examples [03:49, 56302.04 examples/s]
Generating train split: 13183612 examples [03:49, 58982.18 examples/s]
Generating train split: 13202400 examples [03:50, 78081.80 examples/s]
Generating train split: 13217400 examples [03:50, 73088.06 examples/s]
Generating train split: 13235247 examples [03:50, 70604.59 examples/s]
Generating train split: 13250247 examples [03:50, 61855.33 examples/s]
Generating train split: 13262247 examples [03:51, 67228.73 examples/s]
Generating train split: 13303247 examples [03:51, 69389.01 examples/s]
Generating train split: 13319247 examples [03:51, 69319.91 examples/s]
Generating train split: 13334247 examples [03:52, 65709.28 examples/s]
Generating train split: 13353247 examples [03:52, 77924.88 examples/s]
Generating train split: 13366247 examples [03:52, 84888.73 examples/s]
Generating train split: 13395167 examples [03:52, 111256.20 examples/s]
Generating train split: 13411167 examples [03:52, 86226.50 examples/s]
Generating train split: 13422167 examples [03:52, 88083.04 examples/s]
Generating train split: 13448167 examples [03:53, 109141.81 examples/s]
Generating train split: 13477415 examples [03:53, 138466.73 examples/s]
Generating train split: 13512415 examples [03:53, 159299.92 examples/s]
Generating train split: 13538415 examples [03:53, 138044.72 examples/s]
Generating train split: 13563415 examples [03:53, 150846.32 examples/s]
Generating train split: 13600415 examples [03:53, 185184.31 examples/s]
Generating train split: 13631415 examples [03:54, 199599.69 examples/s]
Generating train split: 13669415 examples [03:54, 219066.86 examples/s]
Generating train split: 13715161 examples [03:54, 254963.53 examples/s]
Generating train split: 13743161 examples [03:54, 252727.95 examples/s]
Generating train split: 13770161 examples [03:54, 244871.38 examples/s]
Generating train split: 13799161 examples [03:54, 246346.44 examples/s]
Generating train split: 13836161 examples [03:54, 277750.90 examples/s]
Generating train split: 13888161 examples [03:54, 309066.59 examples/s]
Generating train split: 13921161 examples [03:55, 254598.86 examples/s]
Generating train split: 13981811 examples [03:55, 315487.24 examples/s]
Generating train split: 14019811 examples [03:55, 265749.14 examples/s]
Generating train split: 14056811 examples [03:55, 212978.99 examples/s]
Generating train split: 14081811 examples [03:55, 203609.94 examples/s]
Generating train split: 14114811 examples [03:55, 197735.83 examples/s]
Generating train split: 14144878 examples [03:56, 193605.46 examples/s]
Generating train split: 14165878 examples [03:56, 153751.88 examples/s]
Generating train split: 14214878 examples [03:56, 190571.26 examples/s]
Generating train split: 14274665 examples [03:56, 256871.29 examples/s]
Generating train split: 14334351 examples [03:56, 326119.81 examples/s]
Generating train split: 14376351 examples [03:56, 326806.42 examples/s]
Generating train split: 14427351 examples [03:57, 344343.47 examples/s]
Generating train split: 14496805 examples [03:57, 418667.66 examples/s]
Generating train split: 14547805 examples [03:57, 440678.30 examples/s]
Generating train split: 14601805 examples [03:57, 429860.75 examples/s]
Generating train split: 14650805 examples [03:57, 443186.22 examples/s]
Generating train split: 14699047 examples [03:57, 439373.31 examples/s]
Generating train split: 14747337 examples [03:57, 380305.16 examples/s]
Generating train split: 14790567 examples [03:57, 346318.50 examples/s]
Generating train split: 14834567 examples [03:58, 319714.44 examples/s]
Generating train split: 14868567 examples [03:58, 308513.93 examples/s]
Generating train split: 14903567 examples [03:58, 293339.78 examples/s]
Generating train split: 14948630 examples [03:58, 302649.77 examples/s]
Generating train split: 14978856 examples [03:59, 62577.15 examples/s]
+[titan] 2025-07-23 11:40:36,057 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 11:40:36,957 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 11:40:37,082 - root - INFO - Building dataloader...
+[titan] 2025-07-23 11:40:37,085 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 11:40:37,088 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 11:40:39,687 - fla.layers.mamba2 - WARNING - The fast path is not available because one of `(selective_state_update)` is None. Falling back to the naive implementation. To install follow https://github.com/state-spaces/mamba/#installation
+[titan] 2025-07-23 11:40:39,687 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 11:40:39,804 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 11:40:39,857 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 11:40:39,857 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 11:40:39,857 - root - WARNING - No norm found in model
+[titan] 2025-07-23 11:40:39,858 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 11:40:40,108 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 11:40:40,431 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 11:40:40,596 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 11:40:40,842 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 11:40:40,845 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 11:40:40,873 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+wandb: Network error (InvalidURL), entering retry loop.
+wandb: W&B API key is configured. Use `wandb login --relogin` to force relogin
+wandb: Network error (InvalidURL), entering retry loop.
+[titan] 2025-07-23 11:42:33,611 - root - ERROR - Failed to create WandB logger: Run initialization has timed out after 90.0 sec. Please try increasing the timeout with the `init_timeout` setting: `wandb.init(settings=wandb.Settings(init_timeout=120))`.
+[titan] 2025-07-23 11:42:33,728 - root - INFO - TensorBoard logging enabled. Logs will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1140
+[titan] 2025-07-23 11:42:33,729 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 11:42:33,774 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 11:42:43,730 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 11:42:43,732 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 11:42:43,732 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 11:42:43,732 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 11:42:43,732 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 11:42:43,733 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 11:42:43,733 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 11:42:43,733 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 11:42:43,733 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 11:42:43,733 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+Traceback (most recent call last):
+ File "", line 198, in _run_module_as_main
+ File "", line 88, in _run_code
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+ main(config)
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+ return f(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+ output = model(
+ ^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+ return self._call_impl(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+ return inner()
+ ^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+ result = forward_call(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+ return func(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+ outputs = self.backbone(
+ ^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+ return self._call_impl(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+ return forward_call(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+ hidden_states = mixer_block(
+ ^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+ return self._call_impl(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+ return inner()
+ ^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+ result = forward_call(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+ return fn(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+ return self._call_impl(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+ return forward_call(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+ hidden_states = self.norm(hidden_states)
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+ hidden_states = self.mixer(
+ ^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+ return self._call_impl(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+ return forward_call(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 601, in forward
+ return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 528, in torch_forward
+ G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]
+ ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~
+torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 0 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003896 has 316.00 MiB memory in use. Process 696027 has 316.00 MiB memory in use. Process 1850004 has 21.15 GiB memory in use. Process 2711975 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
+[rank0]: Traceback (most recent call last):
+[rank0]: File "", line 198, in _run_module_as_main
+[rank0]: File "", line 88, in _run_code
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank0]: main(config)
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank0]: return f(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank0]: output = model(
+[rank0]: ^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank0]: return self._call_impl(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank0]: return inner()
+[rank0]: ^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank0]: result = forward_call(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank0]: return func(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank0]: outputs = self.backbone(
+[rank0]: ^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank0]: return self._call_impl(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank0]: return forward_call(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank0]: hidden_states = mixer_block(
+[rank0]: ^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank0]: return self._call_impl(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank0]: return inner()
+[rank0]: ^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank0]: result = forward_call(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank0]: return fn(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank0]: return self._call_impl(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank0]: return forward_call(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank0]: hidden_states = self.norm(hidden_states)
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank0]: hidden_states = self.mixer(
+[rank0]: ^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank0]: return self._call_impl(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank0]: return forward_call(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 601, in forward
+[rank0]: return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 528, in torch_forward
+[rank0]: G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]
+[rank0]: ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~
+[rank0]: torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 0 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003896 has 316.00 MiB memory in use. Process 696027 has 316.00 MiB memory in use. Process 1850004 has 21.15 GiB memory in use. Process 2711975 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/0/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/0/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/1/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/1/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..638000fbb0b442a886bd907cd0fb625f533d7755
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/1/error.json
@@ -0,0 +1 @@
+{"message": {"message": "OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 1 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003901 has 316.00 MiB memory in use. Process 696029 has 316.00 MiB memory in use. Process 1850005 has 21.15 GiB memory in use. Process 2711976 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 601, in forward\n return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 528, in torch_forward\n G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]\n ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~\ntorch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 1 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003901 has 316.00 MiB memory in use. Process 696029 has 316.00 MiB memory in use. Process 1850005 has 21.15 GiB memory in use. Process 2711976 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)\n", "timestamp": "1753242219"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/1/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/1/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..304d0f6abe92231f1dc8ce9bd1789798328736d5
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/1/stderr.log
@@ -0,0 +1,387 @@
+[titan] 2025-07-23 11:13:05,108 - root - INFO - Starting job: default job
+[titan] 2025-07-23 11:13:05,108 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 11:13:05,108 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 11:13:07,346 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 11:13:07,348 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 11:13:07,420 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 11:13:07,421 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 11:13:07,421 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 11:13:07,438 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 11:13:07,712 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 11:13:07,712 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:07,712 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:08,426 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 11:13:08,426 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:13:08,426 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:08,427 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:39,939 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:39,974 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 11:35:39,974 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:35:39,974 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:39,974 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:07,574 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:08,012 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 11:36:08,012 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:36:08,013 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:08,013 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:40:36,226 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 11:40:36,926 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 11:40:37,051 - root - INFO - Building dataloader...
+[titan] 2025-07-23 11:40:37,053 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 11:40:37,055 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 11:40:39,687 - fla.layers.mamba2 - WARNING - The fast path is not available because one of `(selective_state_update)` is None. Falling back to the naive implementation. To install follow https://github.com/state-spaces/mamba/#installation
+[titan] 2025-07-23 11:40:39,688 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 11:40:39,804 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 11:40:39,856 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 11:40:39,856 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 11:40:39,857 - root - WARNING - No norm found in model
+[titan] 2025-07-23 11:40:39,857 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 11:40:40,092 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 11:40:40,433 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 11:40:40,596 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 11:40:40,838 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 11:40:40,840 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 11:40:40,867 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 11:40:40,868 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 11:40:41,114 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 11:40:48,994 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 11:40:48,995 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 11:40:49,029 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 11:40:49,030 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 11:40:49,031 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 11:40:49,031 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 11:40:49,037 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 11:40:49,037 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 11:40:49,037 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 11:40:49,038 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank1]: Traceback (most recent call last):
+[rank1]: File "", line 198, in _run_module_as_main
+[rank1]: File "", line 88, in _run_code
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank1]: main(config)
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank1]: return f(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank1]: output = model(
+[rank1]: ^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank1]: return self._call_impl(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank1]: return inner()
+[rank1]: ^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank1]: result = forward_call(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank1]: return func(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank1]: outputs = self.backbone(
+[rank1]: ^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank1]: return self._call_impl(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank1]: return forward_call(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank1]: hidden_states = mixer_block(
+[rank1]: ^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank1]: return self._call_impl(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank1]: return inner()
+[rank1]: ^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank1]: result = forward_call(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank1]: return fn(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank1]: return self._call_impl(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank1]: return forward_call(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank1]: hidden_states = self.norm(hidden_states)
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank1]: hidden_states = self.mixer(
+[rank1]: ^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank1]: return self._call_impl(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank1]: return forward_call(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 601, in forward
+[rank1]: return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 528, in torch_forward
+[rank1]: G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]
+[rank1]: ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~
+[rank1]: torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 1 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003901 has 316.00 MiB memory in use. Process 696029 has 316.00 MiB memory in use. Process 1850005 has 21.15 GiB memory in use. Process 2711976 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/1/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/1/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/2/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/2/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..0b596cfbb56f476e3b35f6217e2b1db4207e6d51
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/2/error.json
@@ -0,0 +1 @@
+{"message": {"message": "OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 2 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003900 has 316.00 MiB memory in use. Process 696030 has 316.00 MiB memory in use. Process 1850006 has 21.15 GiB memory in use. Process 2711977 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 601, in forward\n return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 528, in torch_forward\n G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]\n ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~\ntorch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 2 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003900 has 316.00 MiB memory in use. Process 696030 has 316.00 MiB memory in use. Process 1850006 has 21.15 GiB memory in use. Process 2711977 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)\n", "timestamp": "1753242218"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/2/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/2/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..202703884b85a375c0d4c08c56effcf07e0650de
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/2/stderr.log
@@ -0,0 +1,387 @@
+[titan] 2025-07-23 11:13:04,960 - root - INFO - Starting job: default job
+[titan] 2025-07-23 11:13:04,960 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 11:13:04,961 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 11:13:07,258 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 11:13:07,260 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 11:13:07,304 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 11:13:07,304 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 11:13:07,304 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 11:13:07,361 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 11:13:07,703 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 11:13:07,704 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:07,705 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:08,425 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 11:13:08,426 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:13:08,426 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:08,426 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:38,900 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:38,934 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 11:35:38,934 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:35:38,934 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:38,934 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:07,570 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:08,012 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 11:36:08,012 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:36:08,013 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:08,013 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:40:36,210 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 11:40:37,019 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 11:40:37,242 - root - INFO - Building dataloader...
+[titan] 2025-07-23 11:40:37,245 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 11:40:37,247 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 11:40:39,687 - fla.layers.mamba2 - WARNING - The fast path is not available because one of `(selective_state_update)` is None. Falling back to the naive implementation. To install follow https://github.com/state-spaces/mamba/#installation
+[titan] 2025-07-23 11:40:39,688 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 11:40:39,806 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 11:40:39,861 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 11:40:39,861 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 11:40:39,862 - root - WARNING - No norm found in model
+[titan] 2025-07-23 11:40:39,862 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 11:40:40,076 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 11:40:40,433 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 11:40:40,596 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 11:40:40,843 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 11:40:40,846 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 11:40:40,874 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 11:40:40,874 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 11:40:41,116 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 11:40:51,348 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 11:40:51,350 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 11:40:51,350 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 11:40:51,350 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 11:40:51,351 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 11:40:51,351 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 11:40:51,351 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 11:40:51,351 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 11:40:51,351 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 11:40:51,352 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank2]: Traceback (most recent call last):
+[rank2]: File "", line 198, in _run_module_as_main
+[rank2]: File "", line 88, in _run_code
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank2]: main(config)
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank2]: return f(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank2]: output = model(
+[rank2]: ^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank2]: return self._call_impl(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank2]: return inner()
+[rank2]: ^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank2]: result = forward_call(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank2]: return func(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank2]: outputs = self.backbone(
+[rank2]: ^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank2]: return self._call_impl(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank2]: return forward_call(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank2]: hidden_states = mixer_block(
+[rank2]: ^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank2]: return self._call_impl(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank2]: return inner()
+[rank2]: ^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank2]: result = forward_call(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank2]: return fn(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank2]: return self._call_impl(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank2]: return forward_call(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank2]: hidden_states = self.norm(hidden_states)
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank2]: hidden_states = self.mixer(
+[rank2]: ^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank2]: return self._call_impl(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank2]: return forward_call(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 601, in forward
+[rank2]: return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 528, in torch_forward
+[rank2]: G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]
+[rank2]: ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~
+[rank2]: torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 2 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003900 has 316.00 MiB memory in use. Process 696030 has 316.00 MiB memory in use. Process 1850006 has 21.15 GiB memory in use. Process 2711977 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/2/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/2/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/3/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/3/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..32dc2db2fe23b59bdc9c17331b5c63c7fd259d8a
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/3/error.json
@@ -0,0 +1 @@
+{"message": {"message": "OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 3 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003902 has 316.00 MiB memory in use. Process 696032 has 316.00 MiB memory in use. Process 1850007 has 21.15 GiB memory in use. Process 2711978 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 601, in forward\n return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 528, in torch_forward\n G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]\n ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~\ntorch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 3 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003902 has 316.00 MiB memory in use. Process 696032 has 316.00 MiB memory in use. Process 1850007 has 21.15 GiB memory in use. Process 2711978 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)\n", "timestamp": "1753242219"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/3/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/3/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..98273274ffe0f271fd88b61c1ab643c2d00b1c26
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/3/stderr.log
@@ -0,0 +1,388 @@
+[titan] 2025-07-23 11:13:05,081 - root - INFO - Starting job: default job
+[titan] 2025-07-23 11:13:05,082 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 11:13:05,082 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 11:13:07,378 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 11:13:07,380 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 11:13:07,458 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 11:13:07,458 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 11:13:07,458 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 11:13:07,476 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 11:13:07,703 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 11:13:07,704 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:07,704 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:08,425 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 11:13:08,426 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:13:08,426 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:08,427 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+
Generating train split: 0 examples [00:00, ? examples/s]
Generating train split: 1000 examples [00:03, 325.16 examples/s]
Generating train split: 22000 examples [00:03, 9631.42 examples/s]
Generating train split: 35000 examples [00:03, 15828.52 examples/s]
Generating train split: 49000 examples [00:03, 22892.88 examples/s]
Generating train split: 98000 examples [00:03, 63832.48 examples/s]
Generating train split: 126000 examples [00:03, 86583.41 examples/s]
Generating train split: 149000 examples [00:03, 102641.38 examples/s]
Generating train split: 188000 examples [00:04, 147944.16 examples/s]
Generating train split: 219000 examples [00:07, 22588.37 examples/s]
Generating train split: 284000 examples [00:07, 42681.22 examples/s]
Generating train split: 318000 examples [00:08, 54194.72 examples/s]
Generating train split: 347000 examples [00:08, 58956.17 examples/s]
Generating train split: 390000 examples [00:08, 82705.69 examples/s]
Generating train split: 448000 examples [00:08, 124212.44 examples/s]
Generating train split: 487000 examples [00:08, 152242.68 examples/s]
Generating train split: 523000 examples [00:09, 163915.53 examples/s]
Generating train split: 558000 examples [00:12, 33846.31 examples/s]
Generating train split: 580000 examples [00:12, 37479.91 examples/s]
Generating train split: 601000 examples [00:12, 43312.87 examples/s]
Generating train split: 649000 examples [00:12, 67294.95 examples/s]
Generating train split: 690000 examples [00:13, 93023.12 examples/s]
Generating train split: 720000 examples [00:13, 110469.35 examples/s]
Generating train split: 749000 examples [00:13, 110153.76 examples/s]
Generating train split: 778000 examples [00:13, 131580.56 examples/s]
Generating train split: 814000 examples [00:13, 155726.06 examples/s]
Generating train split: 847000 examples [00:13, 180094.75 examples/s]
Generating train split: 874000 examples [00:13, 192551.14 examples/s]
Generating train split: 903000 examples [00:17, 28599.42 examples/s]
Generating train split: 940000 examples [00:17, 41277.89 examples/s]
Generating train split: 962000 examples [00:17, 46492.68 examples/s]
Generating train split: 984000 examples [00:17, 55393.33 examples/s]
Generating train split: 1040000 examples [00:17, 95113.69 examples/s]
Generating train split: 1071000 examples [00:17, 115883.69 examples/s]
Generating train split: 1121000 examples [00:17, 163794.81 examples/s]
Generating train split: 1159000 examples [00:18, 164933.60 examples/s]
Generating train split: 1193000 examples [00:18, 187221.40 examples/s]
Generating train split: 1222000 examples [00:18, 160502.79 examples/s]
Generating train split: 1249000 examples [00:18, 156973.30 examples/s]
Generating train split: 1271000 examples [00:21, 30631.46 examples/s]
Generating train split: 1287000 examples [00:21, 32534.74 examples/s]
Generating train split: 1320000 examples [00:21, 47954.42 examples/s]
Generating train split: 1344000 examples [00:22, 61118.15 examples/s]
Generating train split: 1364000 examples [00:22, 69129.99 examples/s]
Generating train split: 1398000 examples [00:22, 96575.56 examples/s]
Generating train split: 1422000 examples [00:22, 99395.01 examples/s]
Generating train split: 1465000 examples [00:22, 144592.39 examples/s]
Generating train split: 1494000 examples [00:22, 155950.08 examples/s]
Generating train split: 1522000 examples [00:22, 168518.25 examples/s]
Generating train split: 1551000 examples [00:23, 167076.32 examples/s]
Generating train split: 1584000 examples [00:23, 146732.63 examples/s]
Generating train split: 1606000 examples [00:23, 135723.53 examples/s]
Generating train split: 1622000 examples [00:26, 25837.12 examples/s]
Generating train split: 1641000 examples [00:26, 32696.62 examples/s]
Generating train split: 1658000 examples [00:26, 40033.37 examples/s]
Generating train split: 1681000 examples [00:26, 53775.32 examples/s]
Generating train split: 1702000 examples [00:26, 67047.53 examples/s]
Generating train split: 1732000 examples [00:26, 91878.11 examples/s]
Generating train split: 1755000 examples [00:27, 102236.06 examples/s]
Generating train split: 1776000 examples [00:27, 118569.48 examples/s]
Generating train split: 1799000 examples [00:27, 120349.83 examples/s]
Generating train split: 1825000 examples [00:27, 145690.23 examples/s]
Generating train split: 1854000 examples [00:27, 160830.08 examples/s]
Generating train split: 1877000 examples [00:27, 172784.90 examples/s]
Generating train split: 1905000 examples [00:27, 176672.61 examples/s]
Generating train split: 1927000 examples [00:28, 149199.70 examples/s]
Generating train split: 1948000 examples [00:28, 157660.96 examples/s]
Generating train split: 1971000 examples [00:28, 96037.33 examples/s]
Generating train split: 1985000 examples [00:30, 24725.96 examples/s]
Generating train split: 1996000 examples [00:31, 26718.91 examples/s]
Generating train split: 2016000 examples [00:31, 36914.56 examples/s]
Generating train split: 2043000 examples [00:31, 53651.70 examples/s]
Generating train split: 2064000 examples [00:31, 68451.60 examples/s]
Generating train split: 2090000 examples [00:31, 91806.05 examples/s]
Generating train split: 2109000 examples [00:31, 101704.11 examples/s]
Generating train split: 2156000 examples [00:31, 128755.64 examples/s]
Generating train split: 2174000 examples [00:32, 125882.26 examples/s]
Generating train split: 2191000 examples [00:32, 128595.99 examples/s]
Generating train split: 2220000 examples [00:32, 150982.03 examples/s]
Generating train split: 2242000 examples [00:32, 134344.54 examples/s]
Generating train split: 2268000 examples [00:32, 153252.26 examples/s]
Generating train split: 2290000 examples [00:32, 154308.37 examples/s]
Generating train split: 2307000 examples [00:32, 148316.13 examples/s]
Generating train split: 2334000 examples [00:33, 93959.84 examples/s]
Generating train split: 2347000 examples [00:35, 25883.31 examples/s]
Generating train split: 2358000 examples [00:35, 29879.80 examples/s]
Generating train split: 2370000 examples [00:36, 29255.52 examples/s]
Generating train split: 2396000 examples [00:36, 45598.55 examples/s]
Generating train split: 2419000 examples [00:36, 61560.90 examples/s]
Generating train split: 2453000 examples [00:36, 89253.27 examples/s]
Generating train split: 2476000 examples [00:36, 103895.64 examples/s]
Generating train split: 2503000 examples [00:36, 126314.80 examples/s]
Generating train split: 2527000 examples [00:36, 118285.04 examples/s]
Generating train split: 2549000 examples [00:37, 129727.83 examples/s]
Generating train split: 2576000 examples [00:37, 152113.31 examples/s]
Generating train split: 2597000 examples [00:37, 161570.30 examples/s]
Generating train split: 2618000 examples [00:37, 103018.13 examples/s]
Generating train split: 2634000 examples [00:37, 103232.70 examples/s]
Generating train split: 2673000 examples [00:37, 143343.17 examples/s]
Generating train split: 2695000 examples [00:40, 28146.66 examples/s]
Generating train split: 2725000 examples [00:40, 40615.37 examples/s]
Generating train split: 2746000 examples [00:40, 42845.94 examples/s]
Generating train split: 2775000 examples [00:41, 58092.24 examples/s]
Generating train split: 2821000 examples [00:41, 91618.02 examples/s]
Generating train split: 2849000 examples [00:41, 102251.78 examples/s]
Generating train split: 2872000 examples [00:41, 117735.61 examples/s]
Generating train split: 2918000 examples [00:41, 163580.59 examples/s]
Generating train split: 2945000 examples [00:41, 141659.95 examples/s]
Generating train split: 2970000 examples [00:42, 130234.57 examples/s]
Generating train split: 2997000 examples [00:42, 112650.36 examples/s]
Generating train split: 3029000 examples [00:42, 141038.68 examples/s]
Generating train split: 3050000 examples [00:44, 33226.30 examples/s]
Generating train split: 3065000 examples [00:45, 32424.10 examples/s]
Generating train split: 3093000 examples [00:45, 45831.81 examples/s]
Generating train split: 3116000 examples [00:45, 57440.38 examples/s]
Generating train split: 3137000 examples [00:45, 61460.77 examples/s]
Generating train split: 3188000 examples [00:46, 104640.67 examples/s]
Generating train split: 3210000 examples [00:46, 117051.39 examples/s]
Generating train split: 3232000 examples [00:46, 111257.11 examples/s]
Generating train split: 3260000 examples [00:46, 132830.90 examples/s]
Generating train split: 3288000 examples [00:46, 126298.57 examples/s]
Generating train split: 3305000 examples [00:46, 132883.49 examples/s]
Generating train split: 3336000 examples [00:46, 161729.22 examples/s]
Generating train split: 3358000 examples [00:47, 166494.16 examples/s]
Generating train split: 3380000 examples [00:47, 99311.42 examples/s]
Generating train split: 3397000 examples [00:49, 28741.65 examples/s]
Generating train split: 3408000 examples [00:49, 31577.20 examples/s]
Generating train split: 3419000 examples [00:50, 29575.09 examples/s]
Generating train split: 3429000 examples [00:50, 34123.96 examples/s]
Generating train split: 3462000 examples [00:50, 61045.57 examples/s]
Generating train split: 3496000 examples [00:50, 87918.98 examples/s]
Generating train split: 3517000 examples [00:50, 93373.06 examples/s]
Generating train split: 3559000 examples [00:50, 139971.93 examples/s]
Generating train split: 3586000 examples [00:51, 132363.14 examples/s]
Generating train split: 3606000 examples [00:51, 124685.83 examples/s]
Generating train split: 3629000 examples [00:51, 137278.01 examples/s]
Generating train split: 3662000 examples [00:51, 123415.69 examples/s]
Generating train split: 3693000 examples [00:51, 151442.42 examples/s]
Generating train split: 3728000 examples [00:51, 175190.92 examples/s]
Generating train split: 3751000 examples [00:52, 84817.66 examples/s]
Generating train split: 3767000 examples [00:54, 32139.05 examples/s]
Generating train split: 3779000 examples [00:54, 36336.61 examples/s]
Generating train split: 3796000 examples [00:54, 44874.73 examples/s]
Generating train split: 3812000 examples [00:55, 44417.54 examples/s]
Generating train split: 3823000 examples [00:55, 48510.43 examples/s]
Generating train split: 3874000 examples [00:55, 100568.96 examples/s]
Generating train split: 3907000 examples [00:55, 128801.65 examples/s]
Generating train split: 3932000 examples [00:55, 123173.77 examples/s]
Generating train split: 3965000 examples [00:55, 151577.10 examples/s]
Generating train split: 3988000 examples [00:55, 129441.97 examples/s]
Generating train split: 4011000 examples [00:56, 131851.89 examples/s]
Generating train split: 4033000 examples [00:56, 115040.68 examples/s]
Generating train split: 4053000 examples [00:56, 93534.87 examples/s]
Generating train split: 4071000 examples [00:56, 101024.05 examples/s]
Generating train split: 4096000 examples [00:57, 115178.57 examples/s]
Generating train split: 4112000 examples [00:58, 29677.38 examples/s]
Generating train split: 4123000 examples [00:59, 29506.16 examples/s]
Generating train split: 4140000 examples [00:59, 37728.52 examples/s]
Generating train split: 4156000 examples [00:59, 46290.37 examples/s]
Generating train split: 4167000 examples [00:59, 50573.19 examples/s]
Generating train split: 4189000 examples [00:59, 61060.44 examples/s]
Generating train split: 4215000 examples [01:00, 86827.36 examples/s]
Generating train split: 4241000 examples [01:00, 112926.22 examples/s]
Generating train split: 4275000 examples [01:00, 154206.41 examples/s]
Generating train split: 4302000 examples [01:00, 128899.43 examples/s]
Generating train split: 4338000 examples [01:00, 156054.56 examples/s]
Generating train split: 4360000 examples [01:00, 160902.87 examples/s]
Generating train split: 4381000 examples [01:00, 165991.01 examples/s]
Generating train split: 4401000 examples [01:01, 125213.25 examples/s]
Generating train split: 4417000 examples [01:01, 81543.42 examples/s]
Generating train split: 4439000 examples [01:01, 101043.54 examples/s]
Generating train split: 4455000 examples [01:03, 27606.03 examples/s]
Generating train split: 4471000 examples [01:03, 34174.80 examples/s]
Generating train split: 4486000 examples [01:04, 34677.71 examples/s]
Generating train split: 4498000 examples [01:04, 38127.40 examples/s]
Generating train split: 4537000 examples [01:04, 69954.41 examples/s]
Generating train split: 4559000 examples [01:04, 86089.59 examples/s]
Generating train split: 4580000 examples [01:04, 88323.56 examples/s]
Generating train split: 4636000 examples [01:04, 153831.21 examples/s]
Generating train split: 4664000 examples [01:05, 158760.84 examples/s]
Generating train split: 4690000 examples [01:05, 118714.10 examples/s]
Generating train split: 4723000 examples [01:05, 127327.40 examples/s]
Generating train split: 4745000 examples [01:05, 111515.31 examples/s]
Generating train split: 4762000 examples [01:06, 119971.98 examples/s]
Generating train split: 4778000 examples [01:06, 79159.80 examples/s]
Generating train split: 4795000 examples [01:06, 78182.80 examples/s]
Generating train split: 4806000 examples [01:08, 30022.02 examples/s]
Generating train split: 4816000 examples [01:08, 30250.52 examples/s]
Generating train split: 4830000 examples [01:08, 35246.45 examples/s]
Generating train split: 4847000 examples [01:08, 47595.65 examples/s]
Generating train split: 4859000 examples [01:09, 45645.36 examples/s]
Generating train split: 4868000 examples [01:09, 43168.08 examples/s]
Generating train split: 4885000 examples [01:09, 58871.99 examples/s]
Generating train split: 4906000 examples [01:09, 78655.48 examples/s]
Generating train split: 4921000 examples [01:09, 82964.07 examples/s]
Generating train split: 4975000 examples [01:09, 164930.09 examples/s]
Generating train split: 5012000 examples [01:09, 191455.07 examples/s]
Generating train split: 5041000 examples [01:10, 152361.29 examples/s]
Generating train split: 5063000 examples [01:10, 162485.42 examples/s]
Generating train split: 5087000 examples [01:10, 154122.03 examples/s]
Generating train split: 5108000 examples [01:10, 111880.83 examples/s]
Generating train split: 5125000 examples [01:11, 110899.25 examples/s]
Generating train split: 5142000 examples [01:11, 87012.80 examples/s]
Generating train split: 5153000 examples [01:12, 29938.32 examples/s]
Generating train split: 5164000 examples [01:12, 33964.65 examples/s]
Generating train split: 5175000 examples [01:13, 27941.73 examples/s]
Generating train split: 5192000 examples [01:13, 38310.85 examples/s]
Generating train split: 5202000 examples [01:13, 41578.48 examples/s]
Generating train split: 5218000 examples [01:13, 54558.59 examples/s]
Generating train split: 5233000 examples [01:14, 65344.51 examples/s]
Generating train split: 5261000 examples [01:14, 99472.90 examples/s]
Generating train split: 5285000 examples [01:14, 115091.55 examples/s]
Generating train split: 5306000 examples [01:14, 117290.21 examples/s]
Generating train split: 5321000 examples [01:14, 118706.01 examples/s]
Generating train split: 5351000 examples [01:14, 155024.23 examples/s]
Generating train split: 5390000 examples [01:14, 205550.80 examples/s]
Generating train split: 5418000 examples [01:15, 129293.14 examples/s]
Generating train split: 5440000 examples [01:15, 91422.32 examples/s]
Generating train split: 5457000 examples [01:15, 95827.09 examples/s]
Generating train split: 5472000 examples [01:16, 70635.99 examples/s]
Generating train split: 5483000 examples [01:17, 35546.94 examples/s]
Generating train split: 5493000 examples [01:17, 32111.00 examples/s]
Generating train split: 5513000 examples [01:18, 33175.91 examples/s]
Generating train split: 5523000 examples [01:18, 36413.90 examples/s]
Generating train split: 5539000 examples [01:18, 48071.85 examples/s]
Generating train split: 5556000 examples [01:18, 61456.20 examples/s]
Generating train split: 5572000 examples [01:18, 70949.32 examples/s]
Generating train split: 5606000 examples [01:18, 112890.11 examples/s]
Generating train split: 5627000 examples [01:19, 121366.27 examples/s]
Generating train split: 5670000 examples [01:19, 176010.96 examples/s]
Generating train split: 5693000 examples [01:19, 123770.29 examples/s]
Generating train split: 5715000 examples [01:19, 138766.32 examples/s]
Generating train split: 5735000 examples [01:19, 126121.65 examples/s]
Generating train split: 5757000 examples [01:19, 130050.65 examples/s]
Generating train split: 5773000 examples [01:20, 96434.87 examples/s]
Generating train split: 5790000 examples [01:20, 103014.33 examples/s]
Generating train split: 5807000 examples [01:20, 92178.68 examples/s]
Generating train split: 5823000 examples [01:20, 101239.66 examples/s]
Generating train split: 5838000 examples [01:22, 32920.25 examples/s]
Generating train split: 5853000 examples [01:22, 40654.83 examples/s]
Generating train split: 5863000 examples [01:22, 30285.96 examples/s]
Generating train split: 5872000 examples [01:22, 34067.17 examples/s]
Generating train split: 5889000 examples [01:23, 45741.80 examples/s]
Generating train split: 5909000 examples [01:23, 60223.96 examples/s]
Generating train split: 5938000 examples [01:23, 77729.92 examples/s]
Generating train split: 5950000 examples [01:23, 82220.54 examples/s]
Generating train split: 5979000 examples [01:23, 111240.40 examples/s]
Generating train split: 5995000 examples [01:23, 109359.85 examples/s]
Generating train split: 6027000 examples [01:24, 144605.89 examples/s]
Generating train split: 6055000 examples [01:24, 172485.87 examples/s]
Generating train split: 6078000 examples [01:24, 125145.27 examples/s]
Generating train split: 6097000 examples [01:24, 125570.69 examples/s]
Generating train split: 6113000 examples [01:24, 95612.96 examples/s]
Generating train split: 6136000 examples [01:25, 106461.94 examples/s]
Generating train split: 6152000 examples [01:25, 102488.91 examples/s]
Generating train split: 6168000 examples [01:26, 41841.26 examples/s]
Generating train split: 6179000 examples [01:26, 33411.00 examples/s]
Generating train split: 6189000 examples [01:27, 37314.06 examples/s]
Generating train split: 6199000 examples [01:27, 42008.41 examples/s]
Generating train split: 6209000 examples [01:27, 34300.94 examples/s]
Generating train split: 6219000 examples [01:27, 36389.34 examples/s]
Generating train split: 6229000 examples [01:27, 42750.40 examples/s]
Generating train split: 6260000 examples [01:28, 68636.06 examples/s]
Generating train split: 6277000 examples [01:28, 81954.79 examples/s]
Generating train split: 6303000 examples [01:28, 103174.29 examples/s]
Generating train split: 6326000 examples [01:28, 118658.43 examples/s]
Generating train split: 6349000 examples [01:28, 137147.00 examples/s]
Generating train split: 6371000 examples [01:28, 130837.25 examples/s]
Generating train split: 6412000 examples [01:29, 173968.33 examples/s]
Generating train split: 6434000 examples [01:29, 112285.30 examples/s]
Generating train split: 6449000 examples [01:29, 115879.66 examples/s]
Generating train split: 6466000 examples [01:29, 111398.48 examples/s]
Generating train split: 6481000 examples [01:29, 109752.82 examples/s]
Generating train split: 6496000 examples [01:30, 95728.68 examples/s]
Generating train split: 6513000 examples [01:31, 39323.04 examples/s]
Generating train split: 6523000 examples [01:31, 44453.09 examples/s]
Generating train split: 6538000 examples [01:31, 53731.60 examples/s]
Generating train split: 6548000 examples [01:32, 33564.62 examples/s]
Generating train split: 6568000 examples [01:32, 40346.14 examples/s]
Generating train split: 6585000 examples [01:32, 53126.80 examples/s]
Generating train split: 6605000 examples [01:32, 71064.22 examples/s]
Generating train split: 6620000 examples [01:32, 69898.97 examples/s]
Generating train split: 6641000 examples [01:32, 91218.12 examples/s]
Generating train split: 6661000 examples [01:33, 109673.87 examples/s]
Generating train split: 6678000 examples [01:33, 114023.38 examples/s]
Generating train split: 6722000 examples [01:33, 180425.00 examples/s]
Generating train split: 6749000 examples [01:33, 156198.31 examples/s]
Generating train split: 6770000 examples [01:33, 127231.03 examples/s]
Generating train split: 6791000 examples [01:34, 111812.25 examples/s]
Generating train split: 6806000 examples [01:34, 108672.62 examples/s]
Generating train split: 6823000 examples [01:34, 112151.91 examples/s]
Generating train split: 6845000 examples [01:34, 125647.37 examples/s]
Generating train split: 6863000 examples [01:34, 117402.95 examples/s]
Generating train split: 6876000 examples [01:36, 31704.44 examples/s]
Generating train split: 6897000 examples [01:36, 43243.78 examples/s]
Generating train split: 6908000 examples [01:36, 31190.23 examples/s]
Generating train split: 6929000 examples [01:37, 43028.01 examples/s]
Generating train split: 6944000 examples [01:37, 45416.79 examples/s]
Generating train split: 6972000 examples [01:37, 67446.64 examples/s]
Generating train split: 6994000 examples [01:37, 85476.71 examples/s]
Generating train split: 7017000 examples [01:37, 102267.34 examples/s]
Generating train split: 7052000 examples [01:37, 143209.92 examples/s]
Generating train split: 7077000 examples [01:38, 120607.88 examples/s]
Generating train split: 7098000 examples [01:38, 130984.11 examples/s]
Generating train split: 7116000 examples [01:38, 136601.07 examples/s]
Generating train split: 7135000 examples [01:38, 111183.57 examples/s]
Generating train split: 7154000 examples [01:38, 124359.32 examples/s]
Generating train split: 7173000 examples [01:39, 87739.09 examples/s]
Generating train split: 7189000 examples [01:39, 95793.40 examples/s]
Generating train split: 7204000 examples [01:39, 73442.92 examples/s]
Generating train split: 7216000 examples [01:40, 54546.23 examples/s]
Generating train split: 7227000 examples [01:40, 46657.49 examples/s]
Generating train split: 7241000 examples [01:40, 55473.54 examples/s]
Generating train split: 7252000 examples [01:40, 42546.47 examples/s]
Generating train split: 7263000 examples [01:41, 30011.44 examples/s]
Generating train split: 7273000 examples [01:41, 32341.43 examples/s]
Generating train split: 7284000 examples [01:42, 34097.31 examples/s]
Generating train split: 7294000 examples [01:42, 39963.64 examples/s]
Generating train split: 7304000 examples [01:42, 47856.07 examples/s]
Generating train split: 7320000 examples [01:42, 65512.86 examples/s]
Generating train split: 7351000 examples [01:42, 99985.15 examples/s]
Generating train split: 7384000 examples [01:42, 143898.81 examples/s]
Generating train split: 7406000 examples [01:42, 130064.04 examples/s]
Generating train split: 7437000 examples [01:43, 165870.18 examples/s]
Generating train split: 7462000 examples [01:43, 161797.63 examples/s]
Generating train split: 7497000 examples [01:43, 197442.13 examples/s]
Generating train split: 7523000 examples [01:43, 152136.33 examples/s]
Generating train split: 7543000 examples [01:44, 100803.02 examples/s]
Generating train split: 7566000 examples [01:44, 94581.82 examples/s]
Generating train split: 7582000 examples [01:44, 101993.83 examples/s]
Generating train split: 7598000 examples [01:45, 51220.52 examples/s]
Generating train split: 7609000 examples [01:45, 55870.69 examples/s]
Generating train split: 7620000 examples [01:46, 36163.84 examples/s]
Generating train split: 7631000 examples [01:46, 41830.49 examples/s]
Generating train split: 7641000 examples [01:46, 48079.47 examples/s]
Generating train split: 7652000 examples [01:46, 52063.37 examples/s]
Generating train split: 7669000 examples [01:46, 59380.70 examples/s]
Generating train split: 7687000 examples [01:46, 78010.24 examples/s]
Generating train split: 7698000 examples [01:46, 74961.43 examples/s]
Generating train split: 7709000 examples [01:47, 78833.24 examples/s]
Generating train split: 7720000 examples [01:47, 72057.17 examples/s]
Generating train split: 7736000 examples [01:47, 86023.18 examples/s]
Generating train split: 7755000 examples [01:47, 105887.58 examples/s]
Generating train split: 7789000 examples [01:47, 110332.69 examples/s]
Generating train split: 7813000 examples [01:47, 133664.44 examples/s]
Generating train split: 7843000 examples [01:47, 160144.96 examples/s]
Generating train split: 7865415 examples [01:48, 166777.73 examples/s]
Generating train split: 7885415 examples [01:48, 117732.76 examples/s]
Generating train split: 7905415 examples [01:48, 87885.70 examples/s]
Generating train split: 7917415 examples [01:49, 67541.40 examples/s]
Generating train split: 7928415 examples [01:49, 73112.94 examples/s]
Generating train split: 7943415 examples [01:50, 41321.73 examples/s]
Generating train split: 7954415 examples [01:50, 48111.31 examples/s]
Generating train split: 7965415 examples [01:50, 47227.70 examples/s]
Generating train split: 7976415 examples [01:50, 39699.08 examples/s]
Generating train split: 7987415 examples [01:51, 33037.63 examples/s]
Generating train split: 8007415 examples [01:51, 43842.08 examples/s]
Generating train split: 8030415 examples [01:51, 64062.29 examples/s]
Generating train split: 8054415 examples [01:51, 84169.10 examples/s]
Generating train split: 8087415 examples [01:51, 98894.19 examples/s]
Generating train split: 8104415 examples [01:52, 108326.28 examples/s]
Generating train split: 8120415 examples [01:52, 109834.11 examples/s]
Generating train split: 8137415 examples [01:52, 111834.91 examples/s]
Generating train split: 8153415 examples [01:52, 101655.05 examples/s]
Generating train split: 8165415 examples [01:52, 100517.86 examples/s]
Generating train split: 8181415 examples [01:52, 111941.92 examples/s]
Generating train split: 8201415 examples [01:52, 131426.96 examples/s]
Generating train split: 8216415 examples [01:53, 93816.37 examples/s]
Generating train split: 8232415 examples [01:53, 98215.68 examples/s]
Generating train split: 8260995 examples [01:53, 123158.21 examples/s]
Generating train split: 8277995 examples [01:53, 112388.21 examples/s]
Generating train split: 8294995 examples [01:54, 51971.08 examples/s]
Generating train split: 8305995 examples [01:54, 46556.30 examples/s]
Generating train split: 8321995 examples [01:54, 56545.57 examples/s]
Generating train split: 8333995 examples [01:55, 64463.73 examples/s]
Generating train split: 8343995 examples [01:55, 64818.28 examples/s]
Generating train split: 8353995 examples [01:56, 28392.31 examples/s]
Generating train split: 8370995 examples [01:56, 38537.66 examples/s]
Generating train split: 8387995 examples [01:56, 51661.76 examples/s]
Generating train split: 8425995 examples [01:56, 92797.14 examples/s]
Generating train split: 8449995 examples [01:56, 114091.44 examples/s]
Generating train split: 8470995 examples [01:56, 104208.53 examples/s]
Generating train split: 8498995 examples [01:57, 125719.54 examples/s]
Generating train split: 8516995 examples [01:57, 122633.61 examples/s]
Generating train split: 8537995 examples [01:57, 113943.70 examples/s]
Generating train split: 8558995 examples [01:57, 102945.27 examples/s]
Generating train split: 8587995 examples [01:57, 126839.07 examples/s]
Generating train split: 8603995 examples [01:58, 81982.34 examples/s]
Generating train split: 8629995 examples [01:58, 104944.71 examples/s]
Generating train split: 8646995 examples [01:58, 104789.56 examples/s]
Generating train split: 8663995 examples [01:59, 49855.90 examples/s]
Generating train split: 8676995 examples [01:59, 45175.01 examples/s]
Generating train split: 8694995 examples [01:59, 56355.88 examples/s]
Generating train split: 8705995 examples [02:00, 53487.13 examples/s]
Generating train split: 8716995 examples [02:00, 49367.97 examples/s]
Generating train split: 8732995 examples [02:00, 60065.42 examples/s]
Generating train split: 8744995 examples [02:00, 51589.00 examples/s]
Generating train split: 8760995 examples [02:01, 53318.29 examples/s]
Generating train split: 8771995 examples [02:01, 54300.25 examples/s]
Generating train split: 8812995 examples [02:01, 102203.52 examples/s]
Generating train split: 8831995 examples [02:01, 88886.98 examples/s]
Generating train split: 8853995 examples [02:01, 103160.61 examples/s]
Generating train split: 8870995 examples [02:02, 105520.82 examples/s]
Generating train split: 8890995 examples [02:02, 92208.74 examples/s]
Generating train split: 8918995 examples [02:02, 115354.34 examples/s]
Generating train split: 8932995 examples [02:02, 117109.05 examples/s]
Generating train split: 8948995 examples [02:02, 95885.00 examples/s]
Generating train split: 8962995 examples [02:02, 102036.37 examples/s]
Generating train split: 8978995 examples [02:03, 81322.65 examples/s]
Generating train split: 8999995 examples [02:04, 45436.83 examples/s]
Generating train split: 9017995 examples [02:04, 57092.02 examples/s]
Generating train split: 9034995 examples [02:04, 44776.72 examples/s]
Generating train split: 9046995 examples [02:05, 43372.45 examples/s]
Generating train split: 9056995 examples [02:05, 41809.35 examples/s]
Generating train split: 9067995 examples [02:05, 49224.77 examples/s]
Generating train split: 9089995 examples [02:05, 56737.71 examples/s]
Generating train split: 9111995 examples [02:05, 77646.40 examples/s]
Generating train split: 9127995 examples [02:06, 63122.99 examples/s]
Generating train split: 9155995 examples [02:06, 89763.10 examples/s]
Generating train split: 9170995 examples [02:06, 90531.67 examples/s]
Generating train split: 9189995 examples [02:06, 96934.00 examples/s]
Generating train split: 9205995 examples [02:06, 99740.82 examples/s]
Generating train split: 9228995 examples [02:07, 124137.37 examples/s]
Generating train split: 9245995 examples [02:07, 116899.80 examples/s]
Generating train split: 9261995 examples [02:07, 118332.76 examples/s]
Generating train split: 9293995 examples [02:07, 147111.85 examples/s]
Generating train split: 9319995 examples [02:07, 146566.13 examples/s]
Generating train split: 9341995 examples [02:07, 157958.35 examples/s]
Generating train split: 9359995 examples [02:08, 73055.48 examples/s]
Generating train split: 9372995 examples [02:09, 42839.03 examples/s]
Generating train split: 9389995 examples [02:09, 52375.75 examples/s]
Generating train split: 9405995 examples [02:09, 53693.51 examples/s]
Generating train split: 9422995 examples [02:09, 61474.77 examples/s]
Generating train split: 9433995 examples [02:09, 66247.90 examples/s]
Generating train split: 9443995 examples [02:09, 69894.20 examples/s]
Generating train split: 9454995 examples [02:10, 55916.76 examples/s]
Generating train split: 9464995 examples [02:10, 52228.98 examples/s]
Generating train split: 9474995 examples [02:10, 57130.09 examples/s]
Generating train split: 9498995 examples [02:10, 88013.51 examples/s]
Generating train split: 9515995 examples [02:10, 99947.98 examples/s]
Generating train split: 9531995 examples [02:11, 73013.97 examples/s]
Generating train split: 9561995 examples [02:11, 103165.45 examples/s]
Generating train split: 9578995 examples [02:11, 110345.19 examples/s]
Generating train split: 9606995 examples [02:11, 130830.57 examples/s]
Generating train split: 9623995 examples [02:11, 107028.26 examples/s]
Generating train split: 9639995 examples [02:12, 79609.27 examples/s]
Generating train split: 9656995 examples [02:12, 86908.52 examples/s]
Generating train split: 9673995 examples [02:12, 97363.64 examples/s]
Generating train split: 9696995 examples [02:12, 110175.88 examples/s]
Generating train split: 9714995 examples [02:13, 65919.65 examples/s]
Generating train split: 9726995 examples [02:13, 40533.17 examples/s]
Generating train split: 9745995 examples [02:14, 53527.63 examples/s]
Generating train split: 9756995 examples [02:14, 43582.13 examples/s]
Generating train split: 9773995 examples [02:14, 55415.25 examples/s]
Generating train split: 9790995 examples [02:14, 67388.61 examples/s]
Generating train split: 9806995 examples [02:14, 69900.18 examples/s]
Generating train split: 9818995 examples [02:15, 73790.91 examples/s]
Generating train split: 9836995 examples [02:15, 84509.28 examples/s]
Generating train split: 9851995 examples [02:15, 75851.85 examples/s]
Generating train split: 9862995 examples [02:15, 74712.90 examples/s]
Generating train split: 9884995 examples [02:15, 93593.43 examples/s]
Generating train split: 9898995 examples [02:16, 75504.12 examples/s]
Generating train split: 9926995 examples [02:16, 107829.71 examples/s]
Generating train split: 9943995 examples [02:16, 107001.59 examples/s]
Generating train split: 9959995 examples [02:16, 105001.41 examples/s]
Generating train split: 9979995 examples [02:16, 118890.12 examples/s]
Generating train split: 10007995 examples [02:16, 142864.18 examples/s]
Generating train split: 10028995 examples [02:17, 110790.22 examples/s]
Generating train split: 10045995 examples [02:17, 85861.10 examples/s]
Generating train split: 10056995 examples [02:18, 50425.03 examples/s]
Generating train split: 10067995 examples [02:18, 49903.58 examples/s]
Generating train split: 10079995 examples [02:18, 53660.80 examples/s]
Generating train split: 10090995 examples [02:18, 41025.52 examples/s]
Generating train split: 10112995 examples [02:18, 60362.86 examples/s]
Generating train split: 10126995 examples [02:19, 50349.10 examples/s]
Generating train split: 10146995 examples [02:19, 68390.81 examples/s]
Generating train split: 10163995 examples [02:19, 80563.71 examples/s]
Generating train split: 10179995 examples [02:19, 70404.75 examples/s]
Generating train split: 10195995 examples [02:20, 77904.06 examples/s]
Generating train split: 10206995 examples [02:20, 66107.38 examples/s]
Generating train split: 10223995 examples [02:20, 79982.93 examples/s]
Generating train split: 10239995 examples [02:20, 90768.36 examples/s]
Generating train split: 10253995 examples [02:20, 100181.85 examples/s]
Generating train split: 10270995 examples [02:20, 106651.98 examples/s]
Generating train split: 10284995 examples [02:21, 93555.95 examples/s]
Generating train split: 10312995 examples [02:21, 131091.26 examples/s]
Generating train split: 10329995 examples [02:21, 110312.42 examples/s]
Generating train split: 10345995 examples [02:21, 111123.93 examples/s]
Generating train split: 10373995 examples [02:21, 132942.28 examples/s]
Generating train split: 10391995 examples [02:21, 130751.40 examples/s]
Generating train split: 10407995 examples [02:21, 117147.78 examples/s]
Generating train split: 10424995 examples [02:22, 88752.81 examples/s]
Generating train split: 10439995 examples [02:23, 38927.86 examples/s]
Generating train split: 10449659 examples [02:23, 43790.19 examples/s]
Generating train split: 10475659 examples [02:23, 52795.26 examples/s]
Generating train split: 10488659 examples [02:23, 59016.14 examples/s]
Generating train split: 10508659 examples [02:23, 75199.81 examples/s]
Generating train split: 10519659 examples [02:24, 60194.30 examples/s]
Generating train split: 10530659 examples [02:24, 59522.08 examples/s]
Generating train split: 10541659 examples [02:24, 46215.32 examples/s]
Generating train split: 10551659 examples [02:25, 50388.92 examples/s]
Generating train split: 10562659 examples [02:25, 58577.98 examples/s]
Generating train split: 10573659 examples [02:25, 60982.75 examples/s]
Generating train split: 10600659 examples [02:25, 86519.70 examples/s]
Generating train split: 10624659 examples [02:25, 107972.18 examples/s]
Generating train split: 10640659 examples [02:25, 77139.26 examples/s]
Generating train split: 10661659 examples [02:26, 96505.86 examples/s]
Generating train split: 10677659 examples [02:26, 95705.92 examples/s]
Generating train split: 10692659 examples [02:26, 97011.45 examples/s]
Generating train split: 10714659 examples [02:26, 115695.52 examples/s]
Generating train split: 10731659 examples [02:26, 125113.05 examples/s]
Generating train split: 10753659 examples [02:27, 69957.54 examples/s]
Generating train split: 10765659 examples [02:27, 70814.92 examples/s]
Generating train split: 10776659 examples [02:28, 36930.57 examples/s]
Generating train split: 10809659 examples [02:28, 62314.13 examples/s]
Generating train split: 10833659 examples [02:28, 66646.21 examples/s]
Generating train split: 10845659 examples [02:28, 66322.26 examples/s]
Generating train split: 10862659 examples [02:28, 79337.23 examples/s]
Generating train split: 10878659 examples [02:29, 69844.46 examples/s]
Generating train split: 10889659 examples [02:29, 64110.99 examples/s]
Generating train split: 10899659 examples [02:29, 48249.17 examples/s]
Generating train split: 10921659 examples [02:30, 61001.56 examples/s]
Generating train split: 10944659 examples [02:30, 83911.03 examples/s]
Generating train split: 10956659 examples [02:30, 81967.29 examples/s]
Generating train split: 10984659 examples [02:30, 113226.97 examples/s]
Generating train split: 11012659 examples [02:30, 143068.55 examples/s]
Generating train split: 11034659 examples [02:30, 125876.10 examples/s]
Generating train split: 11050659 examples [02:31, 87402.73 examples/s]
Generating train split: 11066659 examples [02:31, 77445.73 examples/s]
Generating train split: 11089659 examples [02:31, 96777.35 examples/s]
Generating train split: 11105659 examples [02:32, 54403.89 examples/s]
Generating train split: 11116659 examples [02:32, 58591.77 examples/s]
Generating train split: 11127659 examples [02:32, 48223.41 examples/s]
Generating train split: 11138659 examples [02:33, 40209.03 examples/s]
Generating train split: 11169659 examples [02:33, 68570.58 examples/s]
Generating train split: 11190659 examples [02:33, 71843.91 examples/s]
Generating train split: 11206659 examples [02:33, 74534.50 examples/s]
Generating train split: 11217659 examples [02:33, 73262.77 examples/s]
Generating train split: 11228659 examples [02:34, 61351.12 examples/s]
Generating train split: 11239659 examples [02:34, 54764.24 examples/s]
Generating train split: 11250659 examples [02:34, 53506.91 examples/s]
Generating train split: 11261659 examples [02:34, 59196.19 examples/s]
Generating train split: 11287659 examples [02:35, 75087.82 examples/s]
Generating train split: 11303659 examples [02:35, 88244.24 examples/s]
Generating train split: 11321659 examples [02:35, 102491.17 examples/s]
Generating train split: 11342659 examples [02:35, 112850.87 examples/s]
Generating train split: 11357659 examples [02:35, 109869.64 examples/s]
Generating train split: 11390659 examples [02:35, 147560.28 examples/s]
Generating train split: 11406659 examples [02:35, 127288.51 examples/s]
Generating train split: 11439659 examples [02:35, 169063.50 examples/s]
Generating train split: 11461659 examples [02:37, 46933.36 examples/s]
Generating train split: 11478659 examples [02:37, 51135.13 examples/s]
Generating train split: 11494659 examples [02:37, 50380.44 examples/s]
Generating train split: 11505659 examples [02:38, 51244.87 examples/s]
Generating train split: 11516659 examples [02:38, 56787.84 examples/s]
Generating train split: 11538659 examples [02:38, 78417.45 examples/s]
Generating train split: 11560659 examples [02:38, 85688.92 examples/s]
Generating train split: 11575659 examples [02:38, 84702.09 examples/s]
Generating train split: 11591659 examples [02:38, 97117.91 examples/s]
Generating train split: 11606659 examples [02:39, 48761.99 examples/s]
Generating train split: 11617659 examples [02:39, 54198.22 examples/s]
Generating train split: 11645659 examples [02:39, 83951.07 examples/s]
Generating train split: 11683659 examples [02:39, 122642.75 examples/s]
Generating train split: 11706659 examples [02:40, 135183.52 examples/s]
Generating train split: 11726659 examples [02:40, 128124.23 examples/s]
Generating train split: 11742659 examples [02:40, 88124.47 examples/s]
Generating train split: 11758659 examples [02:40, 86639.52 examples/s]
Generating train split: 11774659 examples [02:40, 98230.88 examples/s]
Generating train split: 11807659 examples [02:41, 70544.93 examples/s]
Generating train split: 11818659 examples [02:41, 73324.21 examples/s]
Generating train split: 11829659 examples [02:42, 40533.91 examples/s]
Generating train split: 11857659 examples [02:42, 61043.31 examples/s]
Generating train split: 11873659 examples [02:42, 71307.26 examples/s]
Generating train split: 11889659 examples [02:42, 74332.93 examples/s]
Generating train split: 11904659 examples [02:43, 62504.41 examples/s]
Generating train split: 11925659 examples [02:43, 75031.62 examples/s]
Generating train split: 11939659 examples [02:43, 65712.40 examples/s]
Generating train split: 11950659 examples [02:44, 45033.08 examples/s]
Generating train split: 11961659 examples [02:44, 51352.81 examples/s]
Generating train split: 11971659 examples [02:44, 56576.80 examples/s]
Generating train split: 11999659 examples [02:44, 88926.66 examples/s]
Generating train split: 12026659 examples [02:44, 121109.37 examples/s]
Generating train split: 12065659 examples [02:44, 156860.63 examples/s]
Generating train split: 12087659 examples [02:44, 155341.93 examples/s]
Generating train split: 12105659 examples [02:45, 124658.06 examples/s]
Generating train split: 12121659 examples [02:45, 75696.42 examples/s]
Generating train split: 12138659 examples [02:45, 81223.91 examples/s]
Generating train split: 12165659 examples [02:45, 108937.77 examples/s]
Generating train split: 12182659 examples [02:46, 61226.68 examples/s]
Generating train split: 12198659 examples [02:46, 59179.67 examples/s]
Generating train split: 12210659 examples [02:46, 65699.50 examples/s]
Generating train split: 12221659 examples [02:47, 60645.46 examples/s]
Generating train split: 12230659 examples [02:47, 59473.00 examples/s]
Generating train split: 12257659 examples [02:47, 81780.06 examples/s]
Generating train split: 12288659 examples [02:47, 77184.58 examples/s]
Generating train split: 12299659 examples [02:48, 62703.95 examples/s]
Generating train split: 12309659 examples [02:48, 41601.62 examples/s]
Generating train split: 12337659 examples [02:49, 63854.12 examples/s]
Generating train split: 12366659 examples [02:49, 79930.61 examples/s]
Generating train split: 12392659 examples [02:49, 98608.53 examples/s]
Generating train split: 12408659 examples [02:49, 98828.29 examples/s]
Generating train split: 12426659 examples [02:49, 99685.37 examples/s]
Generating train split: 12458659 examples [02:49, 104595.25 examples/s]
Generating train split: 12474659 examples [02:50, 113175.61 examples/s]
Generating train split: 12491659 examples [02:50, 66600.72 examples/s]
Generating train split: 12512659 examples [02:50, 78140.87 examples/s]
Generating train split: 12523659 examples [02:51, 48087.68 examples/s]
Generating train split: 12533659 examples [02:51, 46535.68 examples/s]
Generating train split: 12544659 examples [02:51, 46732.62 examples/s]
Generating train split: 12554659 examples [02:52, 53023.55 examples/s]
Generating train split: 12565659 examples [02:52, 49016.78 examples/s]
Generating train split: 12596659 examples [02:52, 83234.31 examples/s]
Generating train split: 12612659 examples [02:52, 95058.58 examples/s]
Generating train split: 12633659 examples [02:52, 76920.22 examples/s]
Generating train split: 12648659 examples [02:53, 55114.83 examples/s]
Generating train split: 12665659 examples [02:53, 58261.71 examples/s]
Generating train split: 12676659 examples [02:53, 60703.41 examples/s]
Generating train split: 12693659 examples [02:53, 74706.51 examples/s]
Generating train split: 12716659 examples [02:54, 98189.10 examples/s]
Generating train split: 12738659 examples [02:54, 117346.91 examples/s]
Generating train split: 12766659 examples [02:54, 109667.39 examples/s]
Generating train split: 12799659 examples [02:54, 112290.30 examples/s]
Generating train split: 12816659 examples [02:55, 92556.34 examples/s]
Generating train split: 12840659 examples [02:55, 108041.88 examples/s]
Generating train split: 12853659 examples [02:55, 67693.73 examples/s]
Generating train split: 12865659 examples [02:55, 71282.49 examples/s]
Generating train split: 12882659 examples [02:55, 85016.55 examples/s]
Generating train split: 12898659 examples [02:56, 45264.08 examples/s]
Generating train split: 12915659 examples [02:56, 57123.05 examples/s]
Generating train split: 12947659 examples [02:56, 85949.39 examples/s]
Generating train split: 12969659 examples [02:57, 101841.55 examples/s]
Generating train split: 12986659 examples [02:57, 55714.90 examples/s]
Generating train split: 13004659 examples [02:57, 67311.06 examples/s]
Generating train split: 13021659 examples [02:58, 53511.74 examples/s]
Generating train split: 13032659 examples [02:58, 52005.65 examples/s]
Generating train split: 13042659 examples [02:58, 54239.74 examples/s]
Generating train split: 13053659 examples [02:58, 56388.31 examples/s]
Generating train split: 13071659 examples [02:59, 72664.67 examples/s]
Generating train split: 13104659 examples [02:59, 114542.31 examples/s]
Generating train split: 13124659 examples [02:59, 110826.36 examples/s]
Generating train split: 13141659 examples [02:59, 120376.32 examples/s]
Generating train split: 13165659 examples [02:59, 121041.66 examples/s]
Generating train split: 13181659 examples [02:59, 96011.14 examples/s]
Generating train split: 13202659 examples [03:00, 101205.65 examples/s]
Generating train split: 13219659 examples [03:00, 62397.60 examples/s]
Generating train split: 13228659 examples [03:00, 59177.33 examples/s]
Generating train split: 13244659 examples [03:01, 63347.34 examples/s]
Generating train split: 13256659 examples [03:01, 61980.76 examples/s]
Generating train split: 13267659 examples [03:01, 63363.17 examples/s]
Generating train split: 13278659 examples [03:01, 57622.48 examples/s]
Generating train split: 13299659 examples [03:01, 78949.64 examples/s]
Generating train split: 13338659 examples [03:02, 53449.17 examples/s]
Generating train split: 13359659 examples [03:03, 55253.69 examples/s]
Generating train split: 13376659 examples [03:03, 66103.81 examples/s]
Generating train split: 13387659 examples [03:03, 59915.38 examples/s]
Generating train split: 13405659 examples [03:03, 71744.16 examples/s]
Generating train split: 13417659 examples [03:03, 77954.83 examples/s]
Generating train split: 13429659 examples [03:03, 76875.37 examples/s]
Generating train split: 13452659 examples [03:03, 102904.81 examples/s]
Generating train split: 13485659 examples [03:04, 128118.89 examples/s]
Generating train split: 13506659 examples [03:04, 131774.28 examples/s]
Generating train split: 13522659 examples [03:04, 136733.27 examples/s]
Generating train split: 13539659 examples [03:04, 120299.10 examples/s]
Generating train split: 13556659 examples [03:04, 95558.54 examples/s]
Generating train split: 13568659 examples [03:04, 98387.56 examples/s]
Generating train split: 13580659 examples [03:05, 88123.42 examples/s]
Generating train split: 13594659 examples [03:05, 86037.48 examples/s]
Generating train split: 13605659 examples [03:05, 76392.50 examples/s]
Generating train split: 13614659 examples [03:06, 38209.95 examples/s]
Generating train split: 13623659 examples [03:06, 42654.68 examples/s]
Generating train split: 13634659 examples [03:06, 46270.78 examples/s]
Generating train split: 13645659 examples [03:06, 54525.29 examples/s]
Generating train split: 13674659 examples [03:06, 90847.33 examples/s]
Generating train split: 13689659 examples [03:06, 101008.08 examples/s]
Generating train split: 13703659 examples [03:07, 42481.69 examples/s]
Generating train split: 13719659 examples [03:07, 46877.14 examples/s]
Generating train split: 13736659 examples [03:08, 57929.85 examples/s]
Generating train split: 13751659 examples [03:08, 60018.75 examples/s]
Generating train split: 13778659 examples [03:08, 86857.72 examples/s]
Generating train split: 13795659 examples [03:08, 97704.97 examples/s]
Generating train split: 13812659 examples [03:08, 100844.74 examples/s]
Generating train split: 13849659 examples [03:08, 150701.32 examples/s]
Generating train split: 13872659 examples [03:09, 125796.99 examples/s]
Generating train split: 13893659 examples [03:09, 111050.38 examples/s]
Generating train split: 13910659 examples [03:09, 74844.23 examples/s]
Generating train split: 13922659 examples [03:09, 77646.04 examples/s]
Generating train split: 13938659 examples [03:10, 84568.87 examples/s]
Generating train split: 13958659 examples [03:10, 71364.76 examples/s]
Generating train split: 13968659 examples [03:10, 66217.90 examples/s]
Generating train split: 13979659 examples [03:10, 62495.84 examples/s]
Generating train split: 13989659 examples [03:10, 64462.49 examples/s]
Generating train split: 14010659 examples [03:11, 53142.84 examples/s]
Generating train split: 14027659 examples [03:11, 66299.85 examples/s]
Generating train split: 14040659 examples [03:11, 56791.28 examples/s]
Generating train split: 14059659 examples [03:12, 63417.93 examples/s]
Generating train split: 14070659 examples [03:13, 32915.51 examples/s]
Generating train split: 14086659 examples [03:13, 43631.27 examples/s]
Generating train split: 14121659 examples [03:13, 76494.19 examples/s]
Generating train split: 14156659 examples [03:13, 111880.85 examples/s]
Generating train split: 14177659 examples [03:13, 114053.79 examples/s]
Generating train split: 14197659 examples [03:13, 110999.80 examples/s]
Generating train split: 14213659 examples [03:13, 111083.08 examples/s]
Generating train split: 14234659 examples [03:14, 115681.70 examples/s]
Generating train split: 14249659 examples [03:14, 114468.81 examples/s]
Generating train split: 14264659 examples [03:14, 57327.88 examples/s]
Generating train split: 14284659 examples [03:14, 73503.10 examples/s]
Generating train split: 14300659 examples [03:15, 84229.88 examples/s]
Generating train split: 14317659 examples [03:15, 95408.33 examples/s]
Generating train split: 14331659 examples [03:15, 91994.38 examples/s]
Generating train split: 14347659 examples [03:15, 104498.26 examples/s]
Generating train split: 14362659 examples [03:15, 67137.69 examples/s]
Generating train split: 14374659 examples [03:16, 43193.10 examples/s]
Generating train split: 14385659 examples [03:16, 46983.11 examples/s]
Generating train split: 14396659 examples [03:16, 53454.21 examples/s]
Generating train split: 14407659 examples [03:16, 59061.51 examples/s]
Generating train split: 14417659 examples [03:17, 52743.99 examples/s]
Generating train split: 14428659 examples [03:17, 52695.23 examples/s]
Generating train split: 14439659 examples [03:17, 45676.83 examples/s]
Generating train split: 14456659 examples [03:17, 63476.51 examples/s]
Generating train split: 14472659 examples [03:17, 74989.38 examples/s]
Generating train split: 14490659 examples [03:18, 88705.13 examples/s]
Generating train split: 14535659 examples [03:18, 158162.41 examples/s]
Generating train split: 14556659 examples [03:18, 164362.27 examples/s]
Generating train split: 14576659 examples [03:18, 83737.89 examples/s]
Generating train split: 14594659 examples [03:19, 69906.97 examples/s]
Generating train split: 14611659 examples [03:19, 70000.24 examples/s]
Generating train split: 14624659 examples [03:19, 59972.47 examples/s]
Generating train split: 14642659 examples [03:19, 72072.21 examples/s]
Generating train split: 14656659 examples [03:20, 75972.21 examples/s]
Generating train split: 14672659 examples [03:20, 84498.46 examples/s]
Generating train split: 14683659 examples [03:20, 84591.93 examples/s]
Generating train split: 14696659 examples [03:20, 89615.25 examples/s]
Generating train split: 14707659 examples [03:20, 58826.94 examples/s]
Generating train split: 14718659 examples [03:21, 59770.46 examples/s]
Generating train split: 14741659 examples [03:21, 84458.63 examples/s]
Generating train split: 14753659 examples [03:21, 66225.16 examples/s]
Generating train split: 14764659 examples [03:22, 40380.86 examples/s]
Generating train split: 14775659 examples [03:22, 40466.17 examples/s]
Generating train split: 14786659 examples [03:22, 44703.72 examples/s]
Generating train split: 14797659 examples [03:22, 51350.71 examples/s]
Generating train split: 14819659 examples [03:22, 73326.41 examples/s]
Generating train split: 14841659 examples [03:22, 93362.23 examples/s]
Generating train split: 14857659 examples [03:23, 103357.84 examples/s]
Generating train split: 14890659 examples [03:23, 149560.35 examples/s]
Generating train split: 14919659 examples [03:23, 167125.87 examples/s]
Generating train split: 14941659 examples [03:23, 160960.17 examples/s]
Generating train split: 14963659 examples [03:24, 65533.49 examples/s]
Generating train split: 14978659 examples [03:24, 64786.79 examples/s]
Generating train split: 14990659 examples [03:24, 63305.92 examples/s]
Generating train split: 15017659 examples [03:24, 89199.54 examples/s]
Generating train split: 15033659 examples [03:24, 93363.56 examples/s]
Generating train split: 15048659 examples [03:25, 76400.93 examples/s]
Generating train split: 15064659 examples [03:25, 53540.18 examples/s]
Generating train split: 15075659 examples [03:25, 55979.16 examples/s]
Generating train split: 15112659 examples [03:26, 92444.97 examples/s]
Generating train split: 15130659 examples [03:26, 50576.61 examples/s]
Generating train split: 15148659 examples [03:27, 57789.95 examples/s]
Generating train split: 15159659 examples [03:27, 56657.83 examples/s]
Generating train split: 15183659 examples [03:27, 79192.40 examples/s]
Generating train split: 15224659 examples [03:27, 107136.95 examples/s]
Generating train split: 15239659 examples [03:27, 105473.75 examples/s]
Generating train split: 15255659 examples [03:27, 108947.94 examples/s]
Generating train split: 15271659 examples [03:28, 94758.22 examples/s]
Generating train split: 15282659 examples [03:28, 88467.07 examples/s]
Generating train split: 15293659 examples [03:28, 85970.89 examples/s]
Generating train split: 15304659 examples [03:28, 82820.10 examples/s]
Generating train split: 15315659 examples [03:28, 64239.19 examples/s]
Generating train split: 15326659 examples [03:29, 61844.29 examples/s]
Generating train split: 15336659 examples [03:29, 58470.22 examples/s]
Generating train split: 15343659 examples [03:29, 52154.29 examples/s]
Generating train split: 15360659 examples [03:29, 67933.12 examples/s]
Generating train split: 15375659 examples [03:29, 78821.05 examples/s]
Generating train split: 15392659 examples [03:29, 91595.98 examples/s]
Generating train split: 15404931 examples [03:30, 82423.53 examples/s]
Generating train split: 15414931 examples [03:30, 68134.17 examples/s]
Generating train split: 15435931 examples [03:30, 87977.34 examples/s]
Generating train split: 15446931 examples [03:30, 66652.18 examples/s]
Generating train split: 15458931 examples [03:30, 67604.82 examples/s]
Generating train split: 15467931 examples [03:31, 56781.22 examples/s]
Generating train split: 15478931 examples [03:31, 36230.09 examples/s]
Generating train split: 15486931 examples [03:32, 32735.41 examples/s]
Generating train split: 15505931 examples [03:32, 47929.36 examples/s]
Generating train split: 15516931 examples [03:32, 56150.55 examples/s]
Generating train split: 15556931 examples [03:32, 107290.54 examples/s]
Generating train split: 15577931 examples [03:32, 120281.52 examples/s]
Generating train split: 15593931 examples [03:32, 105365.27 examples/s]
Generating train split: 15610931 examples [03:32, 117250.07 examples/s]
Generating train split: 15626931 examples [03:33, 80610.90 examples/s]
Generating train split: 15644931 examples [03:33, 68088.30 examples/s]
Generating train split: 15654931 examples [03:33, 64917.88 examples/s]
Generating train split: 15665931 examples [03:34, 58815.17 examples/s]
Generating train split: 15681931 examples [03:34, 72376.22 examples/s]
Generating train split: 15695931 examples [03:34, 70315.25 examples/s]
Generating train split: 15705931 examples [03:34, 74759.60 examples/s]
Generating train split: 15727931 examples [03:34, 97593.26 examples/s]
Generating train split: 15742931 examples [03:34, 83888.70 examples/s]
Generating train split: 15752931 examples [03:34, 77397.23 examples/s]
Generating train split: 15772931 examples [03:35, 79515.83 examples/s]
Generating train split: 15793931 examples [03:35, 97323.54 examples/s]
Generating train split: 15804931 examples [03:35, 70800.79 examples/s]
Generating train split: 15820931 examples [03:35, 80692.54 examples/s]
Generating train split: 15831931 examples [03:36, 62677.46 examples/s]
Generating train split: 15841931 examples [03:36, 46538.43 examples/s]
Generating train split: 15852931 examples [03:36, 53235.41 examples/s]
Generating train split: 15862931 examples [03:36, 51494.46 examples/s]
Generating train split: 15873931 examples [03:37, 53853.04 examples/s]
Generating train split: 15889931 examples [03:37, 67249.28 examples/s]
Generating train split: 15916931 examples [03:37, 101699.52 examples/s]
Generating train split: 15944931 examples [03:37, 128121.61 examples/s]
Generating train split: 15961931 examples [03:37, 132877.72 examples/s]
Generating train split: 15977931 examples [03:37, 115907.83 examples/s]
Generating train split: 15992931 examples [03:37, 101587.15 examples/s]
Generating train split: 16008931 examples [03:38, 71715.97 examples/s]
Generating train split: 16029931 examples [03:38, 70100.52 examples/s]
Generating train split: 16042931 examples [03:39, 52650.04 examples/s]
Generating train split: 16054931 examples [03:39, 54117.59 examples/s]
Generating train split: 16065931 examples [03:39, 60030.96 examples/s]
Generating train split: 16086931 examples [03:39, 78722.12 examples/s]
Generating train split: 16102931 examples [03:39, 90301.89 examples/s]
Generating train split: 16116931 examples [03:40, 59201.93 examples/s]
Generating train split: 16136931 examples [03:40, 76232.15 examples/s]
Generating train split: 16147931 examples [03:40, 59064.24 examples/s]
Generating train split: 16158931 examples [03:40, 45920.46 examples/s]
Generating train split: 16180931 examples [03:41, 53743.56 examples/s]
Generating train split: 16191931 examples [03:41, 57858.30 examples/s]
Generating train split: 16207931 examples [03:41, 69726.89 examples/s]
Generating train split: 16223931 examples [03:41, 75405.95 examples/s]
Generating train split: 16249931 examples [03:41, 103805.96 examples/s]
Generating train split: 16277931 examples [03:42, 113098.88 examples/s]
Generating train split: 16297931 examples [03:42, 114386.20 examples/s]
Generating train split: 16314931 examples [03:42, 114244.05 examples/s]
Generating train split: 16330931 examples [03:42, 78557.97 examples/s]
Generating train split: 16341931 examples [03:42, 70606.90 examples/s]
Generating train split: 16351931 examples [03:43, 73753.94 examples/s]
Generating train split: 16361931 examples [03:43, 69438.13 examples/s]
Generating train split: 16377931 examples [03:43, 85334.26 examples/s]
Generating train split: 16388931 examples [03:43, 63107.84 examples/s]
Generating train split: 16399931 examples [03:43, 70635.28 examples/s]
Generating train split: 16421931 examples [03:44, 66916.90 examples/s]
Generating train split: 16431931 examples [03:44, 63824.10 examples/s]
Generating train split: 16442931 examples [03:44, 62918.99 examples/s]
Generating train split: 16464931 examples [03:44, 55029.99 examples/s]
Generating train split: 16487931 examples [03:45, 75024.73 examples/s]
Generating train split: 16499931 examples [03:45, 69462.96 examples/s]
Generating train split: 16510931 examples [03:45, 67602.28 examples/s]
Generating train split: 16521931 examples [03:45, 50758.24 examples/s]
Generating train split: 16531931 examples [03:45, 57036.44 examples/s]
Generating train split: 16542931 examples [03:46, 59304.40 examples/s]
Generating train split: 16553931 examples [03:46, 66317.91 examples/s]
Generating train split: 16581931 examples [03:46, 101559.16 examples/s]
Generating train split: 16607931 examples [03:46, 127032.03 examples/s]
Generating train split: 16629931 examples [03:46, 143211.23 examples/s]
Generating train split: 16649931 examples [03:46, 136542.36 examples/s]
Generating train split: 16665931 examples [03:47, 84975.09 examples/s]
Generating train split: 16688931 examples [03:47, 71338.50 examples/s]
Generating train split: 16700931 examples [03:47, 58493.42 examples/s]
Generating train split: 16712931 examples [03:48, 54853.86 examples/s]
Generating train split: 16728931 examples [03:48, 67662.51 examples/s]
Generating train split: 16740931 examples [03:48, 56641.21 examples/s]
Generating train split: 16768931 examples [03:48, 85217.51 examples/s]
Generating train split: 16785931 examples [03:49, 77549.38 examples/s]
Generating train split: 16796931 examples [03:49, 66329.16 examples/s]
Generating train split: 16817931 examples [03:49, 87712.05 examples/s]
Generating train split: 16832931 examples [03:49, 65119.86 examples/s]
Generating train split: 16848931 examples [03:49, 76752.96 examples/s]
Generating train split: 16864931 examples [03:50, 63303.16 examples/s]
Generating train split: 16875931 examples [03:50, 46503.56 examples/s]
Generating train split: 16887931 examples [03:50, 50497.44 examples/s]
Generating train split: 16898931 examples [03:51, 57740.12 examples/s]
Generating train split: 16920931 examples [03:51, 73971.21 examples/s]
Generating train split: 16937931 examples [03:51, 87758.28 examples/s]
Generating train split: 16959931 examples [03:51, 111880.09 examples/s]
Generating train split: 16987931 examples [03:51, 143767.65 examples/s]
Generating train split: 17010931 examples [03:51, 155940.05 examples/s]
Generating train split: 17032931 examples [03:52, 87354.03 examples/s]
Generating train split: 17047931 examples [03:52, 65006.80 examples/s]
Generating train split: 17070931 examples [03:52, 82795.35 examples/s]
Generating train split: 17087931 examples [03:52, 76418.69 examples/s]
Generating train split: 17120931 examples [03:53, 107098.62 examples/s]
Generating train split: 17135931 examples [03:53, 97747.33 examples/s]
Generating train split: 17151931 examples [03:53, 77859.81 examples/s]
Generating train split: 17162931 examples [03:53, 82285.68 examples/s]
Generating train split: 17173931 examples [03:54, 55109.74 examples/s]
Generating train split: 17190931 examples [03:54, 66451.91 examples/s]
Generating train split: 17201931 examples [03:54, 64899.35 examples/s]
Generating train split: 17212931 examples [03:54, 58960.50 examples/s]
Generating train split: 17224931 examples [03:54, 66415.47 examples/s]
Generating train split: 17235931 examples [03:55, 60179.93 examples/s]
Generating train split: 17245931 examples [03:55, 39693.74 examples/s]
Generating train split: 17256931 examples [03:55, 48406.96 examples/s]
Generating train split: 17267931 examples [03:55, 53896.66 examples/s]
Generating train split: 17294931 examples [03:56, 86738.84 examples/s]
Generating train split: 17317931 examples [03:56, 110964.34 examples/s]
Generating train split: 17345931 examples [03:56, 143716.82 examples/s]
Generating train split: 17363931 examples [03:56, 142237.71 examples/s]
Generating train split: 17385931 examples [03:56, 126479.33 examples/s]
Generating train split: 17400931 examples [03:57, 60732.77 examples/s]
Generating train split: 17416931 examples [03:57, 72197.44 examples/s]
Generating train split: 17431931 examples [03:57, 74141.78 examples/s]
Generating train split: 17449931 examples [03:57, 84748.37 examples/s]
Generating train split: 17465931 examples [03:57, 81075.91 examples/s]
Generating train split: 17488931 examples [03:58, 99257.64 examples/s]
Generating train split: 17504931 examples [03:58, 77361.86 examples/s]
Generating train split: 17515931 examples [03:58, 76732.46 examples/s]
Generating train split: 17526931 examples [03:59, 42415.40 examples/s]
Generating train split: 17548931 examples [03:59, 55134.30 examples/s]
Generating train split: 17559931 examples [03:59, 49480.66 examples/s]
Generating train split: 17580931 examples [03:59, 69023.35 examples/s]
Generating train split: 17591931 examples [03:59, 65478.00 examples/s]
Generating train split: 17602931 examples [04:00, 68640.20 examples/s]
Generating train split: 17611931 examples [04:00, 48629.46 examples/s]
Generating train split: 17628931 examples [04:00, 59108.61 examples/s]
Generating train split: 17639931 examples [04:00, 61991.71 examples/s]
Generating train split: 17667931 examples [04:00, 94687.35 examples/s]
Generating train split: 17693931 examples [04:01, 120933.13 examples/s]
Generating train split: 17710931 examples [04:01, 129976.29 examples/s]
Generating train split: 17730009 examples [04:01, 136644.13 examples/s]
Generating train split: 17756009 examples [04:01, 163662.96 examples/s]
Generating train split: 17779009 examples [04:02, 80716.57 examples/s]
Generating train split: 17793009 examples [04:02, 73311.39 examples/s]
Generating train split: 17810009 examples [04:02, 74893.81 examples/s]
Generating train split: 17822009 examples [04:02, 67733.42 examples/s]
Generating train split: 17834009 examples [04:02, 71074.05 examples/s]
Generating train split: 17857009 examples [04:03, 88678.40 examples/s]
Generating train split: 17869009 examples [04:03, 62346.38 examples/s]
Generating train split: 17891009 examples [04:03, 66683.29 examples/s]
Generating train split: 17902009 examples [04:03, 56820.89 examples/s]
Generating train split: 17918009 examples [04:04, 69837.36 examples/s]
Generating train split: 17929009 examples [04:04, 69417.28 examples/s]
Generating train split: 17943009 examples [04:04, 49957.96 examples/s]
Generating train split: 17953009 examples [04:05, 41483.59 examples/s]
Generating train split: 17970009 examples [04:05, 51412.65 examples/s]
Generating train split: 17983009 examples [04:05, 60058.15 examples/s]
Generating train split: 18020009 examples [04:05, 103846.15 examples/s]
Generating train split: 18047009 examples [04:05, 115971.33 examples/s]
Generating train split: 18063009 examples [04:05, 115908.98 examples/s]
Generating train split: 18080009 examples [04:06, 87556.01 examples/s]
Generating train split: 18101009 examples [04:06, 107104.83 examples/s]
Generating train split: 18117009 examples [04:06, 111988.37 examples/s]
Generating train split: 18133009 examples [04:06, 95143.17 examples/s]
Generating train split: 18155009 examples [04:06, 94317.47 examples/s]
Generating train split: 18166009 examples [04:07, 44972.10 examples/s]
Generating train split: 18194009 examples [04:07, 66981.86 examples/s]
Generating train split: 18215009 examples [04:07, 83766.70 examples/s]
Generating train split: 18231009 examples [04:08, 64466.94 examples/s]
Generating train split: 18248009 examples [04:08, 75520.96 examples/s]
Generating train split: 18264009 examples [04:08, 77385.65 examples/s]
Generating train split: 18277009 examples [04:09, 59161.08 examples/s]
Generating train split: 18286009 examples [04:09, 60991.92 examples/s]
Generating train split: 18297009 examples [04:09, 42952.70 examples/s]
Generating train split: 18307009 examples [04:09, 49152.44 examples/s]
Generating train split: 18318009 examples [04:09, 48683.12 examples/s]
Generating train split: 18329009 examples [04:10, 57042.75 examples/s]
Generating train split: 18346009 examples [04:10, 74844.04 examples/s]
Generating train split: 18358009 examples [04:10, 75627.19 examples/s]
Generating train split: 18380009 examples [04:10, 103553.14 examples/s]
Generating train split: 18418009 examples [04:10, 147879.46 examples/s]
Generating train split: 18435009 examples [04:10, 136354.27 examples/s]
Generating train split: 18453009 examples [04:11, 108833.29 examples/s]
Generating train split: 18466009 examples [04:11, 91157.06 examples/s]
Generating train split: 18487009 examples [04:11, 106051.36 examples/s]
Generating train split: 18503009 examples [04:11, 84498.46 examples/s]
Generating train split: 18525009 examples [04:11, 95575.58 examples/s]
Generating train split: 18536009 examples [04:12, 91354.86 examples/s]
Generating train split: 18549009 examples [04:12, 83511.65 examples/s]
Generating train split: 18560009 examples [04:12, 57608.65 examples/s]
Generating train split: 18571009 examples [04:12, 64899.55 examples/s]
Generating train split: 18586009 examples [04:12, 75003.59 examples/s]
Generating train split: 18608009 examples [04:12, 93318.92 examples/s]
Generating train split: 18622009 examples [04:13, 62389.63 examples/s]
Generating train split: 18633009 examples [04:13, 61970.79 examples/s]
Generating train split: 18645009 examples [04:13, 51276.03 examples/s]
Generating train split: 18652009 examples [04:14, 30237.04 examples/s]
Generating train split: 18657009 examples [04:14, 27755.46 examples/s]
Generating train split: 18678009 examples [04:15, 44302.86 examples/s]
Generating train split: 18699009 examples [04:15, 57281.06 examples/s]
Generating train split: 18732009 examples [04:15, 90612.95 examples/s]
Generating train split: 18775009 examples [04:15, 133676.76 examples/s]
Generating train split: 18796009 examples [04:15, 137920.51 examples/s]
Generating train split: 18825009 examples [04:15, 154051.39 examples/s]
Generating train split: 18847009 examples [04:15, 141727.29 examples/s]
Generating train split: 18864009 examples [04:16, 64197.17 examples/s]
Generating train split: 18880009 examples [04:16, 72898.72 examples/s]
Generating train split: 18898009 examples [04:17, 78928.47 examples/s]
Generating train split: 18916009 examples [04:17, 57942.23 examples/s]
Generating train split: 18927009 examples [04:17, 58795.79 examples/s]
Generating train split: 18938009 examples [04:17, 65428.93 examples/s]
Generating train split: 18949009 examples [04:18, 56332.29 examples/s]
Generating train split: 18959009 examples [04:18, 62272.52 examples/s]
Generating train split: 18980009 examples [04:18, 87094.23 examples/s]
Generating train split: 18997009 examples [04:18, 80199.31 examples/s]
Generating train split: 19008009 examples [04:18, 67512.07 examples/s]
Generating train split: 19019009 examples [04:19, 42020.80 examples/s]
Generating train split: 19029009 examples [04:19, 41298.22 examples/s]
Generating train split: 19049009 examples [04:19, 60052.87 examples/s]
Generating train split: 19076009 examples [04:19, 84832.39 examples/s]
Generating train split: 19097009 examples [04:20, 105459.66 examples/s]
Generating train split: 19112009 examples [04:20, 101482.61 examples/s]
Generating train split: 19133009 examples [04:20, 114631.49 examples/s]
Generating train split: 19157009 examples [04:20, 132318.02 examples/s]
Generating train split: 19177009 examples [04:20, 141538.98 examples/s]
Generating train split: 19210009 examples [04:20, 168566.07 examples/s]
Generating train split: 19232009 examples [04:21, 122702.66 examples/s]
Generating train split: 19247009 examples [04:21, 84254.54 examples/s]
Generating train split: 19264009 examples [04:21, 69554.62 examples/s]
Generating train split: 19275009 examples [04:21, 70382.01 examples/s]
Generating train split: 19291009 examples [04:22, 67932.30 examples/s]
Generating train split: 19312009 examples [04:22, 63767.36 examples/s]
Generating train split: 19325009 examples [04:22, 65247.62 examples/s]
Generating train split: 19333009 examples [04:23, 45285.51 examples/s]
Generating train split: 19355009 examples [04:23, 53921.20 examples/s]
Generating train split: 19366009 examples [04:23, 48955.47 examples/s]
Generating train split: 19373009 examples [04:24, 37001.67 examples/s]
Generating train split: 19384009 examples [04:24, 44316.42 examples/s]
Generating train split: 19401009 examples [04:24, 60468.75 examples/s]
Generating train split: 19411009 examples [04:24, 56010.00 examples/s]
Generating train split: 19466009 examples [04:24, 135311.25 examples/s]
Generating train split: 19490009 examples [04:24, 133624.63 examples/s]
Generating train split: 19525009 examples [04:25, 117284.89 examples/s]
Generating train split: 19546009 examples [04:25, 109066.05 examples/s]
Generating train split: 19563009 examples [04:25, 102120.59 examples/s]
Generating train split: 19580009 examples [04:26, 84147.09 examples/s]
Generating train split: 19591009 examples [04:26, 63341.32 examples/s]
Generating train split: 19602009 examples [04:26, 65371.60 examples/s]
Generating train split: 19618009 examples [04:26, 74816.66 examples/s]
Generating train split: 19630009 examples [04:26, 81193.38 examples/s]
Generating train split: 19641009 examples [04:26, 79953.30 examples/s]
Generating train split: 19661009 examples [04:27, 83455.43 examples/s]
Generating train split: 19673009 examples [04:27, 64989.73 examples/s]
Generating train split: 19696009 examples [04:27, 85168.35 examples/s]
Generating train split: 19706009 examples [04:28, 51797.88 examples/s]
Generating train split: 19718009 examples [04:28, 51774.56 examples/s]
Generating train split: 19729009 examples [04:29, 32833.40 examples/s]
Generating train split: 19757009 examples [04:29, 55202.68 examples/s]
Generating train split: 19768009 examples [04:29, 59342.02 examples/s]
Generating train split: 19787009 examples [04:29, 65563.94 examples/s]
Generating train split: 19817009 examples [04:29, 96647.73 examples/s]
Generating train split: 19833009 examples [04:29, 103598.70 examples/s]
Generating train split: 19863009 examples [04:29, 140144.83 examples/s]
Generating train split: 19884009 examples [04:29, 154744.81 examples/s]
Generating train split: 19905009 examples [04:30, 155560.94 examples/s]
Generating train split: 19928009 examples [04:30, 143312.81 examples/s]
Generating train split: 19945009 examples [04:30, 127924.62 examples/s]
Generating train split: 19961009 examples [04:30, 105812.67 examples/s]
Generating train split: 19977009 examples [04:30, 87691.16 examples/s]
Generating train split: 19988009 examples [04:31, 54602.84 examples/s]
Generating train split: 19997009 examples [04:31, 58633.37 examples/s]
Generating train split: 20007009 examples [04:31, 54669.20 examples/s]
Generating train split: 20019009 examples [04:32, 51846.92 examples/s]
Generating train split: 20029009 examples [04:32, 58072.73 examples/s]
Generating train split: 20039009 examples [04:32, 56423.21 examples/s]
Generating train split: 20053009 examples [04:32, 70181.66 examples/s]
Generating train split: 20062009 examples [04:32, 74031.49 examples/s]
Generating train split: 20078009 examples [04:32, 54362.53 examples/s]
Generating train split: 20088009 examples [04:33, 47220.22 examples/s]
Generating train split: 20098009 examples [04:33, 31476.74 examples/s]
Generating train split: 20119009 examples [04:33, 50516.10 examples/s]
Generating train split: 20135009 examples [04:34, 61395.38 examples/s]
Generating train split: 20165009 examples [04:34, 93300.80 examples/s]
Generating train split: 20189009 examples [04:34, 99053.66 examples/s]
Generating train split: 20205009 examples [04:34, 106259.71 examples/s]
Generating train split: 20222009 examples [04:34, 117565.89 examples/s]
Generating train split: 20254009 examples [04:34, 155292.84 examples/s]
Generating train split: 20276009 examples [04:35, 104851.30 examples/s]
Generating train split: 20298009 examples [04:35, 88099.21 examples/s]
Generating train split: 20315009 examples [04:35, 77410.34 examples/s]
Generating train split: 20331009 examples [04:36, 67309.91 examples/s]
Generating train split: 20343009 examples [04:36, 57082.77 examples/s]
Generating train split: 20353009 examples [04:36, 56604.49 examples/s]
Generating train split: 20374009 examples [04:36, 75983.54 examples/s]
Generating train split: 20386009 examples [04:37, 57115.64 examples/s]
Generating train split: 20397009 examples [04:37, 59442.58 examples/s]
Generating train split: 20409009 examples [04:37, 67130.41 examples/s]
Generating train split: 20424009 examples [04:37, 80209.33 examples/s]
Generating train split: 20439009 examples [04:38, 29260.61 examples/s]
Generating train split: 20460009 examples [04:38, 42694.94 examples/s]
Generating train split: 20487009 examples [04:39, 64460.07 examples/s]
Generating train split: 20510009 examples [04:39, 84174.04 examples/s]
Generating train split: 20549009 examples [04:39, 129401.17 examples/s]
Generating train split: 20572009 examples [04:39, 111718.37 examples/s]
Generating train split: 20592009 examples [04:39, 117280.75 examples/s]
Generating train split: 20614009 examples [04:39, 127924.66 examples/s]
Generating train split: 20633009 examples [04:40, 96354.57 examples/s]
Generating train split: 20654009 examples [04:40, 113942.62 examples/s]
Generating train split: 20675009 examples [04:40, 102725.23 examples/s]
Generating train split: 20691009 examples [04:41, 65243.97 examples/s]
Generating train split: 20702009 examples [04:41, 53609.21 examples/s]
Generating train split: 20725009 examples [04:41, 73948.28 examples/s]
Generating train split: 20738009 examples [04:41, 68176.33 examples/s]
Generating train split: 20754009 examples [04:42, 53900.86 examples/s]
Generating train split: 20766009 examples [04:42, 60051.76 examples/s]
Generating train split: 20777009 examples [04:42, 36548.29 examples/s]
Generating train split: 20797009 examples [04:43, 52494.96 examples/s]
Generating train split: 20813009 examples [04:43, 50056.52 examples/s]
Generating train split: 20828009 examples [04:43, 61231.83 examples/s]
Generating train split: 20839009 examples [04:43, 58001.98 examples/s]
Generating train split: 20883009 examples [04:43, 114599.63 examples/s]
Generating train split: 20904009 examples [04:44, 127159.02 examples/s]
Generating train split: 20931009 examples [04:44, 144112.62 examples/s]
Generating train split: 20954009 examples [04:44, 100420.60 examples/s]
Generating train split: 20972009 examples [04:44, 110274.45 examples/s]
Generating train split: 20993009 examples [04:44, 127833.37 examples/s]
Generating train split: 21014009 examples [04:45, 70321.36 examples/s]
Generating train split: 21029009 examples [04:45, 62721.43 examples/s]
Generating train split: 21040009 examples [04:45, 56128.10 examples/s]
Generating train split: 21049009 examples [04:46, 45486.68 examples/s]
Generating train split: 21070009 examples [04:46, 64265.63 examples/s]
Generating train split: 21087009 examples [04:46, 76148.46 examples/s]
Generating train split: 21104009 examples [04:47, 47124.09 examples/s]
Generating train split: 21114009 examples [04:47, 50042.30 examples/s]
Generating train split: 21125009 examples [04:47, 39692.81 examples/s]
Generating train split: 21138009 examples [04:47, 49214.85 examples/s]
Generating train split: 21154009 examples [04:48, 55513.96 examples/s]
Generating train split: 21170009 examples [04:48, 70305.41 examples/s]
Generating train split: 21211009 examples [04:48, 121288.42 examples/s]
Generating train split: 21233009 examples [04:48, 117616.10 examples/s]
Generating train split: 21261009 examples [04:48, 145204.48 examples/s]
Generating train split: 21283009 examples [04:48, 129654.18 examples/s]
Generating train split: 21302009 examples [04:49, 90154.81 examples/s]
Generating train split: 21321009 examples [04:49, 99362.02 examples/s]
Generating train split: 21335009 examples [04:49, 72170.84 examples/s]
Generating train split: 21350009 examples [04:49, 82661.65 examples/s]
Generating train split: 21365009 examples [04:50, 77285.15 examples/s]
Generating train split: 21375009 examples [04:50, 52773.13 examples/s]
Generating train split: 21389009 examples [04:51, 37470.46 examples/s]
Generating train split: 21399009 examples [04:51, 43036.11 examples/s]
Generating train split: 21419009 examples [04:51, 61061.61 examples/s]
Generating train split: 21435009 examples [04:51, 65187.15 examples/s]
Generating train split: 21445009 examples [04:52, 42477.07 examples/s]
Generating train split: 21467009 examples [04:52, 43932.95 examples/s]
Generating train split: 21488009 examples [04:52, 60043.82 examples/s]
Generating train split: 21505009 examples [04:52, 73790.51 examples/s]
Generating train split: 21519009 examples [04:53, 72734.72 examples/s]
Generating train split: 21563009 examples [04:53, 124249.85 examples/s]
Generating train split: 21596009 examples [04:53, 140019.62 examples/s]
Generating train split: 21617009 examples [04:53, 98973.66 examples/s]
Generating train split: 21634009 examples [04:54, 81756.57 examples/s]
Generating train split: 21662009 examples [04:54, 102998.59 examples/s]
Generating train split: 21678009 examples [04:54, 89428.76 examples/s]
Generating train split: 21693009 examples [04:54, 84069.84 examples/s]
Generating train split: 21709009 examples [04:55, 80260.57 examples/s]
Generating train split: 21719009 examples [04:55, 49139.40 examples/s]
Generating train split: 21736009 examples [04:55, 54580.94 examples/s]
Generating train split: 21747009 examples [04:56, 40960.80 examples/s]
Generating train split: 21758009 examples [04:56, 47994.65 examples/s]
Generating train split: 21780009 examples [04:56, 59484.03 examples/s]
Generating train split: 21806009 examples [04:56, 84325.19 examples/s]
Generating train split: 21822009 examples [04:57, 64095.25 examples/s]
Generating train split: 21833009 examples [04:57, 69654.65 examples/s]
Generating train split: 21843009 examples [04:57, 51767.23 examples/s]
Generating train split: 21866009 examples [04:57, 71571.25 examples/s]
Generating train split: 21879009 examples [04:58, 68677.66 examples/s]
Generating train split: 21890009 examples [04:58, 70954.55 examples/s]
Generating train split: 21934009 examples [04:58, 134441.90 examples/s]
Generating train split: 21956009 examples [04:58, 120517.71 examples/s]
Generating train split: 21978009 examples [04:58, 133811.99 examples/s]
Generating train split: 22000009 examples [04:58, 146762.62 examples/s]
Generating train split: 22021009 examples [04:59, 85462.09 examples/s]
Generating train split: 22042009 examples [04:59, 87865.14 examples/s]
Generating train split: 22057009 examples [04:59, 62978.27 examples/s]
Generating train split: 22067009 examples [05:00, 67125.85 examples/s]
Generating train split: 22078009 examples [05:00, 43614.31 examples/s]
Generating train split: 22089009 examples [05:00, 42248.16 examples/s]
Generating train split: 22096009 examples [05:01, 31643.03 examples/s]
Generating train split: 22117009 examples [05:01, 49376.27 examples/s]
Generating train split: 22133009 examples [05:01, 56337.43 examples/s]
Generating train split: 22143009 examples [05:01, 60871.48 examples/s]
Generating train split: 22163009 examples [05:02, 69278.63 examples/s]
Generating train split: 22174009 examples [05:02, 73699.13 examples/s]
Generating train split: 22193009 examples [05:02, 93683.90 examples/s]
Generating train split: 22214009 examples [05:02, 93489.88 examples/s]
Generating train split: 22236009 examples [05:02, 115038.98 examples/s]
Generating train split: 22263009 examples [05:02, 146785.28 examples/s]
Generating train split: 22282009 examples [05:02, 148841.42 examples/s]
Generating train split: 22299009 examples [05:02, 148621.06 examples/s]
Generating train split: 22316009 examples [05:03, 106537.41 examples/s]
Generating train split: 22333009 examples [05:03, 68296.99 examples/s]
Generating train split: 22346009 examples [05:03, 71190.88 examples/s]
Generating train split: 22360009 examples [05:04, 64372.63 examples/s]
Generating train split: 22376009 examples [05:04, 75589.13 examples/s]
Generating train split: 22392009 examples [05:04, 85412.98 examples/s]
Generating train split: 22403009 examples [05:05, 45186.97 examples/s]
Generating train split: 22414009 examples [05:05, 32534.35 examples/s]
Generating train split: 22420009 examples [05:05, 31699.75 examples/s]
Generating train split: 22436009 examples [05:05, 45237.35 examples/s]
Generating train split: 22445009 examples [05:06, 39558.43 examples/s]
Generating train split: 22463009 examples [05:06, 54373.04 examples/s]
Generating train split: 22477009 examples [05:06, 59916.09 examples/s]
Generating train split: 22488009 examples [05:06, 63042.14 examples/s]
Generating train split: 22502009 examples [05:06, 70459.38 examples/s]
Generating train split: 22512009 examples [05:07, 73487.44 examples/s]
Generating train split: 22524009 examples [05:07, 78684.09 examples/s]
Generating train split: 22539009 examples [05:07, 90725.28 examples/s]
Generating train split: 22560009 examples [05:07, 108038.88 examples/s]
Generating train split: 22584009 examples [05:07, 132911.52 examples/s]
Generating train split: 22605009 examples [05:07, 147945.15 examples/s]
Generating train split: 22637009 examples [05:07, 180631.35 examples/s]
Generating train split: 22659009 examples [05:07, 172328.98 examples/s]
Generating train split: 22680009 examples [05:08, 75568.49 examples/s]
Generating train split: 22701009 examples [05:08, 82971.87 examples/s]
Generating train split: 22718009 examples [05:09, 72598.72 examples/s]
Generating train split: 22733009 examples [05:09, 81888.11 examples/s]
Generating train split: 22749009 examples [05:09, 93197.87 examples/s]
Generating train split: 22766009 examples [05:09, 58952.76 examples/s]
Generating train split: 22776009 examples [05:10, 41297.73 examples/s]
Generating train split: 22787009 examples [05:10, 44012.46 examples/s]
Generating train split: 22798009 examples [05:10, 48647.38 examples/s]
Generating train split: 22809009 examples [05:11, 41738.48 examples/s]
Generating train split: 22815009 examples [05:11, 40147.17 examples/s]
Generating train split: 22835009 examples [05:11, 59205.52 examples/s]
Generating train split: 22853009 examples [05:11, 66804.62 examples/s]
Generating train split: 22885009 examples [05:11, 102867.23 examples/s]
Generating train split: 22903009 examples [05:11, 101607.33 examples/s]
Generating train split: 22917009 examples [05:12, 84588.12 examples/s]
Generating train split: 22941009 examples [05:12, 106049.26 examples/s]
Generating train split: 22964009 examples [05:12, 115239.06 examples/s]
Generating train split: 22990009 examples [05:12, 137611.78 examples/s]
Generating train split: 23028009 examples [05:12, 173353.27 examples/s]
Generating train split: 23050009 examples [05:12, 165386.80 examples/s]
Generating train split: 23069167 examples [05:13, 145393.95 examples/s]
Generating train split: 23107167 examples [05:13, 179760.22 examples/s]
Generating train split: 23131167 examples [05:13, 178642.23 examples/s]
Generating train split: 23163167 examples [05:13, 198679.44 examples/s]
Generating train split: 23186167 examples [05:13, 176566.35 examples/s]
Generating train split: 23212167 examples [05:13, 192340.42 examples/s]
Generating train split: 23238167 examples [05:13, 206704.39 examples/s]
Generating train split: 23268167 examples [05:13, 219249.19 examples/s]
Generating train split: 23293167 examples [05:15, 63965.62 examples/s]
Generating train split: 23310167 examples [05:15, 61940.73 examples/s]
Generating train split: 23324167 examples [05:15, 67324.71 examples/s]
Generating train split: 23338167 examples [05:15, 74158.20 examples/s]
Generating train split: 23354167 examples [05:15, 75110.87 examples/s]
Generating train split: 23391167 examples [05:15, 116278.88 examples/s]
Generating train split: 23410167 examples [05:16, 113273.90 examples/s]
Generating train split: 23428167 examples [05:16, 124328.90 examples/s]
Generating train split: 23446167 examples [05:16, 91076.43 examples/s]
Generating train split: 23467167 examples [05:16, 107555.30 examples/s]
Generating train split: 23483167 examples [05:16, 112555.76 examples/s]
Generating train split: 23501167 examples [05:17, 73757.72 examples/s]
Generating train split: 23512167 examples [05:17, 65481.66 examples/s]
Generating train split: 23547167 examples [05:17, 105994.29 examples/s]
Generating train split: 23577167 examples [05:17, 139055.53 examples/s]
Generating train split: 23602167 examples [05:17, 156578.80 examples/s]
Generating train split: 23626167 examples [05:19, 33475.20 examples/s]
Generating train split: 23643167 examples [05:20, 34889.43 examples/s]
Generating train split: 23657167 examples [05:20, 39804.35 examples/s]
Generating train split: 23678167 examples [05:20, 52747.93 examples/s]
Generating train split: 23698167 examples [05:20, 64825.20 examples/s]
Generating train split: 23726167 examples [05:20, 89398.21 examples/s]
Generating train split: 23743167 examples [05:20, 100467.49 examples/s]
Generating train split: 23765167 examples [05:21, 104023.58 examples/s]
Generating train split: 23783167 examples [05:21, 73761.51 examples/s]
Generating train split: 23805167 examples [05:21, 92191.39 examples/s]
Generating train split: 23823167 examples [05:22, 83162.62 examples/s]
Generating train split: 23837167 examples [05:22, 82211.61 examples/s]
Generating train split: 23852167 examples [05:22, 70789.26 examples/s]
Generating train split: 23887167 examples [05:22, 112525.41 examples/s]
Generating train split: 23911167 examples [05:22, 134808.77 examples/s]
Generating train split: 23932167 examples [05:24, 39148.69 examples/s]
Generating train split: 23947167 examples [05:24, 35042.25 examples/s]
Generating train split: 23973167 examples [05:24, 50518.50 examples/s]
Generating train split: 23988167 examples [05:24, 58908.98 examples/s]
Generating train split: 24004167 examples [05:25, 46992.89 examples/s]
Generating train split: 24043167 examples [05:25, 78163.60 examples/s]
Generating train split: 24064167 examples [05:25, 86341.64 examples/s]
Generating train split: 24082167 examples [05:25, 98030.89 examples/s]
Generating train split: 24105167 examples [05:26, 111519.39 examples/s]
Generating train split: 24121167 examples [05:26, 100024.83 examples/s]
Generating train split: 24138167 examples [05:26, 93927.79 examples/s]
Generating train split: 24154167 examples [05:26, 97627.17 examples/s]
Generating train split: 24166167 examples [05:26, 93705.71 examples/s]
Generating train split: 24194167 examples [05:26, 129443.78 examples/s]
Generating train split: 24210167 examples [05:27, 127058.71 examples/s]
Generating train split: 24238167 examples [05:27, 160747.82 examples/s]
Generating train split: 24258167 examples [05:27, 96921.17 examples/s]
Generating train split: 24274167 examples [05:29, 33114.86 examples/s]
Generating train split: 24286167 examples [05:29, 38186.80 examples/s]
Generating train split: 24298167 examples [05:29, 45134.73 examples/s]
Generating train split: 24313167 examples [05:29, 38332.22 examples/s]
Generating train split: 24342167 examples [05:29, 60265.83 examples/s]
Generating train split: 24381167 examples [05:30, 93142.84 examples/s]
Generating train split: 24400167 examples [05:30, 99420.10 examples/s]
Generating train split: 24417167 examples [05:30, 108185.89 examples/s]
Generating train split: 24435167 examples [05:30, 100059.16 examples/s]
Generating train split: 24451167 examples [05:31, 67436.61 examples/s]
Generating train split: 24469167 examples [05:31, 76538.28 examples/s]
Generating train split: 24495167 examples [05:31, 103536.18 examples/s]
Generating train split: 24511167 examples [05:31, 100765.00 examples/s]
Generating train split: 24543167 examples [05:31, 115760.21 examples/s]
Generating train split: 24563167 examples [05:31, 128639.93 examples/s]
Generating train split: 24579167 examples [05:31, 131835.15 examples/s]
Generating train split: 24615167 examples [05:31, 177915.09 examples/s]
Generating train split: 24637167 examples [05:32, 93489.24 examples/s]
Generating train split: 24658167 examples [05:34, 35108.53 examples/s]
Generating train split: 24678167 examples [05:34, 44608.02 examples/s]
Generating train split: 24695167 examples [05:34, 53736.77 examples/s]
Generating train split: 24713167 examples [05:34, 52375.56 examples/s]
Generating train split: 24735167 examples [05:34, 67132.96 examples/s]
Generating train split: 24753167 examples [05:35, 74483.06 examples/s]
Generating train split: 24781167 examples [05:35, 99707.81 examples/s]
Generating train split: 24805167 examples [05:35, 95784.74 examples/s]
Generating train split: 24824167 examples [05:35, 89525.22 examples/s]
Generating train split: 24839167 examples [05:35, 80752.90 examples/s]
Generating train split: 24864167 examples [05:36, 99254.45 examples/s]
Generating train split: 24890167 examples [05:36, 123451.32 examples/s]
Generating train split: 24910167 examples [05:36, 81540.29 examples/s]
Generating train split: 24940167 examples [05:36, 105475.71 examples/s]
Generating train split: 24957167 examples [05:36, 114722.41 examples/s]
Generating train split: 24973167 examples [05:37, 119728.81 examples/s]
Generating train split: 24989167 examples [05:37, 73762.03 examples/s]
Generating train split: 25004167 examples [05:38, 38035.63 examples/s]
Generating train split: 25016167 examples [05:38, 33807.98 examples/s]
Generating train split: 25043167 examples [05:39, 51520.50 examples/s]
Generating train split: 25066167 examples [05:39, 67448.93 examples/s]
Generating train split: 25083167 examples [05:39, 52457.36 examples/s]
Generating train split: 25116167 examples [05:39, 79517.71 examples/s]
Generating train split: 25131506 examples [05:39, 84288.60 examples/s]
Generating train split: 25168506 examples [05:40, 124512.35 examples/s]
Generating train split: 25190506 examples [05:40, 105710.76 examples/s]
Generating train split: 25208506 examples [05:40, 91573.82 examples/s]
Generating train split: 25224506 examples [05:40, 101345.43 examples/s]
Generating train split: 25250506 examples [05:40, 109644.12 examples/s]
Generating train split: 25277506 examples [05:41, 137331.04 examples/s]
Generating train split: 25299506 examples [05:41, 109670.79 examples/s]
Generating train split: 25316506 examples [05:41, 106940.28 examples/s]
Generating train split: 25342506 examples [05:41, 129636.61 examples/s]
Generating train split: 25360506 examples [05:43, 38502.98 examples/s]
Generating train split: 25377506 examples [05:43, 47796.93 examples/s]
Generating train split: 25392506 examples [05:43, 42230.74 examples/s]
Generating train split: 25428506 examples [05:43, 62575.12 examples/s]
Generating train split: 25445506 examples [05:44, 58199.37 examples/s]
Generating train split: 25456506 examples [05:44, 58659.64 examples/s]
Generating train split: 25498506 examples [05:44, 97828.27 examples/s]
Generating train split: 25515506 examples [05:44, 106069.45 examples/s]
Generating train split: 25542506 examples [05:44, 112020.36 examples/s]
Generating train split: 25564506 examples [05:45, 117624.95 examples/s]
Generating train split: 25580506 examples [05:45, 114477.06 examples/s]
Generating train split: 25596506 examples [05:45, 111979.95 examples/s]
Generating train split: 25616506 examples [05:45, 89961.60 examples/s]
Generating train split: 25638506 examples [05:45, 110276.23 examples/s]
Generating train split: 25655506 examples [05:46, 104211.64 examples/s]
Generating train split: 25668506 examples [05:46, 80135.33 examples/s]
Generating train split: 25690506 examples [05:46, 102343.75 examples/s]
Generating train split: 25712506 examples [05:47, 55723.79 examples/s]
Generating train split: 25727506 examples [05:47, 39856.06 examples/s]
Generating train split: 25749506 examples [05:48, 40356.42 examples/s]
Generating train split: 25782506 examples [05:48, 63163.28 examples/s]
Generating train split: 25798506 examples [05:48, 68232.53 examples/s]
Generating train split: 25822506 examples [05:48, 80922.32 examples/s]
Generating train split: 25838506 examples [05:48, 91172.97 examples/s]
Generating train split: 25855506 examples [05:49, 83905.53 examples/s]
Generating train split: 25879506 examples [05:49, 105044.69 examples/s]
Generating train split: 25901506 examples [05:49, 123437.11 examples/s]
Generating train split: 25918506 examples [05:49, 116508.85 examples/s]
Generating train split: 25943506 examples [05:49, 98354.48 examples/s]
Generating train split: 25961506 examples [05:50, 103812.26 examples/s]
Generating train split: 25975506 examples [05:50, 76703.80 examples/s]
Generating train split: 26007506 examples [05:50, 112023.42 examples/s]
Generating train split: 26035506 examples [05:50, 141161.42 examples/s]
Generating train split: 26056506 examples [05:50, 117357.52 examples/s]
Generating train split: 26072506 examples [05:52, 32806.49 examples/s]
Generating train split: 26089506 examples [05:52, 41162.21 examples/s]
Generating train split: 26106506 examples [05:53, 38693.29 examples/s]
Generating train split: 26127506 examples [05:53, 49086.31 examples/s]
Generating train split: 26162506 examples [05:53, 74937.53 examples/s]
Generating train split: 26178506 examples [05:53, 82079.49 examples/s]
Generating train split: 26200506 examples [05:53, 99022.23 examples/s]
Generating train split: 26217506 examples [05:54, 83950.31 examples/s]
Generating train split: 26241506 examples [05:54, 104690.95 examples/s]
Generating train split: 26259506 examples [05:54, 91471.27 examples/s]
Generating train split: 26280506 examples [05:54, 107739.58 examples/s]
Generating train split: 26295506 examples [05:54, 106153.01 examples/s]
Generating train split: 26314506 examples [05:54, 120460.20 examples/s]
Generating train split: 26331506 examples [05:55, 115464.67 examples/s]
Generating train split: 26348506 examples [05:55, 101191.00 examples/s]
Generating train split: 26380506 examples [05:55, 139016.94 examples/s]
Generating train split: 26403506 examples [05:56, 66747.84 examples/s]
Generating train split: 26418506 examples [05:57, 32997.34 examples/s]
Generating train split: 26429506 examples [05:57, 37983.31 examples/s]
Generating train split: 26451506 examples [05:57, 51294.57 examples/s]
Generating train split: 26468506 examples [05:57, 50320.74 examples/s]
Generating train split: 26480506 examples [05:58, 54263.89 examples/s]
Generating train split: 26507506 examples [05:58, 79679.26 examples/s]
Generating train split: 26536506 examples [05:58, 108332.28 examples/s]
Generating train split: 26565506 examples [05:58, 122277.75 examples/s]
Generating train split: 26587506 examples [05:58, 129310.59 examples/s]
Generating train split: 26605506 examples [05:58, 101913.29 examples/s]
Generating train split: 26622506 examples [05:59, 102260.32 examples/s]
Generating train split: 26640506 examples [05:59, 96565.65 examples/s]
Generating train split: 26652506 examples [05:59, 99889.08 examples/s]
Generating train split: 26665506 examples [05:59, 103006.82 examples/s]
Generating train split: 26682506 examples [05:59, 117403.41 examples/s]
Generating train split: 26696506 examples [05:59, 109589.19 examples/s]
Generating train split: 26709506 examples [06:00, 96692.82 examples/s]
Generating train split: 26738506 examples [06:00, 131433.95 examples/s]
Generating train split: 26757506 examples [06:00, 57253.12 examples/s]
Generating train split: 26773506 examples [06:02, 30404.74 examples/s]
Generating train split: 26784506 examples [06:02, 35558.72 examples/s]
Generating train split: 26801506 examples [06:02, 45689.17 examples/s]
Generating train split: 26813506 examples [06:02, 51263.44 examples/s]
Generating train split: 26824506 examples [06:02, 47939.18 examples/s]
Generating train split: 26839506 examples [06:02, 59698.41 examples/s]
Generating train split: 26862506 examples [06:03, 83767.36 examples/s]
Generating train split: 26895506 examples [06:03, 122448.92 examples/s]
Generating train split: 26912506 examples [06:03, 115773.14 examples/s]
Generating train split: 26939506 examples [06:03, 135979.40 examples/s]
Generating train split: 26956506 examples [06:03, 117566.83 examples/s]
Generating train split: 26973506 examples [06:03, 111043.85 examples/s]
Generating train split: 26992506 examples [06:03, 117344.15 examples/s]
Generating train split: 27009506 examples [06:04, 122405.15 examples/s]
Generating train split: 27027506 examples [06:04, 90113.31 examples/s]
Generating train split: 27038506 examples [06:04, 60333.16 examples/s]
Generating train split: 27066506 examples [06:05, 81344.82 examples/s]
Generating train split: 27077506 examples [06:05, 49976.42 examples/s]
Generating train split: 27088506 examples [06:06, 32926.13 examples/s]
Generating train split: 27099506 examples [06:06, 36866.45 examples/s]
Generating train split: 27110506 examples [06:06, 40950.24 examples/s]
Generating train split: 27121506 examples [06:06, 42086.18 examples/s]
Generating train split: 27151506 examples [06:07, 70864.58 examples/s]
Generating train split: 27180506 examples [06:07, 102906.46 examples/s]
Generating train split: 27196506 examples [06:07, 103249.98 examples/s]
Generating train split: 27212506 examples [06:07, 76147.32 examples/s]
Generating train split: 27236506 examples [06:07, 98644.50 examples/s]
Generating train split: 27252506 examples [06:08, 87909.24 examples/s]
Generating train split: 27274506 examples [06:08, 108521.45 examples/s]
Generating train split: 27297506 examples [06:08, 126946.63 examples/s]
Generating train split: 27313506 examples [06:08, 131309.21 examples/s]
Generating train split: 27334506 examples [06:08, 148984.63 examples/s]
Generating train split: 27354506 examples [06:08, 138690.27 examples/s]
Generating train split: 27371506 examples [06:09, 82889.35 examples/s]
Generating train split: 27389506 examples [06:09, 81732.66 examples/s]
Generating train split: 27400506 examples [06:09, 63051.67 examples/s]
Generating train split: 27422506 examples [06:10, 45158.50 examples/s]
Generating train split: 27431506 examples [06:10, 49209.25 examples/s]
Generating train split: 27441506 examples [06:11, 32992.66 examples/s]
Generating train split: 27464506 examples [06:11, 49438.91 examples/s]
Generating train split: 27475506 examples [06:11, 52324.98 examples/s]
Generating train split: 27485506 examples [06:11, 44625.55 examples/s]
Generating train split: 27509506 examples [06:11, 68624.85 examples/s]
Generating train split: 27548506 examples [06:12, 113214.19 examples/s]
Generating train split: 27576506 examples [06:12, 134944.38 examples/s]
Generating train split: 27600506 examples [06:12, 126298.11 examples/s]
Generating train split: 27621506 examples [06:12, 102808.88 examples/s]
Generating train split: 27636506 examples [06:12, 92989.30 examples/s]
Generating train split: 27669506 examples [06:13, 126454.13 examples/s]
Generating train split: 27686506 examples [06:13, 112023.24 examples/s]
Generating train split: 27707506 examples [06:13, 120125.97 examples/s]
Generating train split: 27723506 examples [06:13, 77891.87 examples/s]
Generating train split: 27734506 examples [06:13, 77833.10 examples/s]
Generating train split: 27750506 examples [06:14, 86430.15 examples/s]
Generating train split: 27761506 examples [06:14, 54011.78 examples/s]
Generating train split: 27777506 examples [06:15, 29919.91 examples/s]
Generating train split: 27788506 examples [06:15, 31938.61 examples/s]
Generating train split: 27815506 examples [06:16, 52397.63 examples/s]
Generating train split: 27846506 examples [06:16, 53978.38 examples/s]
Generating train split: 27857506 examples [06:16, 57770.63 examples/s]
Generating train split: 27878506 examples [06:16, 74138.30 examples/s]
Generating train split: 27910506 examples [06:16, 108644.10 examples/s]
Generating train split: 27931506 examples [06:17, 101986.20 examples/s]
Generating train split: 27958506 examples [06:17, 125882.81 examples/s]
Generating train split: 27976506 examples [06:17, 113844.45 examples/s]
Generating train split: 27995506 examples [06:17, 115753.72 examples/s]
Generating train split: 28021506 examples [06:17, 141020.13 examples/s]
Generating train split: 28043506 examples [06:18, 84695.61 examples/s]
Generating train split: 28059506 examples [06:18, 69595.33 examples/s]
Generating train split: 28075506 examples [06:18, 72522.24 examples/s]
Generating train split: 28085506 examples [06:19, 47119.95 examples/s]
Generating train split: 28097506 examples [06:19, 53758.95 examples/s]
Generating train split: 28107506 examples [06:20, 28836.11 examples/s]
Generating train split: 28118506 examples [06:20, 30038.86 examples/s]
Generating train split: 28161506 examples [06:20, 64097.86 examples/s]
Generating train split: 28178506 examples [06:20, 74282.91 examples/s]
Generating train split: 28195506 examples [06:21, 73895.88 examples/s]
Generating train split: 28211506 examples [06:21, 74656.18 examples/s]
Generating train split: 28230506 examples [06:21, 88522.80 examples/s]
Generating train split: 28246506 examples [06:21, 100482.84 examples/s]
Generating train split: 28286506 examples [06:21, 157280.73 examples/s]
Generating train split: 28312506 examples [06:21, 167482.62 examples/s]
Generating train split: 28361506 examples [06:22, 162648.78 examples/s]
Generating train split: 28381506 examples [06:22, 89692.98 examples/s]
Generating train split: 28398506 examples [06:23, 82940.00 examples/s]
Generating train split: 28413506 examples [06:23, 62787.09 examples/s]
Generating train split: 28429506 examples [06:23, 72638.93 examples/s]
Generating train split: 28441506 examples [06:24, 51574.96 examples/s]
Generating train split: 28451506 examples [06:24, 41432.11 examples/s]
Generating train split: 28461506 examples [06:24, 37042.85 examples/s]
Generating train split: 28482506 examples [06:25, 53211.06 examples/s]
Generating train split: 28492506 examples [06:25, 39776.86 examples/s]
Generating train split: 28510506 examples [06:25, 54644.19 examples/s]
Generating train split: 28531506 examples [06:25, 75028.10 examples/s]
Generating train split: 28552506 examples [06:25, 91738.74 examples/s]
Generating train split: 28568506 examples [06:25, 102896.70 examples/s]
Generating train split: 28589506 examples [06:26, 117368.11 examples/s]
Generating train split: 28605506 examples [06:26, 98328.81 examples/s]
Generating train split: 28636506 examples [06:26, 132708.63 examples/s]
Generating train split: 28657506 examples [06:26, 137176.43 examples/s]
Generating train split: 28680506 examples [06:26, 152291.36 examples/s]
Generating train split: 28702506 examples [06:27, 92528.36 examples/s]
Generating train split: 28718506 examples [06:27, 73082.39 examples/s]
Generating train split: 28731506 examples [06:27, 79040.28 examples/s]
Generating train split: 28745506 examples [06:28, 48095.37 examples/s]
Generating train split: 28771506 examples [06:28, 67199.45 examples/s]
Generating train split: 28786506 examples [06:29, 32022.83 examples/s]
Generating train split: 28818506 examples [06:29, 52163.93 examples/s]
Generating train split: 28835506 examples [06:29, 57772.20 examples/s]
Generating train split: 28849506 examples [06:30, 65310.20 examples/s]
Generating train split: 28867506 examples [06:30, 58515.93 examples/s]
Generating train split: 28899506 examples [06:30, 89123.15 examples/s]
Generating train split: 28916506 examples [06:30, 95892.88 examples/s]
Generating train split: 28932506 examples [06:30, 104951.44 examples/s]
Generating train split: 28973506 examples [06:30, 161963.61 examples/s]
Generating train split: 29003506 examples [06:31, 183236.77 examples/s]
Generating train split: 29031506 examples [06:31, 157058.12 examples/s]
Generating train split: 29052506 examples [06:31, 76937.86 examples/s]
Generating train split: 29068506 examples [06:32, 66119.09 examples/s]
Generating train split: 29089506 examples [06:32, 82011.13 examples/s]
Generating train split: 29107506 examples [06:32, 93908.42 examples/s]
Generating train split: 29127506 examples [06:33, 37413.35 examples/s]
Generating train split: 29142506 examples [06:34, 34121.74 examples/s]
Generating train split: 29154506 examples [06:34, 38155.07 examples/s]
Generating train split: 29169506 examples [06:34, 47978.03 examples/s]
Generating train split: 29192506 examples [06:34, 67390.01 examples/s]
Generating train split: 29211506 examples [06:34, 83199.06 examples/s]
Generating train split: 29227506 examples [06:35, 94499.13 examples/s]
Generating train split: 29244506 examples [06:35, 101170.99 examples/s]
Generating train split: 29268506 examples [06:35, 122063.04 examples/s]
Generating train split: 29290506 examples [06:35, 113474.24 examples/s]
Generating train split: 29322506 examples [06:35, 144375.97 examples/s]
Generating train split: 29343506 examples [06:35, 121757.29 examples/s]
Generating train split: 29360506 examples [06:36, 126015.78 examples/s]
Generating train split: 29377506 examples [06:36, 74266.05 examples/s]
Generating train split: 29393506 examples [06:36, 79824.88 examples/s]
Generating train split: 29408506 examples [06:37, 56762.19 examples/s]
Generating train split: 29430506 examples [06:37, 72434.40 examples/s]
Generating train split: 29445506 examples [06:38, 31484.69 examples/s]
Generating train split: 29455506 examples [06:38, 35664.10 examples/s]
Generating train split: 29463506 examples [06:38, 37899.40 examples/s]
Generating train split: 29475506 examples [06:39, 39073.77 examples/s]
Generating train split: 29495506 examples [06:39, 56875.39 examples/s]
Generating train split: 29508506 examples [06:39, 57701.94 examples/s]
Generating train split: 29521506 examples [06:39, 64408.59 examples/s]
Generating train split: 29542506 examples [06:39, 86678.31 examples/s]
Generating train split: 29559506 examples [06:39, 96173.07 examples/s]
Generating train split: 29581506 examples [06:39, 116183.37 examples/s]
Generating train split: 29603506 examples [06:40, 126243.98 examples/s]
Generating train split: 29630506 examples [06:40, 155241.17 examples/s]
Generating train split: 29652506 examples [06:40, 152671.19 examples/s]
Generating train split: 29682506 examples [06:40, 183170.69 examples/s]
Generating train split: 29706506 examples [06:40, 115110.37 examples/s]
Generating train split: 29722506 examples [06:41, 81432.09 examples/s]
Generating train split: 29737506 examples [06:41, 89235.60 examples/s]
Generating train split: 29753506 examples [06:42, 52394.53 examples/s]
Generating train split: 29764506 examples [06:42, 35090.96 examples/s]
Generating train split: 29775506 examples [06:42, 38851.96 examples/s]
Generating train split: 29785506 examples [06:43, 29295.50 examples/s]
Generating train split: 29806506 examples [06:43, 44757.85 examples/s]
Generating train split: 29823506 examples [06:43, 50001.34 examples/s]
Generating train split: 29855506 examples [06:44, 80123.81 examples/s]
Generating train split: 29870506 examples [06:44, 70687.39 examples/s]
Generating train split: 29887506 examples [06:44, 74146.02 examples/s]
Generating train split: 29918506 examples [06:44, 89724.40 examples/s]
Generating train split: 29939506 examples [06:44, 105792.13 examples/s]
Generating train split: 29968506 examples [06:45, 135602.01 examples/s]
Generating train split: 29988506 examples [06:45, 144028.76 examples/s]
Generating train split: 30009506 examples [06:45, 130800.40 examples/s]
Generating train split: 30025506 examples [06:45, 116760.84 examples/s]
Generating train split: 30050506 examples [06:45, 137569.60 examples/s]
Generating train split: 30066506 examples [06:46, 85899.08 examples/s]
Generating train split: 30088506 examples [06:46, 105714.20 examples/s]
Generating train split: 30103506 examples [06:46, 109974.59 examples/s]
Generating train split: 30118174 examples [06:47, 45278.12 examples/s]
Generating train split: 30132174 examples [06:47, 36761.14 examples/s]
Generating train split: 30153174 examples [06:47, 49795.19 examples/s]
Generating train split: 30163174 examples [06:48, 51808.62 examples/s]
Generating train split: 30175174 examples [06:48, 49419.28 examples/s]
Generating train split: 30187174 examples [06:48, 54662.41 examples/s]
Generating train split: 30198174 examples [06:48, 60797.73 examples/s]
Generating train split: 30226174 examples [06:48, 93661.60 examples/s]
Generating train split: 30241174 examples [06:48, 94528.71 examples/s]
Generating train split: 30254174 examples [06:49, 95495.86 examples/s]
Generating train split: 30277174 examples [06:49, 106566.38 examples/s]
Generating train split: 30302174 examples [06:49, 128057.54 examples/s]
Generating train split: 30318174 examples [06:49, 98685.83 examples/s]
Generating train split: 30339174 examples [06:49, 113445.62 examples/s]
Generating train split: 30354174 examples [06:49, 119041.54 examples/s]
Generating train split: 30372174 examples [06:50, 112318.86 examples/s]
Generating train split: 30387174 examples [06:50, 102431.18 examples/s]
Generating train split: 30404174 examples [06:50, 67702.05 examples/s]
Generating train split: 30427174 examples [06:50, 89758.03 examples/s]
Generating train split: 30444174 examples [06:51, 47823.09 examples/s]
Generating train split: 30455174 examples [06:51, 52136.84 examples/s]
Generating train split: 30465174 examples [06:51, 54577.35 examples/s]
Generating train split: 30474174 examples [06:52, 36514.97 examples/s]
Generating train split: 30485174 examples [06:52, 41312.13 examples/s]
Generating train split: 30496174 examples [06:52, 49327.27 examples/s]
Generating train split: 30513174 examples [06:52, 62932.01 examples/s]
Generating train split: 30524174 examples [06:53, 57003.81 examples/s]
Generating train split: 30536174 examples [06:53, 59199.15 examples/s]
Generating train split: 30560174 examples [06:53, 86303.21 examples/s]
Generating train split: 30586174 examples [06:53, 107587.58 examples/s]
Generating train split: 30609174 examples [06:53, 128030.51 examples/s]
Generating train split: 30637174 examples [06:53, 153419.64 examples/s]
Generating train split: 30659174 examples [06:54, 119718.23 examples/s]
Generating train split: 30674174 examples [06:54, 106153.59 examples/s]
Generating train split: 30691174 examples [06:54, 111692.14 examples/s]
Generating train split: 30717174 examples [06:54, 135173.55 examples/s]
Generating train split: 30733174 examples [06:54, 99821.67 examples/s]
Generating train split: 30749174 examples [06:55, 93818.64 examples/s]
Generating train split: 30764174 examples [06:55, 70679.13 examples/s]
Generating train split: 30781174 examples [06:55, 80850.01 examples/s]
Generating train split: 30792174 examples [06:55, 56175.68 examples/s]
Generating train split: 30803174 examples [06:56, 42789.09 examples/s]
Generating train split: 30814174 examples [06:56, 40027.58 examples/s]
Generating train split: 30824174 examples [06:57, 34206.00 examples/s]
Generating train split: 30829174 examples [06:57, 32792.97 examples/s]
Generating train split: 30863174 examples [06:57, 68907.27 examples/s]
Generating train split: 30875174 examples [06:57, 55915.14 examples/s]
Generating train split: 30886174 examples [06:57, 62848.55 examples/s]
Generating train split: 30899174 examples [06:58, 70177.54 examples/s]
Generating train split: 30923174 examples [06:58, 95288.61 examples/s]
Generating train split: 30936174 examples [06:58, 89762.44 examples/s]
Generating train split: 30965174 examples [06:58, 125421.93 examples/s]
Generating train split: 30985174 examples [06:58, 135155.85 examples/s]
Generating train split: 31004174 examples [06:58, 128618.14 examples/s]
Generating train split: 31024174 examples [06:58, 121051.22 examples/s]
Generating train split: 31041174 examples [06:59, 87760.05 examples/s]
Generating train split: 31056174 examples [06:59, 97820.07 examples/s]
Generating train split: 31071174 examples [06:59, 104708.15 examples/s]
Generating train split: 31084174 examples [06:59, 109981.38 examples/s]
Generating train split: 31101174 examples [06:59, 87288.97 examples/s]
Generating train split: 31112174 examples [07:00, 66936.85 examples/s]
Generating train split: 31122174 examples [07:00, 48313.28 examples/s]
Generating train split: 31133174 examples [07:01, 35010.56 examples/s]
Generating train split: 31144174 examples [07:01, 34542.43 examples/s]
Generating train split: 31150174 examples [07:01, 29759.18 examples/s]
Generating train split: 31161174 examples [07:01, 34713.91 examples/s]
Generating train split: 31187174 examples [07:02, 62328.75 examples/s]
Generating train split: 31204174 examples [07:02, 74545.74 examples/s]
Generating train split: 31220174 examples [07:02, 84752.57 examples/s]
Generating train split: 31240174 examples [07:02, 98826.50 examples/s]
Generating train split: 31268174 examples [07:02, 117558.13 examples/s]
Generating train split: 31290174 examples [07:02, 133997.66 examples/s]
Generating train split: 31310174 examples [07:03, 112696.94 examples/s]
Generating train split: 31333174 examples [07:03, 121321.73 examples/s]
Generating train split: 31349174 examples [07:03, 93843.92 examples/s]
Generating train split: 31364174 examples [07:03, 69308.33 examples/s]
Generating train split: 31379174 examples [07:03, 80208.82 examples/s]
Generating train split: 31407174 examples [07:04, 96983.61 examples/s]
Generating train split: 31425174 examples [07:04, 105773.87 examples/s]
Generating train split: 31441174 examples [07:05, 49465.61 examples/s]
Generating train split: 31452174 examples [07:05, 37000.36 examples/s]
Generating train split: 31462174 examples [07:05, 42344.45 examples/s]
Generating train split: 31472174 examples [07:06, 38922.99 examples/s]
Generating train split: 31496174 examples [07:06, 59169.68 examples/s]
Generating train split: 31508174 examples [07:06, 59210.46 examples/s]
Generating train split: 31519174 examples [07:06, 60031.14 examples/s]
Generating train split: 31532174 examples [07:06, 55148.65 examples/s]
Generating train split: 31550174 examples [07:07, 72789.16 examples/s]
Generating train split: 31563174 examples [07:07, 81367.93 examples/s]
Generating train split: 31603174 examples [07:07, 143289.11 examples/s]
Generating train split: 31630174 examples [07:07, 147725.18 examples/s]
Generating train split: 31664174 examples [07:07, 177358.02 examples/s]
Generating train split: 31690174 examples [07:07, 156337.59 examples/s]
Generating train split: 31712174 examples [07:08, 125265.18 examples/s]
Generating train split: 31735174 examples [07:08, 83936.36 examples/s]
Generating train split: 31751174 examples [07:08, 92599.35 examples/s]
Generating train split: 31768174 examples [07:08, 97196.62 examples/s]
Generating train split: 31785174 examples [07:09, 69851.20 examples/s]
Generating train split: 31799174 examples [07:09, 50794.69 examples/s]
Generating train split: 31810174 examples [07:10, 38251.33 examples/s]
Generating train split: 31822174 examples [07:10, 42914.22 examples/s]
Generating train split: 31832174 examples [07:10, 46699.23 examples/s]
Generating train split: 31842174 examples [07:10, 53466.11 examples/s]
Generating train split: 31864174 examples [07:10, 76515.42 examples/s]
Generating train split: 31875174 examples [07:11, 56904.63 examples/s]
Generating train split: 31903174 examples [07:11, 89719.93 examples/s]
Generating train split: 31918174 examples [07:11, 73021.29 examples/s]
Generating train split: 31934174 examples [07:11, 72523.86 examples/s]
Generating train split: 31961174 examples [07:11, 102493.56 examples/s]
Generating train split: 31980174 examples [07:12, 110444.25 examples/s]
Generating train split: 32015174 examples [07:12, 145924.03 examples/s]
Generating train split: 32037174 examples [07:12, 123499.65 examples/s]
Generating train split: 32057174 examples [07:12, 100038.20 examples/s]
Generating train split: 32074174 examples [07:13, 94838.70 examples/s]
Generating train split: 32089174 examples [07:13, 84362.02 examples/s]
Generating train split: 32100174 examples [07:13, 86796.25 examples/s]
Generating train split: 32122174 examples [07:13, 66215.51 examples/s]
Generating train split: 32134174 examples [07:14, 44589.08 examples/s]
Generating train split: 32145174 examples [07:14, 34767.68 examples/s]
Generating train split: 32157174 examples [07:15, 40846.21 examples/s]
Generating train split: 32164174 examples [07:15, 37636.07 examples/s]
Generating train split: 32181174 examples [07:15, 53460.86 examples/s]
Generating train split: 32205174 examples [07:15, 79892.36 examples/s]
Generating train split: 32219174 examples [07:15, 68471.80 examples/s]
Generating train split: 32241174 examples [07:15, 88529.47 examples/s]
Generating train split: 32259174 examples [07:16, 71769.46 examples/s]
Generating train split: 32275174 examples [07:16, 75362.32 examples/s]
Generating train split: 32304174 examples [07:16, 98517.38 examples/s]
Generating train split: 32332174 examples [07:16, 127413.69 examples/s]
Generating train split: 32360174 examples [07:17, 131056.44 examples/s]
Generating train split: 32377174 examples [07:17, 110418.25 examples/s]
Generating train split: 32393174 examples [07:17, 90124.75 examples/s]
Generating train split: 32405174 examples [07:17, 78311.88 examples/s]
Generating train split: 32421174 examples [07:17, 84409.99 examples/s]
Generating train split: 32436174 examples [07:18, 94403.30 examples/s]
Generating train split: 32452174 examples [07:18, 65652.96 examples/s]
Generating train split: 32464174 examples [07:18, 46158.57 examples/s]
Generating train split: 32475174 examples [07:19, 34912.07 examples/s]
Generating train split: 32494174 examples [07:19, 48005.43 examples/s]
Generating train split: 32506174 examples [07:19, 55662.88 examples/s]
Generating train split: 32516174 examples [07:19, 54947.48 examples/s]
Generating train split: 32532174 examples [07:20, 68477.38 examples/s]
Generating train split: 32549174 examples [07:20, 84632.33 examples/s]
Generating train split: 32565174 examples [07:20, 75340.02 examples/s]
Generating train split: 32587174 examples [07:20, 99430.42 examples/s]
Generating train split: 32609174 examples [07:20, 121005.72 examples/s]
Generating train split: 32627174 examples [07:20, 97905.92 examples/s]
Generating train split: 32653174 examples [07:21, 112757.09 examples/s]
Generating train split: 32686174 examples [07:21, 99281.93 examples/s]
Generating train split: 32703174 examples [07:21, 101833.97 examples/s]
Generating train split: 32719174 examples [07:21, 100704.13 examples/s]
Generating train split: 32732174 examples [07:21, 98439.17 examples/s]
Generating train split: 32755174 examples [07:22, 121309.76 examples/s]
Generating train split: 32772174 examples [07:22, 70067.66 examples/s]
Generating train split: 32786174 examples [07:22, 75315.96 examples/s]
Generating train split: 32797174 examples [07:23, 36689.36 examples/s]
Generating train split: 32813174 examples [07:23, 46714.17 examples/s]
Generating train split: 32823174 examples [07:23, 52331.70 examples/s]
Generating train split: 32835174 examples [07:24, 39299.44 examples/s]
Generating train split: 32845174 examples [07:24, 39853.46 examples/s]
Generating train split: 32872174 examples [07:24, 66662.76 examples/s]
Generating train split: 32898174 examples [07:24, 87722.75 examples/s]
Generating train split: 32915174 examples [07:25, 84374.43 examples/s]
Generating train split: 32930174 examples [07:25, 92479.98 examples/s]
Generating train split: 32960174 examples [07:25, 123750.33 examples/s]
Generating train split: 32978174 examples [07:25, 106689.87 examples/s]
Generating train split: 33000174 examples [07:25, 125141.67 examples/s]
Generating train split: 33020174 examples [07:25, 102754.60 examples/s]
Generating train split: 33037174 examples [07:26, 84531.18 examples/s]
Generating train split: 33048174 examples [07:26, 85492.16 examples/s]
Generating train split: 33064174 examples [07:26, 72954.94 examples/s]
Generating train split: 33075174 examples [07:26, 77880.61 examples/s]
Generating train split: 33097174 examples [07:27, 70842.71 examples/s]
Generating train split: 33111174 examples [07:27, 80543.56 examples/s]
Generating train split: 33122174 examples [07:27, 53388.68 examples/s]
Generating train split: 33131174 examples [07:27, 57540.66 examples/s]
Generating train split: 33144174 examples [07:28, 37170.62 examples/s]
Generating train split: 33155174 examples [07:28, 31792.10 examples/s]
Generating train split: 33187174 examples [07:29, 57418.05 examples/s]
Generating train split: 33198174 examples [07:29, 61744.27 examples/s]
Generating train split: 33210174 examples [07:29, 69158.94 examples/s]
Generating train split: 33242174 examples [07:29, 89889.96 examples/s]
Generating train split: 33257174 examples [07:29, 85532.08 examples/s]
Generating train split: 33279174 examples [07:29, 104338.09 examples/s]
Generating train split: 33295174 examples [07:29, 112393.61 examples/s]
Generating train split: 33312174 examples [07:30, 93430.97 examples/s]
Generating train split: 33334174 examples [07:30, 107887.54 examples/s]
Generating train split: 33351174 examples [07:30, 78814.07 examples/s]
Generating train split: 33366174 examples [07:30, 89535.92 examples/s]
Generating train split: 33388174 examples [07:31, 102759.92 examples/s]
Generating train split: 33407174 examples [07:31, 85036.57 examples/s]
Generating train split: 33444174 examples [07:31, 128943.87 examples/s]
Generating train split: 33466174 examples [07:32, 52944.19 examples/s]
Generating train split: 33484174 examples [07:32, 49363.30 examples/s]
Generating train split: 33495174 examples [07:33, 49165.98 examples/s]
Generating train split: 33507174 examples [07:33, 51665.63 examples/s]
Generating train split: 33535174 examples [07:33, 76098.91 examples/s]
Generating train split: 33552174 examples [07:33, 89110.86 examples/s]
Generating train split: 33568174 examples [07:33, 99946.85 examples/s]
Generating train split: 33583174 examples [07:33, 98417.53 examples/s]
Generating train split: 33609174 examples [07:33, 126833.38 examples/s]
Generating train split: 33629174 examples [07:34, 116293.46 examples/s]
Generating train split: 33647174 examples [07:34, 97474.95 examples/s]
Generating train split: 33659174 examples [07:34, 86506.58 examples/s]
Generating train split: 33672174 examples [07:34, 71669.09 examples/s]
Generating train split: 33707174 examples [07:35, 109547.61 examples/s]
Generating train split: 33722174 examples [07:35, 75202.07 examples/s]
Generating train split: 33746174 examples [07:35, 97923.68 examples/s]
Generating train split: 33761174 examples [07:35, 69276.14 examples/s]
Generating train split: 33778174 examples [07:36, 81573.47 examples/s]
Generating train split: 33793174 examples [07:37, 38071.60 examples/s]
Generating train split: 33809174 examples [07:37, 42211.42 examples/s]
Generating train split: 33827174 examples [07:37, 53753.94 examples/s]
Generating train split: 33838174 examples [07:37, 55837.31 examples/s]
Generating train split: 33850174 examples [07:37, 58499.45 examples/s]
Generating train split: 33860174 examples [07:37, 62147.77 examples/s]
Generating train split: 33879174 examples [07:38, 81634.78 examples/s]
Generating train split: 33901174 examples [07:38, 105051.22 examples/s]
Generating train split: 33918174 examples [07:38, 101281.36 examples/s]
Generating train split: 33933174 examples [07:38, 106226.68 examples/s]
Generating train split: 33947174 examples [07:38, 106646.71 examples/s]
Generating train split: 33963174 examples [07:38, 85356.65 examples/s]
Generating train split: 33985174 examples [07:38, 110507.64 examples/s]
Generating train split: 34007174 examples [07:39, 120149.38 examples/s]
Generating train split: 34023174 examples [07:39, 83973.84 examples/s]
Generating train split: 34040174 examples [07:39, 90437.82 examples/s]
Generating train split: 34055174 examples [07:39, 92626.64 examples/s]
Generating train split: 34067174 examples [07:40, 74347.88 examples/s]
Generating train split: 34106174 examples [07:40, 127071.88 examples/s]
Generating train split: 34128174 examples [07:41, 43736.05 examples/s]
Generating train split: 34145174 examples [07:41, 39932.73 examples/s]
Generating train split: 34156174 examples [07:42, 42617.80 examples/s]
Generating train split: 34173174 examples [07:42, 53307.27 examples/s]
Generating train split: 34194174 examples [07:42, 70737.55 examples/s]
Generating train split: 34209174 examples [07:42, 73439.24 examples/s]
Generating train split: 34232174 examples [07:42, 89877.79 examples/s]
Generating train split: 34248174 examples [07:43, 74576.54 examples/s]
Generating train split: 34264174 examples [07:43, 82089.69 examples/s]
Generating train split: 34275174 examples [07:43, 79968.13 examples/s]
Generating train split: 34292174 examples [07:43, 90912.84 examples/s]
Generating train split: 34314174 examples [07:43, 115597.65 examples/s]
Generating train split: 34330174 examples [07:43, 124219.47 examples/s]
Generating train split: 34345174 examples [07:44, 69164.90 examples/s]
Generating train split: 34366174 examples [07:44, 85025.13 examples/s]
Generating train split: 34389174 examples [07:44, 107136.49 examples/s]
Generating train split: 34404174 examples [07:44, 89865.87 examples/s]
Generating train split: 34420174 examples [07:44, 97127.74 examples/s]
Generating train split: 34437174 examples [07:45, 38288.65 examples/s]
Generating train split: 34449174 examples [07:46, 40042.46 examples/s]
Generating train split: 34460174 examples [07:46, 45715.78 examples/s]
Generating train split: 34478174 examples [07:46, 58126.36 examples/s]
Generating train split: 34504174 examples [07:46, 84570.15 examples/s]
Generating train split: 34521174 examples [07:46, 78057.34 examples/s]
Generating train split: 34544174 examples [07:46, 96866.16 examples/s]
Generating train split: 34561174 examples [07:47, 92912.64 examples/s]
Generating train split: 34595174 examples [07:47, 125321.47 examples/s]
Generating train split: 34613174 examples [07:47, 98617.22 examples/s]
Generating train split: 34637174 examples [07:47, 118519.80 examples/s]
Generating train split: 34653174 examples [07:48, 86271.61 examples/s]
Generating train split: 34669174 examples [07:48, 95075.80 examples/s]
Generating train split: 34686174 examples [07:48, 88785.46 examples/s]
Generating train split: 34701174 examples [07:48, 67485.86 examples/s]
Generating train split: 34719174 examples [07:48, 82551.31 examples/s]
Generating train split: 34735174 examples [07:49, 73727.94 examples/s]
Generating train split: 34746174 examples [07:49, 68935.81 examples/s]
Generating train split: 34761174 examples [07:49, 81020.09 examples/s]
Generating train split: 34772174 examples [07:50, 41040.78 examples/s]
Generating train split: 34783174 examples [07:50, 39721.38 examples/s]
Generating train split: 34801174 examples [07:50, 53436.66 examples/s]
Generating train split: 34810174 examples [07:50, 47645.78 examples/s]
Generating train split: 34827174 examples [07:50, 62995.60 examples/s]
Generating train split: 34863174 examples [07:51, 105103.59 examples/s]
Generating train split: 34878174 examples [07:51, 94537.24 examples/s]
Generating train split: 34903174 examples [07:51, 114524.55 examples/s]
Generating train split: 34919174 examples [07:51, 92306.86 examples/s]
Generating train split: 34935174 examples [07:51, 101247.01 examples/s]
Generating train split: 34951174 examples [07:52, 84060.31 examples/s]
Generating train split: 34968174 examples [07:52, 97503.07 examples/s]
Generating train split: 34985174 examples [07:52, 103987.33 examples/s]
Generating train split: 35002174 examples [07:52, 79401.17 examples/s]
Generating train split: 35013174 examples [07:52, 73468.14 examples/s]
Generating train split: 35023174 examples [07:53, 74772.76 examples/s]
Generating train split: 35036174 examples [07:53, 59647.22 examples/s]
Generating train split: 35059174 examples [07:53, 86505.34 examples/s]
Generating train split: 35087174 examples [07:53, 112431.12 examples/s]
Generating train split: 35103174 examples [07:54, 39997.40 examples/s]
Generating train split: 35114174 examples [07:54, 42639.87 examples/s]
Generating train split: 35125174 examples [07:55, 45715.25 examples/s]
Generating train split: 35147174 examples [07:55, 65202.71 examples/s]
Generating train split: 35163174 examples [07:55, 63397.00 examples/s]
Generating train split: 35180174 examples [07:55, 68464.35 examples/s]
Generating train split: 35209174 examples [07:55, 97166.60 examples/s]
Generating train split: 35226174 examples [07:55, 105198.82 examples/s]
Generating train split: 35255174 examples [07:56, 91805.20 examples/s]
Generating train split: 35272174 examples [07:56, 102047.75 examples/s]
Generating train split: 35288174 examples [07:56, 74361.48 examples/s]
Generating train split: 35316174 examples [07:56, 100674.87 examples/s]
Generating train split: 35333174 examples [07:57, 97730.99 examples/s]
Generating train split: 35351174 examples [07:57, 82042.03 examples/s]
Generating train split: 35367174 examples [07:57, 87449.09 examples/s]
Generating train split: 35381174 examples [07:57, 65490.46 examples/s]
Generating train split: 35392174 examples [07:58, 70755.12 examples/s]
Generating train split: 35403174 examples [07:58, 71856.20 examples/s]
Generating train split: 35414174 examples [07:58, 71022.37 examples/s]
Generating train split: 35426174 examples [07:58, 79446.33 examples/s]
Generating train split: 35437174 examples [07:58, 54815.28 examples/s]
Generating train split: 35449174 examples [07:59, 31075.78 examples/s]
Generating train split: 35460174 examples [07:59, 37743.03 examples/s]
Generating train split: 35477174 examples [07:59, 52268.61 examples/s]
Generating train split: 35494174 examples [08:00, 55381.73 examples/s]
Generating train split: 35511174 examples [08:00, 66931.11 examples/s]
Generating train split: 35523174 examples [08:00, 70971.92 examples/s]
Generating train split: 35540174 examples [08:00, 87114.32 examples/s]
Generating train split: 35583174 examples [08:00, 154639.35 examples/s]
Generating train split: 35604174 examples [08:01, 98002.07 examples/s]
Generating train split: 35622174 examples [08:01, 100897.64 examples/s]
Generating train split: 35637174 examples [08:01, 85634.15 examples/s]
Generating train split: 35669174 examples [08:01, 89795.43 examples/s]
Generating train split: 35685174 examples [08:02, 81777.08 examples/s]
Generating train split: 35703174 examples [08:02, 95587.55 examples/s]
Generating train split: 35720174 examples [08:02, 99046.58 examples/s]
Generating train split: 35732174 examples [08:02, 62106.82 examples/s]
Generating train split: 35742174 examples [08:03, 59918.05 examples/s]
Generating train split: 35754174 examples [08:03, 44369.05 examples/s]
Generating train split: 35765174 examples [08:03, 39158.60 examples/s]
Generating train split: 35776174 examples [08:04, 34072.51 examples/s]
Generating train split: 35787174 examples [08:04, 41372.78 examples/s]
Generating train split: 35799174 examples [08:04, 50659.27 examples/s]
Generating train split: 35833174 examples [08:04, 81502.42 examples/s]
Generating train split: 35864174 examples [08:04, 115670.61 examples/s]
Generating train split: 35880174 examples [08:05, 108212.80 examples/s]
Generating train split: 35901174 examples [08:05, 126232.88 examples/s]
Generating train split: 35944174 examples [08:05, 176788.12 examples/s]
Generating train split: 35965174 examples [08:05, 109613.76 examples/s]
Generating train split: 35984174 examples [08:06, 76482.30 examples/s]
Generating train split: 36014174 examples [08:06, 102509.28 examples/s]
Generating train split: 36035174 examples [08:06, 77457.40 examples/s]
Generating train split: 36053174 examples [08:06, 80632.81 examples/s]
Generating train split: 36066174 examples [08:07, 54180.23 examples/s]
Generating train split: 36076174 examples [08:07, 58598.78 examples/s]
Generating train split: 36087174 examples [08:07, 60920.55 examples/s]
Generating train split: 36099174 examples [08:07, 63787.81 examples/s]
Generating train split: 36110174 examples [08:08, 56670.77 examples/s]
Generating train split: 36121174 examples [08:08, 57158.49 examples/s]
Generating train split: 36131174 examples [08:08, 36134.40 examples/s]
Generating train split: 36142174 examples [08:09, 43718.15 examples/s]
Generating train split: 36161174 examples [08:09, 63704.69 examples/s]
Generating train split: 36187174 examples [08:09, 88414.11 examples/s]
Generating train split: 36209174 examples [08:09, 94538.86 examples/s]
Generating train split: 36224174 examples [08:09, 94999.06 examples/s]
Generating train split: 36254174 examples [08:09, 130457.10 examples/s]
Generating train split: 36273174 examples [08:09, 139778.84 examples/s]
Generating train split: 36295174 examples [08:10, 71589.87 examples/s]
Generating train split: 36311174 examples [08:10, 73477.75 examples/s]
Generating train split: 36340174 examples [08:10, 102921.85 examples/s]
Generating train split: 36368174 examples [08:11, 104434.29 examples/s]
Generating train split: 36384174 examples [08:11, 88776.36 examples/s]
Generating train split: 36405174 examples [08:11, 73228.48 examples/s]
Generating train split: 36417174 examples [08:12, 61684.82 examples/s]
Generating train split: 36438174 examples [08:12, 58216.51 examples/s]
Generating train split: 36448174 examples [08:12, 51778.68 examples/s]
Generating train split: 36459174 examples [08:13, 44700.47 examples/s]
Generating train split: 36471174 examples [08:13, 51927.80 examples/s]
Generating train split: 36483174 examples [08:13, 55602.55 examples/s]
Generating train split: 36500174 examples [08:13, 63308.47 examples/s]
Generating train split: 36524174 examples [08:13, 89934.40 examples/s]
Generating train split: 36551174 examples [08:13, 120583.41 examples/s]
Generating train split: 36585174 examples [08:13, 158886.90 examples/s]
Generating train split: 36605174 examples [08:14, 96742.01 examples/s]
Generating train split: 36622174 examples [08:14, 83441.94 examples/s]
Generating train split: 36642174 examples [08:14, 95683.61 examples/s]
Generating train split: 36659174 examples [08:15, 88811.03 examples/s]
Generating train split: 36682174 examples [08:15, 84821.90 examples/s]
Generating train split: 36710174 examples [08:15, 66005.24 examples/s]
Generating train split: 36722174 examples [08:16, 71193.93 examples/s]
Generating train split: 36733174 examples [08:16, 59437.27 examples/s]
Generating train split: 36743174 examples [08:16, 64516.12 examples/s]
Generating train split: 36753174 examples [08:16, 56235.13 examples/s]
Generating train split: 36770174 examples [08:17, 51194.43 examples/s]
Generating train split: 36782174 examples [08:17, 46075.56 examples/s]
Generating train split: 36798174 examples [08:17, 57922.04 examples/s]
Generating train split: 36810174 examples [08:17, 63877.62 examples/s]
Generating train split: 36825174 examples [08:17, 71018.96 examples/s]
Generating train split: 36836174 examples [08:18, 66413.47 examples/s]
Generating train split: 36849174 examples [08:18, 60466.44 examples/s]
Generating train split: 36882174 examples [08:18, 102208.59 examples/s]
Generating train split: 36915174 examples [08:18, 89869.10 examples/s]
Generating train split: 36931174 examples [08:19, 94148.32 examples/s]
Generating train split: 36944174 examples [08:19, 83865.72 examples/s]
Generating train split: 36963174 examples [08:19, 96846.76 examples/s]
Generating train split: 36980174 examples [08:19, 80730.79 examples/s]
Generating train split: 37002174 examples [08:19, 83311.29 examples/s]
Generating train split: 37025174 examples [08:20, 105018.36 examples/s]
Generating train split: 37047174 examples [08:20, 72985.61 examples/s]
Generating train split: 37061174 examples [08:20, 77841.34 examples/s]
Generating train split: 37074174 examples [08:20, 82552.14 examples/s]
Generating train split: 37085174 examples [08:21, 62156.16 examples/s]
Generating train split: 37101174 examples [08:21, 55135.47 examples/s]
Generating train split: 37110174 examples [08:21, 58602.57 examples/s]
Generating train split: 37120174 examples [08:22, 36571.48 examples/s]
Generating train split: 37140174 examples [08:22, 53505.64 examples/s]
Generating train split: 37153174 examples [08:22, 60826.07 examples/s]
Generating train split: 37177174 examples [08:22, 84441.56 examples/s]
Generating train split: 37189174 examples [08:22, 86730.92 examples/s]
Generating train split: 37205174 examples [08:23, 75047.66 examples/s]
Generating train split: 37231174 examples [08:23, 99756.50 examples/s]
Generating train split: 37257174 examples [08:23, 127803.85 examples/s]
Generating train split: 37273174 examples [08:23, 129112.19 examples/s]
Generating train split: 37290174 examples [08:23, 128324.92 examples/s]
Generating train split: 37313174 examples [08:23, 146043.52 examples/s]
Generating train split: 37330174 examples [08:24, 66491.07 examples/s]
Generating train split: 37347174 examples [08:24, 68446.81 examples/s]
Generating train split: 37370174 examples [08:24, 89002.77 examples/s]
Generating train split: 37389174 examples [08:25, 55007.49 examples/s]
Generating train split: 37405174 examples [08:25, 62430.69 examples/s]
Generating train split: 37419174 examples [08:25, 61756.10 examples/s]
Generating train split: 37430174 examples [08:25, 67163.32 examples/s]
Generating train split: 37440174 examples [08:26, 45979.96 examples/s]
Generating train split: 37453174 examples [08:26, 34610.50 examples/s]
Generating train split: 37461174 examples [08:26, 38681.88 examples/s]
Generating train split: 37471174 examples [08:27, 44180.31 examples/s]
Generating train split: 37502174 examples [08:27, 81980.54 examples/s]
Generating train split: 37516174 examples [08:27, 88568.94 examples/s]
Generating train split: 37539174 examples [08:27, 107138.89 examples/s]
Generating train split: 37554174 examples [08:27, 105108.52 examples/s]
Generating train split: 37571174 examples [08:27, 90497.02 examples/s]
Generating train split: 37585174 examples [08:27, 99232.03 examples/s]
Generating train split: 37601174 examples [08:28, 100454.48 examples/s]
Generating train split: 37620174 examples [08:28, 118659.40 examples/s]
Generating train split: 37642174 examples [08:28, 90649.00 examples/s]
Generating train split: 37664174 examples [08:28, 112571.47 examples/s]
Generating train split: 37680174 examples [08:28, 113833.66 examples/s]
Generating train split: 37707174 examples [08:28, 141879.73 examples/s]
Generating train split: 37728174 examples [08:29, 59558.59 examples/s]
Generating train split: 37745174 examples [08:29, 70340.94 examples/s]
Generating train split: 37762174 examples [08:30, 68608.00 examples/s]
Generating train split: 37776174 examples [08:30, 42133.76 examples/s]
Generating train split: 37801174 examples [08:30, 61286.60 examples/s]
Generating train split: 37817174 examples [08:31, 62345.96 examples/s]
Generating train split: 37834174 examples [08:31, 62083.98 examples/s]
Generating train split: 37844174 examples [08:31, 62456.19 examples/s]
Generating train split: 37855174 examples [08:31, 67561.71 examples/s]
Generating train split: 37885174 examples [08:31, 104840.48 examples/s]
Generating train split: 37909174 examples [08:32, 99631.54 examples/s]
Generating train split: 37926174 examples [08:32, 89191.79 examples/s]
Generating train split: 37941174 examples [08:32, 87194.15 examples/s]
Generating train split: 37953174 examples [08:32, 76667.75 examples/s]
Generating train split: 37964174 examples [08:32, 77615.56 examples/s]
Generating train split: 37973174 examples [08:33, 65725.13 examples/s]
Generating train split: 37997174 examples [08:33, 79669.13 examples/s]
Generating train split: 38008174 examples [08:33, 83166.42 examples/s]
Generating train split: 38019174 examples [08:33, 57577.20 examples/s]
Generating train split: 38032174 examples [08:33, 66009.09 examples/s]
Generating train split: 38048174 examples [08:34, 77377.79 examples/s]
Generating train split: 38058174 examples [08:34, 68364.83 examples/s]
Generating train split: 38075174 examples [08:34, 50768.77 examples/s]
Generating train split: 38086174 examples [08:34, 57306.14 examples/s]
Generating train split: 38097174 examples [08:35, 41812.40 examples/s]
Generating train split: 38108174 examples [08:35, 45895.30 examples/s]
Generating train split: 38116174 examples [08:35, 43385.32 examples/s]
Generating train split: 38133174 examples [08:35, 61810.92 examples/s]
Generating train split: 38152174 examples [08:35, 81860.38 examples/s]
Generating train split: 38169174 examples [08:36, 90775.91 examples/s]
Generating train split: 38184174 examples [08:36, 93873.69 examples/s]
Generating train split: 38204174 examples [08:36, 107070.63 examples/s]
Generating train split: 38220174 examples [08:36, 118390.18 examples/s]
Generating train split: 38251174 examples [08:36, 162038.57 examples/s]
Generating train split: 38274174 examples [08:37, 92792.95 examples/s]
Generating train split: 38291174 examples [08:37, 85758.59 examples/s]
Generating train split: 38307174 examples [08:37, 93988.71 examples/s]
Generating train split: 38323174 examples [08:37, 77853.62 examples/s]
Generating train split: 38340174 examples [08:37, 91133.18 examples/s]
Generating train split: 38355174 examples [08:38, 86184.67 examples/s]
Generating train split: 38366174 examples [08:38, 61077.91 examples/s]
Generating train split: 38378174 examples [08:38, 63007.57 examples/s]
Generating train split: 38390174 examples [08:38, 53677.35 examples/s]
Generating train split: 38406174 examples [08:39, 67380.90 examples/s]
Generating train split: 38416174 examples [08:39, 44826.14 examples/s]
Generating train split: 38432174 examples [08:40, 38352.47 examples/s]
Generating train split: 38464174 examples [08:40, 64083.49 examples/s]
Generating train split: 38476174 examples [08:40, 62573.07 examples/s]
Generating train split: 38494174 examples [08:40, 75937.31 examples/s]
Generating train split: 38518174 examples [08:40, 101370.80 examples/s]
Generating train split: 38535174 examples [08:40, 89456.34 examples/s]
Generating train split: 38554174 examples [08:41, 103382.56 examples/s]
Generating train split: 38571174 examples [08:41, 100640.08 examples/s]
Generating train split: 38592174 examples [08:41, 99780.38 examples/s]
Generating train split: 38609174 examples [08:41, 105040.97 examples/s]
Generating train split: 38624174 examples [08:41, 85052.93 examples/s]
Generating train split: 38635174 examples [08:41, 85758.86 examples/s]
Generating train split: 38648174 examples [08:42, 91097.28 examples/s]
Generating train split: 38665174 examples [08:42, 97507.31 examples/s]
Generating train split: 38686174 examples [08:42, 57945.54 examples/s]
Generating train split: 38697174 examples [08:43, 55024.57 examples/s]
Generating train split: 38713174 examples [08:43, 64376.81 examples/s]
Generating train split: 38723174 examples [08:43, 44486.13 examples/s]
Generating train split: 38733174 examples [08:43, 50400.58 examples/s]
Generating train split: 38744174 examples [08:44, 42561.81 examples/s]
Generating train split: 38771174 examples [08:44, 43403.65 examples/s]
Generating train split: 38797174 examples [08:44, 64760.82 examples/s]
Generating train split: 38812174 examples [08:45, 73751.11 examples/s]
Generating train split: 38834174 examples [08:45, 94567.16 examples/s]
Generating train split: 38854174 examples [08:45, 112565.49 examples/s]
Generating train split: 38882174 examples [08:45, 136596.71 examples/s]
Generating train split: 38903174 examples [08:45, 95889.61 examples/s]
Generating train split: 38919174 examples [08:45, 88839.06 examples/s]
Generating train split: 38941174 examples [08:46, 93901.88 examples/s]
Generating train split: 38956174 examples [08:46, 70770.47 examples/s]
Generating train split: 38972174 examples [08:46, 82596.02 examples/s]
Generating train split: 38988174 examples [08:46, 92519.14 examples/s]
Generating train split: 39005174 examples [08:46, 95286.67 examples/s]
Generating train split: 39020174 examples [08:47, 85701.59 examples/s]
Generating train split: 39030174 examples [08:47, 88203.20 examples/s]
Generating train split: 39041174 examples [08:47, 66017.70 examples/s]
Generating train split: 39053174 examples [08:48, 37291.08 examples/s]
Generating train split: 39065174 examples [08:48, 45436.88 examples/s]
Generating train split: 39082174 examples [08:48, 40906.13 examples/s]
Generating train split: 39111174 examples [08:48, 65160.32 examples/s]
Generating train split: 39122174 examples [08:49, 67314.86 examples/s]
Generating train split: 39133174 examples [08:49, 50773.78 examples/s]
Generating train split: 39149174 examples [08:49, 64545.13 examples/s]
Generating train split: 39173174 examples [08:49, 90475.09 examples/s]
Generating train split: 39213174 examples [08:49, 140956.24 examples/s]
Generating train split: 39236174 examples [08:50, 149181.52 examples/s]
Generating train split: 39257174 examples [08:50, 137906.22 examples/s]
Generating train split: 39274174 examples [08:50, 97245.94 examples/s]
Generating train split: 39290174 examples [08:50, 84666.21 examples/s]
Generating train split: 39305174 examples [08:50, 83302.44 examples/s]
Generating train split: 39317174 examples [08:51, 82761.38 examples/s]
Generating train split: 39328174 examples [08:51, 63611.73 examples/s]
Generating train split: 39338174 examples [08:51, 68619.64 examples/s]
Generating train split: 39354174 examples [08:51, 74221.30 examples/s]
Generating train split: 39365174 examples [08:51, 78709.63 examples/s]
Generating train split: 39376174 examples [08:52, 30347.89 examples/s]
Generating train split: 39391174 examples [08:52, 40450.21 examples/s]
Generating train split: 39402174 examples [08:53, 34665.21 examples/s]
Generating train split: 39429174 examples [08:53, 59448.31 examples/s]
Generating train split: 39446174 examples [08:53, 64618.10 examples/s]
Generating train split: 39469174 examples [08:54, 66563.00 examples/s]
Generating train split: 39483174 examples [08:54, 75938.03 examples/s]
Generating train split: 39495174 examples [08:54, 69503.77 examples/s]
Generating train split: 39539174 examples [08:54, 124395.80 examples/s]
Generating train split: 39561174 examples [08:54, 139201.85 examples/s]
Generating train split: 39583174 examples [08:54, 119558.75 examples/s]
Generating train split: 39599174 examples [08:55, 113463.90 examples/s]
Generating train split: 39613174 examples [08:55, 84149.41 examples/s]
Generating train split: 39631174 examples [08:55, 86863.27 examples/s]
Generating train split: 39653174 examples [08:55, 101520.60 examples/s]
Generating train split: 39671174 examples [08:55, 112370.05 examples/s]
Generating train split: 39688174 examples [08:56, 78272.80 examples/s]
Generating train split: 39700174 examples [08:56, 64309.79 examples/s]
Generating train split: 39717174 examples [08:56, 52795.32 examples/s]
Generating train split: 39727174 examples [08:57, 58253.56 examples/s]
Generating train split: 39739174 examples [08:57, 44739.74 examples/s]
Generating train split: 39754174 examples [08:57, 56780.77 examples/s]
Generating train split: 39777174 examples [08:58, 55828.57 examples/s]
Generating train split: 39790174 examples [08:58, 63350.41 examples/s]
Generating train split: 39802174 examples [08:58, 64623.79 examples/s]
Generating train split: 39817174 examples [08:58, 77395.80 examples/s]
Generating train split: 39857174 examples [08:58, 136516.47 examples/s]
Generating train split: 39880174 examples [08:58, 128587.79 examples/s]
Generating train split: 39898174 examples [08:58, 107395.42 examples/s]
Generating train split: 39912174 examples [08:59, 107741.06 examples/s]
Generating train split: 39933174 examples [08:59, 95299.94 examples/s]
Generating train split: 39945174 examples [08:59, 94311.67 examples/s]
Generating train split: 39956174 examples [08:59, 75395.94 examples/s]
Generating train split: 39973174 examples [08:59, 85183.23 examples/s]
Generating train split: 39983174 examples [09:00, 85790.08 examples/s]
Generating train split: 39994174 examples [09:00, 87159.93 examples/s]
Generating train split: 40019174 examples [09:00, 116502.05 examples/s]
Generating train split: 40035174 examples [09:00, 88710.49 examples/s]
Generating train split: 40048174 examples [09:01, 35197.96 examples/s]
Generating train split: 40066174 examples [09:01, 37299.63 examples/s]
Generating train split: 40077174 examples [09:02, 43605.59 examples/s]
Generating train split: 40100174 examples [09:02, 61617.67 examples/s]
Generating train split: 40114174 examples [09:02, 61313.30 examples/s]
Generating train split: 40125174 examples [09:02, 59564.85 examples/s]
Generating train split: 40141174 examples [09:02, 73959.78 examples/s]
Generating train split: 40152174 examples [09:03, 64058.68 examples/s]
Generating train split: 40180174 examples [09:03, 98838.97 examples/s]
Generating train split: 40214174 examples [09:03, 142889.80 examples/s]
Generating train split: 40237174 examples [09:03, 159449.10 examples/s]
Generating train split: 40261174 examples [09:03, 133399.56 examples/s]
Generating train split: 40278174 examples [09:03, 99572.00 examples/s]
Generating train split: 40295174 examples [09:04, 83259.85 examples/s]
Generating train split: 40308174 examples [09:04, 81510.04 examples/s]
Generating train split: 40319174 examples [09:04, 60082.06 examples/s]
Generating train split: 40332174 examples [09:04, 65241.54 examples/s]
Generating train split: 40342174 examples [09:05, 62157.57 examples/s]
Generating train split: 40358174 examples [09:06, 31341.52 examples/s]
Generating train split: 40375174 examples [09:06, 41864.20 examples/s]
Generating train split: 40387174 examples [09:06, 35819.28 examples/s]
Generating train split: 40410174 examples [09:06, 54347.48 examples/s]
Generating train split: 40421174 examples [09:06, 58692.97 examples/s]
Generating train split: 40432174 examples [09:07, 61408.29 examples/s]
Generating train split: 40459174 examples [09:07, 92048.88 examples/s]
Generating train split: 40498174 examples [09:07, 141052.30 examples/s]
Generating train split: 40522174 examples [09:07, 102224.45 examples/s]
Generating train split: 40538174 examples [09:07, 105863.22 examples/s]
Generating train split: 40560174 examples [09:07, 124628.31 examples/s]
Generating train split: 40582174 examples [09:08, 116469.75 examples/s]
Generating train split: 40599174 examples [09:08, 124761.95 examples/s]
Generating train split: 40620174 examples [09:08, 140042.05 examples/s]
Generating train split: 40637174 examples [09:08, 103820.79 examples/s]
Generating train split: 40653174 examples [09:08, 109099.78 examples/s]
Generating train split: 40675174 examples [09:08, 121274.25 examples/s]
Generating train split: 40690174 examples [09:09, 76588.31 examples/s]
Generating train split: 40704174 examples [09:10, 37859.50 examples/s]
Generating train split: 40716174 examples [09:10, 36619.20 examples/s]
Generating train split: 40727174 examples [09:11, 30525.23 examples/s]
Generating train split: 40748174 examples [09:11, 45546.37 examples/s]
Generating train split: 40758174 examples [09:11, 51027.44 examples/s]
Generating train split: 40789174 examples [09:11, 81371.28 examples/s]
Generating train split: 40803174 examples [09:11, 83906.35 examples/s]
Generating train split: 40836174 examples [09:11, 124801.09 examples/s]
Generating train split: 40869174 examples [09:11, 154562.67 examples/s]
Generating train split: 40901174 examples [09:12, 187470.81 examples/s]
Generating train split: 40925174 examples [09:12, 167059.47 examples/s]
Generating train split: 40946174 examples [09:12, 114994.20 examples/s]
Generating train split: 40963174 examples [09:12, 108026.22 examples/s]
Generating train split: 40982174 examples [09:12, 113387.24 examples/s]
Generating train split: 41001174 examples [09:13, 78341.32 examples/s]
Generating train split: 41018174 examples [09:13, 53574.94 examples/s]
Generating train split: 41030174 examples [09:14, 49789.05 examples/s]
Generating train split: 41040174 examples [09:14, 54628.98 examples/s]
Generating train split: 41050174 examples [09:14, 38027.02 examples/s]
Generating train split: 41056174 examples [09:15, 32307.50 examples/s]
Generating train split: 41080174 examples [09:15, 54694.27 examples/s]
Generating train split: 41091174 examples [09:15, 59098.47 examples/s]
Generating train split: 41103174 examples [09:15, 62805.58 examples/s]
Generating train split: 41125174 examples [09:15, 71475.77 examples/s]
Generating train split: 41136174 examples [09:16, 69619.85 examples/s]
Generating train split: 41155174 examples [09:16, 74879.45 examples/s]
Generating train split: 41191174 examples [09:16, 116130.28 examples/s]
Generating train split: 41208174 examples [09:16, 115510.99 examples/s]
Generating train split: 41251174 examples [09:16, 165950.35 examples/s]
Generating train split: 41273174 examples [09:16, 153740.45 examples/s]
Generating train split: 41300174 examples [09:17, 166826.72 examples/s]
Generating train split: 41322174 examples [09:17, 138310.63 examples/s]
Generating train split: 41340174 examples [09:17, 112875.36 examples/s]
Generating train split: 41355174 examples [09:17, 71017.61 examples/s]
Generating train split: 41367174 examples [09:18, 36467.95 examples/s]
Generating train split: 41379174 examples [09:19, 31680.54 examples/s]
Generating train split: 41390174 examples [09:19, 30992.09 examples/s]
Generating train split: 41411174 examples [09:20, 45785.51 examples/s]
Generating train split: 41427174 examples [09:20, 56125.05 examples/s]
Generating train split: 41445174 examples [09:20, 71345.20 examples/s]
Generating train split: 41472174 examples [09:20, 100754.68 examples/s]
Generating train split: 41493174 examples [09:20, 103490.60 examples/s]
Generating train split: 41521174 examples [09:20, 135206.61 examples/s]
Generating train split: 41544174 examples [09:20, 150713.59 examples/s]
Generating train split: 41564174 examples [09:20, 138776.24 examples/s]
Generating train split: 41586174 examples [09:21, 118549.84 examples/s]
Generating train split: 41605174 examples [09:21, 91251.95 examples/s]
Generating train split: 41619174 examples [09:21, 98237.12 examples/s]
Generating train split: 41635174 examples [09:21, 107794.75 examples/s]
Generating train split: 41657174 examples [09:21, 107861.46 examples/s]
Generating train split: 41674174 examples [09:22, 112644.50 examples/s]
Generating train split: 41690174 examples [09:22, 46603.48 examples/s]
Generating train split: 41701174 examples [09:23, 27034.42 examples/s]
Generating train split: 41713174 examples [09:24, 27273.65 examples/s]
Generating train split: 41754174 examples [09:24, 53945.05 examples/s]
Generating train split: 41791174 examples [09:24, 81695.46 examples/s]
Generating train split: 41810174 examples [09:24, 87423.69 examples/s]
Generating train split: 41827174 examples [09:25, 92377.26 examples/s]
Generating train split: 41842174 examples [09:25, 92300.56 examples/s]
Generating train split: 41864174 examples [09:25, 106903.85 examples/s]
Generating train split: 41878174 examples [09:25, 111794.13 examples/s]
Generating train split: 41896174 examples [09:25, 124146.06 examples/s]
Generating train split: 41925174 examples [09:25, 146534.66 examples/s]
Generating train split: 41946174 examples [09:25, 121455.07 examples/s]
Generating train split: 41963174 examples [09:26, 95853.85 examples/s]
Generating train split: 41984174 examples [09:26, 113476.62 examples/s]
Generating train split: 42000174 examples [09:26, 107076.32 examples/s]
Generating train split: 42022174 examples [09:26, 119770.58 examples/s]
Generating train split: 42038174 examples [09:27, 50318.14 examples/s]
Generating train split: 42051174 examples [09:28, 40218.38 examples/s]
Generating train split: 42061174 examples [09:28, 45258.61 examples/s]
Generating train split: 42072174 examples [09:28, 42275.28 examples/s]
Generating train split: 42080174 examples [09:29, 29343.18 examples/s]
Generating train split: 42090174 examples [09:29, 36015.76 examples/s]
Generating train split: 42133174 examples [09:29, 80801.33 examples/s]
Generating train split: 42164174 examples [09:29, 111539.98 examples/s]
Generating train split: 42187174 examples [09:29, 130306.75 examples/s]
Generating train split: 42226174 examples [09:29, 172637.80 examples/s]
Generating train split: 42253174 examples [09:29, 161109.32 examples/s]
Generating train split: 42277174 examples [09:30, 116572.85 examples/s]
Generating train split: 42298174 examples [09:30, 130326.41 examples/s]
Generating train split: 42316174 examples [09:30, 67097.82 examples/s]
Generating train split: 42333174 examples [09:31, 78554.75 examples/s]
Generating train split: 42352174 examples [09:31, 86499.55 examples/s]
Generating train split: 42368174 examples [09:32, 36102.00 examples/s]
Generating train split: 42378174 examples [09:32, 39636.12 examples/s]
Generating train split: 42389174 examples [09:33, 34738.60 examples/s]
Generating train split: 42405174 examples [09:33, 45700.37 examples/s]
Generating train split: 42426174 examples [09:33, 64118.87 examples/s]
Generating train split: 42441174 examples [09:33, 63627.67 examples/s]
Generating train split: 42453174 examples [09:33, 52338.05 examples/s]
Generating train split: 42480174 examples [09:33, 77097.74 examples/s]
Generating train split: 42512174 examples [09:34, 113232.15 examples/s]
Generating train split: 42536174 examples [09:34, 134051.45 examples/s]
Generating train split: 42578174 examples [09:34, 178921.26 examples/s]
Generating train split: 42609174 examples [09:34, 195877.50 examples/s]
Generating train split: 42637174 examples [09:34, 124364.87 examples/s]
Generating train split: 42658174 examples [09:35, 70294.68 examples/s]
Generating train split: 42673174 examples [09:35, 77423.60 examples/s]
Generating train split: 42700174 examples [09:35, 98822.85 examples/s]
Generating train split: 42722174 examples [09:36, 54540.00 examples/s]
Generating train split: 42735174 examples [09:37, 35542.40 examples/s]
Generating train split: 42755174 examples [09:37, 45580.48 examples/s]
Generating train split: 42783174 examples [09:37, 64936.16 examples/s]
Generating train split: 42803174 examples [09:37, 77518.81 examples/s]
Generating train split: 42821174 examples [09:38, 71857.84 examples/s]
Generating train split: 42834174 examples [09:38, 68421.59 examples/s]
Generating train split: 42855174 examples [09:38, 86074.87 examples/s]
Generating train split: 42872174 examples [09:38, 90090.29 examples/s]
Generating train split: 42894174 examples [09:38, 107562.24 examples/s]
Generating train split: 42932174 examples [09:39, 101507.85 examples/s]
Generating train split: 42960174 examples [09:39, 124616.83 examples/s]
Generating train split: 42979174 examples [09:39, 117443.01 examples/s]
Generating train split: 42998174 examples [09:40, 62652.20 examples/s]
Generating train split: 43021174 examples [09:40, 67475.09 examples/s]
Generating train split: 43032174 examples [09:41, 43101.33 examples/s]
Generating train split: 43042174 examples [09:41, 41496.06 examples/s]
Generating train split: 43056174 examples [09:41, 50565.44 examples/s]
Generating train split: 43071174 examples [09:41, 60126.66 examples/s]
Generating train split: 43090174 examples [09:41, 76491.89 examples/s]
Generating train split: 43102174 examples [09:42, 54215.12 examples/s]
Generating train split: 43121174 examples [09:42, 71043.94 examples/s]
Generating train split: 43144174 examples [09:42, 95940.20 examples/s]
Generating train split: 43171174 examples [09:42, 125506.95 examples/s]
Generating train split: 43189174 examples [09:42, 111405.28 examples/s]
Generating train split: 43218174 examples [09:43, 133056.18 examples/s]
Generating train split: 43235174 examples [09:43, 107315.45 examples/s]
Generating train split: 43251174 examples [09:43, 109980.50 examples/s]
Generating train split: 43284174 examples [09:43, 147370.22 examples/s]
Generating train split: 43304174 examples [09:44, 74074.90 examples/s]
Generating train split: 43320174 examples [09:44, 83184.47 examples/s]
Generating train split: 43337174 examples [09:44, 65961.25 examples/s]
Generating train split: 43352174 examples [09:44, 70869.77 examples/s]
Generating train split: 43363174 examples [09:45, 53913.68 examples/s]
Generating train split: 43371174 examples [09:45, 41771.20 examples/s]
Generating train split: 43378174 examples [09:46, 28674.56 examples/s]
Generating train split: 43387174 examples [09:46, 34098.97 examples/s]
Generating train split: 43408174 examples [09:46, 54689.78 examples/s]
Generating train split: 43428174 examples [09:46, 66519.23 examples/s]
Generating train split: 43439174 examples [09:47, 55988.50 examples/s]
Generating train split: 43456174 examples [09:47, 70184.73 examples/s]
Generating train split: 43484174 examples [09:47, 101730.71 examples/s]
Generating train split: 43498174 examples [09:47, 82478.59 examples/s]
Generating train split: 43529174 examples [09:47, 118842.93 examples/s]
Generating train split: 43555174 examples [09:47, 135586.78 examples/s]
Generating train split: 43576174 examples [09:47, 120212.49 examples/s]
Generating train split: 43600174 examples [09:48, 138713.44 examples/s]
Generating train split: 43618174 examples [09:48, 59992.64 examples/s]
Generating train split: 43656174 examples [09:49, 68148.47 examples/s]
Generating train split: 43670174 examples [09:49, 56430.81 examples/s]
Generating train split: 43681174 examples [09:49, 60782.84 examples/s]
Generating train split: 43693174 examples [09:50, 47623.93 examples/s]
Generating train split: 43703174 examples [09:50, 46184.58 examples/s]
Generating train split: 43719174 examples [09:51, 43652.08 examples/s]
Generating train split: 43738174 examples [09:51, 58887.42 examples/s]
Generating train split: 43754174 examples [09:51, 71522.21 examples/s]
Generating train split: 43767174 examples [09:51, 71545.30 examples/s]
Generating train split: 43780174 examples [09:51, 76044.46 examples/s]
Generating train split: 43800174 examples [09:51, 97877.10 examples/s]
Generating train split: 43813174 examples [09:51, 99821.64 examples/s]
Generating train split: 43835174 examples [09:52, 103992.88 examples/s]
Generating train split: 43863174 examples [09:52, 137412.17 examples/s]
Generating train split: 43898174 examples [09:52, 167861.20 examples/s]
Generating train split: 43922174 examples [09:52, 118877.51 examples/s]
Generating train split: 43938174 examples [09:52, 111388.41 examples/s]
Generating train split: 43954174 examples [09:52, 99095.38 examples/s]
Generating train split: 43970174 examples [09:53, 64584.63 examples/s]
Generating train split: 43986174 examples [09:53, 76071.43 examples/s]
Generating train split: 43997174 examples [09:54, 51197.91 examples/s]
Generating train split: 44008174 examples [09:54, 48581.57 examples/s]
Generating train split: 44017174 examples [09:54, 52836.57 examples/s]
Generating train split: 44033174 examples [09:55, 35209.81 examples/s]
Generating train split: 44046174 examples [09:55, 44273.49 examples/s]
Generating train split: 44057174 examples [09:55, 40875.15 examples/s]
Generating train split: 44072174 examples [09:55, 53562.62 examples/s]
Generating train split: 44104174 examples [09:55, 89167.67 examples/s]
Generating train split: 44119174 examples [09:56, 93654.11 examples/s]
Generating train split: 44144174 examples [09:56, 113370.39 examples/s]
Generating train split: 44163174 examples [09:56, 115368.37 examples/s]
Generating train split: 44185174 examples [09:56, 131841.42 examples/s]
Generating train split: 44204174 examples [09:56, 112409.11 examples/s]
Generating train split: 44226174 examples [09:56, 132285.43 examples/s]
Generating train split: 44243174 examples [09:56, 113450.23 examples/s]
Generating train split: 44260174 examples [09:57, 80078.99 examples/s]
Generating train split: 44274174 examples [09:57, 70297.73 examples/s]
Generating train split: 44295174 examples [09:57, 89466.82 examples/s]
Generating train split: 44308174 examples [09:58, 43363.55 examples/s]
Generating train split: 44322174 examples [09:58, 52245.05 examples/s]
Generating train split: 44340174 examples [09:58, 56890.66 examples/s]
Generating train split: 44350174 examples [09:59, 41497.48 examples/s]
Generating train split: 44357174 examples [09:59, 35550.98 examples/s]
Generating train split: 44371174 examples [09:59, 44753.20 examples/s]
Generating train split: 44402174 examples [10:00, 72014.41 examples/s]
Generating train split: 44413174 examples [10:00, 73183.25 examples/s]
Generating train split: 44430174 examples [10:00, 78629.14 examples/s]
Generating train split: 44447174 examples [10:00, 94063.28 examples/s]
Generating train split: 44472174 examples [10:00, 114116.96 examples/s]
Generating train split: 44488174 examples [10:00, 93446.94 examples/s]
Generating train split: 44504174 examples [10:01, 93159.04 examples/s]
Generating train split: 44515174 examples [10:01, 86869.69 examples/s]
Generating train split: 44539174 examples [10:01, 116199.16 examples/s]
Generating train split: 44560174 examples [10:01, 130994.68 examples/s]
Generating train split: 44576174 examples [10:01, 117162.61 examples/s]
Generating train split: 44592174 examples [10:02, 63987.11 examples/s]
Generating train split: 44604174 examples [10:02, 44033.89 examples/s]
Generating train split: 44625174 examples [10:02, 60926.01 examples/s]
Generating train split: 44639174 examples [10:03, 53807.18 examples/s]
Generating train split: 44649174 examples [10:03, 46356.60 examples/s]
Generating train split: 44659174 examples [10:03, 37157.95 examples/s]
Generating train split: 44669174 examples [10:04, 32983.22 examples/s]
Generating train split: 44682174 examples [10:04, 41675.37 examples/s]
Generating train split: 44719174 examples [10:04, 78345.71 examples/s]
Generating train split: 44741174 examples [10:04, 91452.60 examples/s]
Generating train split: 44762174 examples [10:04, 102721.80 examples/s]
Generating train split: 44778174 examples [10:05, 104478.05 examples/s]
Generating train split: 44795174 examples [10:05, 114475.20 examples/s]
Generating train split: 44815174 examples [10:05, 117542.28 examples/s]
Generating train split: 44832174 examples [10:05, 122733.88 examples/s]
Generating train split: 44848174 examples [10:05, 118416.85 examples/s]
Generating train split: 44864174 examples [10:05, 91929.08 examples/s]
Generating train split: 44882174 examples [10:06, 97590.40 examples/s]
Generating train split: 44893174 examples [10:06, 97019.11 examples/s]
Generating train split: 44928174 examples [10:06, 68983.95 examples/s]
Generating train split: 44940174 examples [10:07, 49170.60 examples/s]
Generating train split: 44950174 examples [10:07, 40973.99 examples/s]
Generating train split: 44971174 examples [10:08, 45823.58 examples/s]
Generating train split: 44981174 examples [10:08, 51052.53 examples/s]
Generating train split: 44992174 examples [10:08, 45930.14 examples/s]
Generating train split: 45003174 examples [10:08, 53461.07 examples/s]
Generating train split: 45014174 examples [10:08, 55179.78 examples/s]
Generating train split: 45031174 examples [10:08, 73364.99 examples/s]
Generating train split: 45052174 examples [10:09, 93278.82 examples/s]
Generating train split: 45088174 examples [10:09, 142650.31 examples/s]
Generating train split: 45131174 examples [10:09, 193336.64 examples/s]
Generating train split: 45157174 examples [10:09, 173807.35 examples/s]
Generating train split: 45181174 examples [10:09, 122355.73 examples/s]
Generating train split: 45197174 examples [10:10, 69245.30 examples/s]
Generating train split: 45219174 examples [10:10, 80193.92 examples/s]
Generating train split: 45244174 examples [10:10, 100682.55 examples/s]
Generating train split: 45265174 examples [10:10, 110827.04 examples/s]
Generating train split: 45281174 examples [10:11, 116830.96 examples/s]
Generating train split: 45296174 examples [10:11, 105965.94 examples/s]
Generating train split: 45314174 examples [10:11, 116795.60 examples/s]
Generating train split: 45335174 examples [10:11, 126333.54 examples/s]
Generating train split: 45352174 examples [10:11, 132824.01 examples/s]
Generating train split: 45389174 examples [10:11, 168104.99 examples/s]
Generating train split: 45416174 examples [10:11, 188255.47 examples/s]
Generating train split: 45438174 examples [10:12, 175783.09 examples/s]
Generating train split: 45459174 examples [10:12, 160304.75 examples/s]
Generating train split: 45476174 examples [10:12, 102847.64 examples/s]
Generating train split: 45496174 examples [10:12, 112194.76 examples/s]
Generating train split: 45512174 examples [10:12, 111186.39 examples/s]
Generating train split: 45532174 examples [10:13, 76021.27 examples/s]
Generating train split: 45547174 examples [10:13, 75589.31 examples/s]
Generating train split: 45578174 examples [10:13, 107084.73 examples/s]
Generating train split: 45601174 examples [10:13, 124986.72 examples/s]
Generating train split: 45617174 examples [10:13, 110438.30 examples/s]
Generating train split: 45632174 examples [10:14, 103978.49 examples/s]
Generating train split: 45649174 examples [10:14, 72306.35 examples/s]
Generating train split: 45659174 examples [10:14, 72684.42 examples/s]
Generating train split: 45671174 examples [10:14, 60315.37 examples/s]
Generating train split: 45695174 examples [10:15, 86163.89 examples/s]
Generating train split: 45709174 examples [10:15, 85273.86 examples/s]
Generating train split: 45728174 examples [10:15, 103410.12 examples/s]
Generating train split: 45744174 examples [10:15, 92761.02 examples/s]
Generating train split: 45759174 examples [10:15, 63038.40 examples/s]
Generating train split: 45772174 examples [10:16, 59521.62 examples/s]
Generating train split: 45780174 examples [10:16, 60341.90 examples/s]
Generating train split: 45795174 examples [10:16, 71763.55 examples/s]
Generating train split: 45807174 examples [10:16, 73885.85 examples/s]
Generating train split: 45825174 examples [10:16, 88766.03 examples/s]
Generating train split: 45835174 examples [10:17, 34900.99 examples/s]
Generating train split: 45843174 examples [10:17, 38771.77 examples/s]
Generating train split: 45853174 examples [10:17, 39786.16 examples/s]
Generating train split: 45876174 examples [10:18, 62730.19 examples/s]
Generating train split: 45886174 examples [10:18, 66610.46 examples/s]
Generating train split: 45897174 examples [10:18, 65961.51 examples/s]
Generating train split: 45907174 examples [10:18, 70502.91 examples/s]
Generating train split: 45925174 examples [10:18, 91356.37 examples/s]
Generating train split: 45941174 examples [10:19, 59187.19 examples/s]
Generating train split: 45961174 examples [10:19, 73603.69 examples/s]
Generating train split: 45976174 examples [10:19, 85604.84 examples/s]
Generating train split: 45990174 examples [10:19, 87515.73 examples/s]
Generating train split: 46015174 examples [10:19, 113399.08 examples/s]
Generating train split: 46033174 examples [10:19, 105176.25 examples/s]
Generating train split: 46054174 examples [10:19, 115011.92 examples/s]
Generating train split: 46068174 examples [10:20, 98119.51 examples/s]
Generating train split: 46084174 examples [10:20, 76728.94 examples/s]
Generating train split: 46095174 examples [10:20, 51780.10 examples/s]
Generating train split: 46117174 examples [10:21, 71684.39 examples/s]
Generating train split: 46132174 examples [10:21, 77362.62 examples/s]
Generating train split: 46152174 examples [10:21, 92725.24 examples/s]
Generating train split: 46164174 examples [10:21, 58076.83 examples/s]
Generating train split: 46175174 examples [10:22, 56365.04 examples/s]
Generating train split: 46186174 examples [10:22, 56505.59 examples/s]
Generating train split: 46197174 examples [10:22, 46151.49 examples/s]
Generating train split: 46220174 examples [10:22, 70324.16 examples/s]
Generating train split: 46242174 examples [10:22, 89578.12 examples/s]
Generating train split: 46258174 examples [10:23, 76960.32 examples/s]
Generating train split: 46275174 examples [10:23, 89673.77 examples/s]
Generating train split: 46291174 examples [10:23, 67793.11 examples/s]
Generating train split: 46308174 examples [10:23, 80916.08 examples/s]
Generating train split: 46331174 examples [10:23, 98335.37 examples/s]
Generating train split: 46349174 examples [10:24, 109576.59 examples/s]
Generating train split: 46368174 examples [10:24, 95336.57 examples/s]
Generating train split: 46382174 examples [10:24, 78761.95 examples/s]
Generating train split: 46407174 examples [10:24, 103598.92 examples/s]
Generating train split: 46424174 examples [10:25, 56716.59 examples/s]
Generating train split: 46435174 examples [10:25, 53126.23 examples/s]
Generating train split: 46452174 examples [10:25, 65860.71 examples/s]
Generating train split: 46473174 examples [10:25, 85246.49 examples/s]
Generating train split: 46496174 examples [10:26, 50899.96 examples/s]
Generating train split: 46529174 examples [10:26, 77989.51 examples/s]
Generating train split: 46555174 examples [10:26, 95178.67 examples/s]
Generating train split: 46572174 examples [10:27, 67697.68 examples/s]
Generating train split: 46599174 examples [10:27, 71190.58 examples/s]
Generating train split: 46615174 examples [10:27, 80691.96 examples/s]
Generating train split: 46632174 examples [10:28, 61292.68 examples/s]
Generating train split: 46650174 examples [10:28, 73080.79 examples/s]
Generating train split: 46666174 examples [10:28, 81047.45 examples/s]
Generating train split: 46697174 examples [10:28, 116081.26 examples/s]
Generating train split: 46722174 examples [10:28, 126912.47 examples/s]
Generating train split: 46744174 examples [10:28, 137500.53 examples/s]
Generating train split: 46762174 examples [10:29, 49073.51 examples/s]
Generating train split: 46779174 examples [10:30, 52460.09 examples/s]
Generating train split: 46791174 examples [10:30, 58579.63 examples/s]
Generating train split: 46818174 examples [10:30, 84837.48 examples/s]
Generating train split: 46838174 examples [10:30, 65830.11 examples/s]
Generating train split: 46855174 examples [10:31, 57721.40 examples/s]
Generating train split: 46883174 examples [10:31, 78182.17 examples/s]
Generating train split: 46900174 examples [10:31, 68775.74 examples/s]
Generating train split: 46922174 examples [10:31, 86722.61 examples/s]
Generating train split: 46939174 examples [10:32, 93525.87 examples/s]
Generating train split: 46957174 examples [10:32, 77729.35 examples/s]
Generating train split: 46968174 examples [10:32, 73744.85 examples/s]
Generating train split: 46996174 examples [10:32, 105859.98 examples/s]
Generating train split: 47013174 examples [10:32, 87656.56 examples/s]
Generating train split: 47037174 examples [10:33, 109630.11 examples/s]
Generating train split: 47052174 examples [10:33, 100460.21 examples/s]
Generating train split: 47079174 examples [10:33, 124150.26 examples/s]
Generating train split: 47095174 examples [10:33, 73069.26 examples/s]
Generating train split: 47107174 examples [10:34, 73562.37 examples/s]
Generating train split: 47119174 examples [10:34, 70050.72 examples/s]
Generating train split: 47130174 examples [10:34, 73307.99 examples/s]
Generating train split: 47157174 examples [10:34, 106985.09 examples/s]
Generating train split: 47186174 examples [10:35, 54150.56 examples/s]
Generating train split: 47203174 examples [10:35, 63711.26 examples/s]
Generating train split: 47216174 examples [10:35, 52538.28 examples/s]
Generating train split: 47240174 examples [10:36, 72379.91 examples/s]
Generating train split: 47257174 examples [10:36, 59093.20 examples/s]
Generating train split: 47273174 examples [10:36, 68716.65 examples/s]
Generating train split: 47291174 examples [10:37, 57233.96 examples/s]
Generating train split: 47303174 examples [10:37, 64251.23 examples/s]
Generating train split: 47331174 examples [10:37, 94120.52 examples/s]
Generating train split: 47347174 examples [10:37, 90406.02 examples/s]
Generating train split: 47360174 examples [10:37, 86737.98 examples/s]
Generating train split: 47378174 examples [10:37, 96861.38 examples/s]
Generating train split: 47394174 examples [10:37, 99880.06 examples/s]
Generating train split: 47409174 examples [10:38, 80770.38 examples/s]
Generating train split: 47420174 examples [10:38, 57923.23 examples/s]
Generating train split: 47441174 examples [10:38, 79085.81 examples/s]
Generating train split: 47458174 examples [10:39, 66036.56 examples/s]
Generating train split: 47486174 examples [10:39, 94891.38 examples/s]
Generating train split: 47513174 examples [10:39, 85871.88 examples/s]
Generating train split: 47526174 examples [10:40, 53564.60 examples/s]
Generating train split: 47546174 examples [10:40, 66128.83 examples/s]
Generating train split: 47557174 examples [10:40, 55395.72 examples/s]
Generating train split: 47586174 examples [10:40, 82558.67 examples/s]
Generating train split: 47602174 examples [10:41, 52872.03 examples/s]
Generating train split: 47619174 examples [10:41, 64354.73 examples/s]
Generating train split: 47635174 examples [10:41, 58130.87 examples/s]
Generating train split: 47668174 examples [10:41, 91175.67 examples/s]
Generating train split: 47701174 examples [10:42, 123963.19 examples/s]
Generating train split: 47721174 examples [10:42, 122161.50 examples/s]
Generating train split: 47741174 examples [10:42, 133645.29 examples/s]
Generating train split: 47763174 examples [10:42, 113373.88 examples/s]
Generating train split: 47781174 examples [10:42, 105714.78 examples/s]
Generating train split: 47796174 examples [10:43, 71776.42 examples/s]
Generating train split: 47809174 examples [10:43, 75581.83 examples/s]
Generating train split: 47819174 examples [10:43, 65448.08 examples/s]
Generating train split: 47831174 examples [10:43, 64002.33 examples/s]
Generating train split: 47854174 examples [10:44, 60411.11 examples/s]
Generating train split: 47861174 examples [10:44, 40318.81 examples/s]
Generating train split: 47898174 examples [10:44, 75233.89 examples/s]
Generating train split: 47915174 examples [10:45, 64602.53 examples/s]
Generating train split: 47931174 examples [10:45, 71906.65 examples/s]
Generating train split: 47947174 examples [10:45, 81635.92 examples/s]
Generating train split: 47959174 examples [10:45, 64353.07 examples/s]
Generating train split: 47969174 examples [10:45, 61559.09 examples/s]
Generating train split: 47982174 examples [10:46, 63910.76 examples/s]
Generating train split: 47992174 examples [10:46, 52175.93 examples/s]
Generating train split: 48011174 examples [10:46, 71899.15 examples/s]
Generating train split: 48034174 examples [10:46, 95229.23 examples/s]
Generating train split: 48064174 examples [10:46, 122803.67 examples/s]
Generating train split: 48081174 examples [10:47, 66747.88 examples/s]
Generating train split: 48096174 examples [10:47, 72024.13 examples/s]
Generating train split: 48114174 examples [10:47, 75037.61 examples/s]
Generating train split: 48135174 examples [10:47, 90344.04 examples/s]
Generating train split: 48147174 examples [10:48, 59819.31 examples/s]
Generating train split: 48158174 examples [10:48, 65527.58 examples/s]
Generating train split: 48173174 examples [10:48, 76385.13 examples/s]
Generating train split: 48187174 examples [10:48, 62761.83 examples/s]
Generating train split: 48203174 examples [10:49, 58905.43 examples/s]
Generating train split: 48213174 examples [10:49, 63711.87 examples/s]
Generating train split: 48242174 examples [10:49, 95419.74 examples/s]
Generating train split: 48258174 examples [10:49, 71783.86 examples/s]
Generating train split: 48280174 examples [10:49, 91911.09 examples/s]
Generating train split: 48298174 examples [10:50, 99705.83 examples/s]
Generating train split: 48315174 examples [10:50, 68108.13 examples/s]
Generating train split: 48326174 examples [10:50, 65797.27 examples/s]
Generating train split: 48347174 examples [10:50, 86644.49 examples/s]
Generating train split: 48367174 examples [10:51, 100428.31 examples/s]
Generating train split: 48389174 examples [10:51, 115060.08 examples/s]
Generating train split: 48408174 examples [10:51, 123792.86 examples/s]
Generating train split: 48424174 examples [10:51, 125959.22 examples/s]
Generating train split: 48438174 examples [10:51, 111950.06 examples/s]
Generating train split: 48453174 examples [10:52, 51191.49 examples/s]
Generating train split: 48469174 examples [10:52, 62523.04 examples/s]
Generating train split: 48481174 examples [10:52, 65684.31 examples/s]
Generating train split: 48497174 examples [10:52, 78867.37 examples/s]
Generating train split: 48513174 examples [10:52, 68762.92 examples/s]
Generating train split: 48525174 examples [10:53, 64711.82 examples/s]
Generating train split: 48537174 examples [10:53, 60786.27 examples/s]
Generating train split: 48559174 examples [10:53, 83216.22 examples/s]
Generating train split: 48572174 examples [10:53, 86716.14 examples/s]
Generating train split: 48583174 examples [10:54, 56050.09 examples/s]
Generating train split: 48595174 examples [10:54, 47788.18 examples/s]
Generating train split: 48605174 examples [10:54, 54556.62 examples/s]
Generating train split: 48631174 examples [10:54, 87226.24 examples/s]
Generating train split: 48648174 examples [10:54, 85594.99 examples/s]
Generating train split: 48660174 examples [10:55, 57462.17 examples/s]
Generating train split: 48673174 examples [10:55, 57049.59 examples/s]
Generating train split: 48695174 examples [10:55, 76146.15 examples/s]
Generating train split: 48722174 examples [10:55, 93092.55 examples/s]
Generating train split: 48745174 examples [10:55, 109221.41 examples/s]
Generating train split: 48760174 examples [10:56, 100501.58 examples/s]
Generating train split: 48775174 examples [10:56, 81278.57 examples/s]
Generating train split: 48785174 examples [10:56, 53504.34 examples/s]
Generating train split: 48795174 examples [10:57, 56984.93 examples/s]
Generating train split: 48807174 examples [10:57, 62527.89 examples/s]
Generating train split: 48827174 examples [10:57, 85534.30 examples/s]
Generating train split: 48843174 examples [10:57, 66357.28 examples/s]
Generating train split: 48854174 examples [10:58, 52443.46 examples/s]
Generating train split: 48866174 examples [10:58, 60758.62 examples/s]
Generating train split: 48894174 examples [10:58, 95595.68 examples/s]
Generating train split: 48910174 examples [10:58, 79581.10 examples/s]
Generating train split: 48931174 examples [10:58, 64160.70 examples/s]
Generating train split: 48943174 examples [10:59, 68059.75 examples/s]
Generating train split: 48954174 examples [10:59, 69325.77 examples/s]
Generating train split: 48979174 examples [10:59, 97362.57 examples/s]
Generating train split: 48995174 examples [10:59, 73259.64 examples/s]
Generating train split: 49010174 examples [11:00, 65033.38 examples/s]
Generating train split: 49033174 examples [11:00, 87337.34 examples/s]
Generating train split: 49065174 examples [11:00, 117657.32 examples/s]
Generating train split: 49081174 examples [11:00, 96659.59 examples/s]
Generating train split: 49099174 examples [11:01, 68088.16 examples/s]
Generating train split: 49111174 examples [11:01, 72704.83 examples/s]
Generating train split: 49122174 examples [11:01, 75629.32 examples/s]
Generating train split: 49133174 examples [11:01, 73838.47 examples/s]
Generating train split: 49143174 examples [11:01, 59063.95 examples/s]
Generating train split: 49153174 examples [11:01, 65042.87 examples/s]
Generating train split: 49174174 examples [11:01, 91052.23 examples/s]
Generating train split: 49186174 examples [11:02, 69006.89 examples/s]
Generating train split: 49198174 examples [11:02, 73107.34 examples/s]
Generating train split: 49208174 examples [11:02, 50181.39 examples/s]
Generating train split: 49225174 examples [11:02, 64467.12 examples/s]
Generating train split: 49235174 examples [11:03, 56720.20 examples/s]
Generating train split: 49251174 examples [11:03, 73045.25 examples/s]
Generating train split: 49261174 examples [11:03, 75211.93 examples/s]
Generating train split: 49285174 examples [11:03, 86010.42 examples/s]
Generating train split: 49297174 examples [11:03, 63871.42 examples/s]
Generating train split: 49308174 examples [11:04, 52111.68 examples/s]
Generating train split: 49340174 examples [11:04, 89620.78 examples/s]
Generating train split: 49356174 examples [11:04, 87721.85 examples/s]
Generating train split: 49372174 examples [11:04, 87693.24 examples/s]
Generating train split: 49388174 examples [11:04, 81816.29 examples/s]
Generating train split: 49405174 examples [11:05, 78665.68 examples/s]
Generating train split: 49426174 examples [11:05, 60226.66 examples/s]
Generating train split: 49437174 examples [11:05, 57599.20 examples/s]
Generating train split: 49453174 examples [11:06, 70473.53 examples/s]
Generating train split: 49467174 examples [11:06, 80321.05 examples/s]
Generating train split: 49478174 examples [11:06, 56275.61 examples/s]
Generating train split: 49498174 examples [11:06, 72550.86 examples/s]
Generating train split: 49508174 examples [11:06, 65784.34 examples/s]
Generating train split: 49538174 examples [11:07, 99746.01 examples/s]
Generating train split: 49554174 examples [11:07, 55552.00 examples/s]
Generating train split: 49565174 examples [11:07, 56420.94 examples/s]
Generating train split: 49585174 examples [11:07, 73771.06 examples/s]
Generating train split: 49604174 examples [11:08, 89118.75 examples/s]
Generating train split: 49622174 examples [11:08, 58857.47 examples/s]
Generating train split: 49632174 examples [11:08, 63923.34 examples/s]
Generating train split: 49643174 examples [11:08, 65576.70 examples/s]
Generating train split: 49674174 examples [11:09, 98380.19 examples/s]
Generating train split: 49700174 examples [11:09, 120504.70 examples/s]
Generating train split: 49716174 examples [11:09, 68851.06 examples/s]
Generating train split: 49730174 examples [11:09, 64200.87 examples/s]
Generating train split: 49746174 examples [11:10, 51292.98 examples/s]
Generating train split: 49757174 examples [11:10, 55196.13 examples/s]
Generating train split: 49789174 examples [11:10, 89857.38 examples/s]
Generating train split: 49815174 examples [11:10, 114659.20 examples/s]
Generating train split: 49836174 examples [11:11, 56894.66 examples/s]
Generating train split: 49868174 examples [11:11, 72401.70 examples/s]
Generating train split: 49885174 examples [11:12, 68136.83 examples/s]
Generating train split: 49896174 examples [11:12, 66172.26 examples/s]
Generating train split: 49906174 examples [11:12, 59102.09 examples/s]
Generating train split: 49929174 examples [11:12, 79448.67 examples/s]
Generating train split: 49951174 examples [11:12, 100230.47 examples/s]
Generating train split: 49968174 examples [11:12, 103309.79 examples/s]
Generating train split: 49984174 examples [11:13, 90710.23 examples/s]
Generating train split: 50001174 examples [11:13, 73276.96 examples/s]
Generating train split: 50019174 examples [11:13, 86916.22 examples/s]
Generating train split: 50035174 examples [11:13, 91349.90 examples/s]
Generating train split: 50050174 examples [11:14, 80187.91 examples/s]
Generating train split: 50066174 examples [11:14, 80858.72 examples/s]
Generating train split: 50076174 examples [11:14, 82460.03 examples/s]
Generating train split: 50087174 examples [11:14, 79149.72 examples/s]
Generating train split: 50097174 examples [11:14, 71483.54 examples/s]
Generating train split: 50110174 examples [11:15, 46266.76 examples/s]
Generating train split: 50131174 examples [11:15, 65219.91 examples/s]
Generating train split: 50152174 examples [11:15, 66523.23 examples/s]
Generating train split: 50174174 examples [11:15, 87444.15 examples/s]
Generating train split: 50193174 examples [11:15, 102541.43 examples/s]
Generating train split: 50209174 examples [11:16, 56531.85 examples/s]
Generating train split: 50220174 examples [11:16, 57416.26 examples/s]
Generating train split: 50242174 examples [11:16, 79006.27 examples/s]
Generating train split: 50259174 examples [11:17, 49462.98 examples/s]
Generating train split: 50281174 examples [11:17, 66528.64 examples/s]
Generating train split: 50297174 examples [11:17, 75354.10 examples/s]
Generating train split: 50314174 examples [11:17, 88198.85 examples/s]
Generating train split: 50331174 examples [11:18, 59061.53 examples/s]
Generating train split: 50346174 examples [11:18, 66744.85 examples/s]
Generating train split: 50363174 examples [11:18, 66731.35 examples/s]
Generating train split: 50379174 examples [11:18, 77231.43 examples/s]
Generating train split: 50390174 examples [11:19, 77861.10 examples/s]
Generating train split: 50400174 examples [11:19, 77298.37 examples/s]
Generating train split: 50410174 examples [11:19, 71182.58 examples/s]
Generating train split: 50439174 examples [11:19, 110391.14 examples/s]
Generating train split: 50455174 examples [11:19, 79258.45 examples/s]
Generating train split: 50470174 examples [11:19, 83777.76 examples/s]
Generating train split: 50481174 examples [11:20, 66126.86 examples/s]
Generating train split: 50493174 examples [11:20, 69522.26 examples/s]
Generating train split: 50509174 examples [11:20, 84011.40 examples/s]
Generating train split: 50523174 examples [11:21, 47168.88 examples/s]
Generating train split: 50545174 examples [11:21, 60018.60 examples/s]
Generating train split: 50566174 examples [11:21, 78713.49 examples/s]
Generating train split: 50593174 examples [11:21, 108311.78 examples/s]
Generating train split: 50615174 examples [11:21, 85282.97 examples/s]
Generating train split: 50631174 examples [11:22, 88135.11 examples/s]
Generating train split: 50647174 examples [11:22, 83317.15 examples/s]
Generating train split: 50662174 examples [11:22, 88212.24 examples/s]
Generating train split: 50673174 examples [11:22, 91798.95 examples/s]
Generating train split: 50685174 examples [11:22, 82869.79 examples/s]
Generating train split: 50698174 examples [11:22, 72790.98 examples/s]
Generating train split: 50710174 examples [11:23, 80859.22 examples/s]
Generating train split: 50723174 examples [11:23, 56764.61 examples/s]
Generating train split: 50734174 examples [11:23, 59127.82 examples/s]
Generating train split: 50744174 examples [11:23, 60043.74 examples/s]
Generating train split: 50754174 examples [11:23, 59280.05 examples/s]
Generating train split: 50775174 examples [11:24, 80335.69 examples/s]
Generating train split: 50787174 examples [11:24, 64427.27 examples/s]
Generating train split: 50808174 examples [11:24, 88017.42 examples/s]
Generating train split: 50823174 examples [11:24, 93689.28 examples/s]
Generating train split: 50839174 examples [11:25, 65706.38 examples/s]
Generating train split: 50848174 examples [11:25, 68510.94 examples/s]
Generating train split: 50860174 examples [11:25, 50124.18 examples/s]
Generating train split: 50873174 examples [11:25, 55266.37 examples/s]
Generating train split: 50884174 examples [11:25, 59850.68 examples/s]
Generating train split: 50895174 examples [11:25, 65789.32 examples/s]
Generating train split: 50904174 examples [11:26, 64715.23 examples/s]
Generating train split: 50929174 examples [11:26, 100345.84 examples/s]
Generating train split: 50944174 examples [11:26, 70351.24 examples/s]
Generating train split: 50956174 examples [11:26, 74799.28 examples/s]
Generating train split: 50967174 examples [11:26, 79070.14 examples/s]
Generating train split: 50978174 examples [11:27, 74239.51 examples/s]
Generating train split: 51009174 examples [11:27, 119168.54 examples/s]
Generating train split: 51025174 examples [11:27, 100699.69 examples/s]
Generating train split: 51039174 examples [11:27, 80880.54 examples/s]
Generating train split: 51057174 examples [11:27, 94667.89 examples/s]
Generating train split: 51072174 examples [11:28, 80328.35 examples/s]
Generating train split: 51083174 examples [11:28, 85239.86 examples/s]
Generating train split: 51098174 examples [11:28, 95381.81 examples/s]
Generating train split: 51112174 examples [11:28, 65178.33 examples/s]
Generating train split: 51123174 examples [11:28, 68782.50 examples/s]
Generating train split: 51133174 examples [11:29, 58134.57 examples/s]
Generating train split: 51165174 examples [11:29, 97562.43 examples/s]
Generating train split: 51181174 examples [11:29, 57377.83 examples/s]
Generating train split: 51192174 examples [11:29, 61229.41 examples/s]
Generating train split: 51203174 examples [11:30, 55614.13 examples/s]
Generating train split: 51222174 examples [11:30, 65366.37 examples/s]
Generating train split: 51232174 examples [11:30, 59095.87 examples/s]
Generating train split: 51250174 examples [11:30, 74084.42 examples/s]
Generating train split: 51266174 examples [11:30, 88939.67 examples/s]
Generating train split: 51287174 examples [11:30, 102128.75 examples/s]
Generating train split: 51303174 examples [11:31, 95904.39 examples/s]
Generating train split: 51319174 examples [11:31, 60670.71 examples/s]
Generating train split: 51341174 examples [11:31, 80394.11 examples/s]
Generating train split: 51358174 examples [11:31, 82568.55 examples/s]
Generating train split: 51374174 examples [11:32, 91014.42 examples/s]
Generating train split: 51390174 examples [11:32, 75764.08 examples/s]
Generating train split: 51401174 examples [11:32, 66121.80 examples/s]
Generating train split: 51418174 examples [11:32, 80152.87 examples/s]
Generating train split: 51431174 examples [11:33, 59143.77 examples/s]
Generating train split: 51450174 examples [11:33, 77227.37 examples/s]
Generating train split: 51472174 examples [11:33, 65578.44 examples/s]
Generating train split: 51482174 examples [11:34, 44687.16 examples/s]
Generating train split: 51494174 examples [11:34, 43894.79 examples/s]
Generating train split: 51515174 examples [11:34, 60958.41 examples/s]
Generating train split: 51525174 examples [11:34, 64105.00 examples/s]
Generating train split: 51536174 examples [11:34, 71303.25 examples/s]
Generating train split: 51546174 examples [11:34, 70320.19 examples/s]
Generating train split: 51556174 examples [11:35, 58823.46 examples/s]
Generating train split: 51573174 examples [11:35, 78173.67 examples/s]
Generating train split: 51590174 examples [11:35, 94051.45 examples/s]
Generating train split: 51612174 examples [11:35, 109742.01 examples/s]
Generating train split: 51627174 examples [11:35, 113362.91 examples/s]
Generating train split: 51645174 examples [11:36, 67220.24 examples/s]
Generating train split: 51655174 examples [11:36, 68239.90 examples/s]
Generating train split: 51671174 examples [11:36, 81984.44 examples/s]
Generating train split: 51682174 examples [11:36, 79146.02 examples/s]
Generating train split: 51693174 examples [11:36, 82216.78 examples/s]
Generating train split: 51707174 examples [11:36, 90623.49 examples/s]
Generating train split: 51728174 examples [11:36, 111754.44 examples/s]
Generating train split: 51744174 examples [11:37, 73795.28 examples/s]
Generating train split: 51755174 examples [11:37, 79074.16 examples/s]
Generating train split: 51770174 examples [11:37, 67836.98 examples/s]
Generating train split: 51796174 examples [11:37, 97239.75 examples/s]
Generating train split: 51819174 examples [11:37, 121215.18 examples/s]
Generating train split: 51836174 examples [11:38, 59157.22 examples/s]
Generating train split: 51848174 examples [11:39, 38734.10 examples/s]
Generating train split: 51860174 examples [11:39, 43789.72 examples/s]
Generating train split: 51891174 examples [11:39, 72872.33 examples/s]
Generating train split: 51910174 examples [11:39, 72884.72 examples/s]
Generating train split: 51933174 examples [11:39, 93299.81 examples/s]
Generating train split: 51949174 examples [11:40, 96929.43 examples/s]
Generating train split: 51969174 examples [11:40, 113451.81 examples/s]
Generating train split: 51985174 examples [11:40, 99411.91 examples/s]
Generating train split: 52007174 examples [11:40, 111903.42 examples/s]
Generating train split: 52023174 examples [11:40, 85452.98 examples/s]
Generating train split: 52036174 examples [11:41, 71926.93 examples/s]
Generating train split: 52051174 examples [11:41, 60073.04 examples/s]
Generating train split: 52062174 examples [11:41, 51914.85 examples/s]
Generating train split: 52092174 examples [11:41, 83171.02 examples/s]
Generating train split: 52106174 examples [11:42, 63785.98 examples/s]
Generating train split: 52123174 examples [11:42, 73427.45 examples/s]
Generating train split: 52134174 examples [11:42, 61046.67 examples/s]
Generating train split: 52145174 examples [11:42, 56815.00 examples/s]
Generating train split: 52155174 examples [11:43, 59513.93 examples/s]
Generating train split: 52163174 examples [11:43, 42615.91 examples/s]
Generating train split: 52173174 examples [11:43, 37420.31 examples/s]
Generating train split: 52184174 examples [11:44, 42694.33 examples/s]
Generating train split: 52196174 examples [11:44, 53403.67 examples/s]
Generating train split: 52213174 examples [11:44, 70676.37 examples/s]
Generating train split: 52223174 examples [11:44, 66616.92 examples/s]
Generating train split: 52261174 examples [11:44, 123490.50 examples/s]
Generating train split: 52281174 examples [11:44, 131577.78 examples/s]
Generating train split: 52309174 examples [11:44, 162394.85 examples/s]
Generating train split: 52329174 examples [11:45, 73566.40 examples/s]
Generating train split: 52345174 examples [11:45, 81195.51 examples/s]
Generating train split: 52362174 examples [11:45, 80188.87 examples/s]
Generating train split: 52387174 examples [11:45, 105077.20 examples/s]
Generating train split: 52406174 examples [11:46, 116124.17 examples/s]
Generating train split: 52425174 examples [11:46, 84460.30 examples/s]
Generating train split: 52442174 examples [11:46, 82405.06 examples/s]
Generating train split: 52456174 examples [11:46, 67498.99 examples/s]
Generating train split: 52473174 examples [11:47, 58180.52 examples/s]
Generating train split: 52484174 examples [11:47, 55022.26 examples/s]
Generating train split: 52494174 examples [11:47, 59437.96 examples/s]
Generating train split: 52504174 examples [11:47, 52444.93 examples/s]
Generating train split: 52516174 examples [11:48, 60481.31 examples/s]
Generating train split: 52537174 examples [11:48, 84725.49 examples/s]
Generating train split: 52549174 examples [11:48, 46828.16 examples/s]
Generating train split: 52569174 examples [11:49, 56523.27 examples/s]
Generating train split: 52596174 examples [11:49, 83860.30 examples/s]
Generating train split: 52611174 examples [11:49, 86168.87 examples/s]
Generating train split: 52628174 examples [11:49, 97450.29 examples/s]
Generating train split: 52643174 examples [11:49, 102136.62 examples/s]
Generating train split: 52659174 examples [11:49, 101351.01 examples/s]
Generating train split: 52672174 examples [11:49, 97650.43 examples/s]
Generating train split: 52683174 examples [11:50, 67013.73 examples/s]
Generating train split: 52696174 examples [11:50, 57589.47 examples/s]
Generating train split: 52726174 examples [11:50, 91002.65 examples/s]
Generating train split: 52740174 examples [11:50, 71938.55 examples/s]
Generating train split: 52761174 examples [11:51, 92318.19 examples/s]
Generating train split: 52786174 examples [11:51, 117939.66 examples/s]
Generating train split: 52803174 examples [11:51, 53431.36 examples/s]
Generating train split: 52815174 examples [11:52, 51188.67 examples/s]
Generating train split: 52833174 examples [11:52, 62222.54 examples/s]
Generating train split: 52844174 examples [11:52, 59455.66 examples/s]
Generating train split: 52862174 examples [11:52, 67560.13 examples/s]
Generating train split: 52883174 examples [11:52, 88501.61 examples/s]
Generating train split: 52896174 examples [11:53, 84074.56 examples/s]
Generating train split: 52907174 examples [11:53, 79948.46 examples/s]
Generating train split: 52917174 examples [11:53, 83276.41 examples/s]
Generating train split: 52944174 examples [11:53, 112678.78 examples/s]
Generating train split: 52959174 examples [11:53, 106337.95 examples/s]
Generating train split: 52972174 examples [11:53, 78812.93 examples/s]
Generating train split: 52997174 examples [11:54, 102951.60 examples/s]
Generating train split: 53013174 examples [11:54, 95533.38 examples/s]
Generating train split: 53024174 examples [11:54, 56440.23 examples/s]
Generating train split: 53037178 examples [11:54, 60293.47 examples/s]
Generating train split: 53048178 examples [11:55, 66481.77 examples/s]
Generating train split: 53080178 examples [11:55, 109373.38 examples/s]
Generating train split: 53104178 examples [11:55, 132436.69 examples/s]
Generating train split: 53125178 examples [11:55, 103278.11 examples/s]
Generating train split: 53150178 examples [11:56, 79163.94 examples/s]
Generating train split: 53167178 examples [11:56, 76064.19 examples/s]
Generating train split: 53178178 examples [11:56, 58943.98 examples/s]
Generating train split: 53190178 examples [11:56, 53677.76 examples/s]
Generating train split: 53201178 examples [11:57, 53514.83 examples/s]
Generating train split: 53208178 examples [11:57, 41317.49 examples/s]
Generating train split: 53219178 examples [11:57, 49568.48 examples/s]
Generating train split: 53237178 examples [11:57, 69191.49 examples/s]
Generating train split: 53248178 examples [11:57, 66682.55 examples/s]
Generating train split: 53260178 examples [11:58, 73488.41 examples/s]
Generating train split: 53286178 examples [11:58, 102536.52 examples/s]
Generating train split: 53303178 examples [11:58, 104459.91 examples/s]
Generating train split: 53317178 examples [11:58, 83044.34 examples/s]
Generating train split: 53330178 examples [11:58, 72352.93 examples/s]
Generating train split: 53340178 examples [11:58, 76137.98 examples/s]
Generating train split: 53352178 examples [11:59, 62236.03 examples/s]
Generating train split: 53373178 examples [11:59, 67483.03 examples/s]
Generating train split: 53393178 examples [11:59, 87827.20 examples/s]
Generating train split: 53416178 examples [11:59, 110668.87 examples/s]
Generating train split: 53432178 examples [11:59, 88303.89 examples/s]
Generating train split: 53448178 examples [12:00, 74802.78 examples/s]
Generating train split: 53465178 examples [12:00, 63445.42 examples/s]
Generating train split: 53475178 examples [12:00, 57668.34 examples/s]
Generating train split: 53486178 examples [12:01, 59444.64 examples/s]
Generating train split: 53498178 examples [12:01, 60119.39 examples/s]
Generating train split: 53509178 examples [12:01, 59492.56 examples/s]
Generating train split: 53521178 examples [12:01, 66930.79 examples/s]
Generating train split: 53531178 examples [12:01, 54400.25 examples/s]
Generating train split: 53542178 examples [12:02, 47761.40 examples/s]
Generating train split: 53553178 examples [12:02, 48128.04 examples/s]
Generating train split: 53585178 examples [12:02, 85153.80 examples/s]
Generating train split: 53603178 examples [12:02, 97649.00 examples/s]
Generating train split: 53618178 examples [12:02, 107460.52 examples/s]
Generating train split: 53638178 examples [12:02, 123939.39 examples/s]
Generating train split: 53656178 examples [12:03, 82991.59 examples/s]
Generating train split: 53668178 examples [12:03, 60844.51 examples/s]
Generating train split: 53677178 examples [12:03, 58473.96 examples/s]
Generating train split: 53694178 examples [12:04, 54711.68 examples/s]
Generating train split: 53731178 examples [12:04, 96360.15 examples/s]
Generating train split: 53768178 examples [12:04, 134502.59 examples/s]
Generating train split: 53791178 examples [12:04, 135409.99 examples/s]
Generating train split: 53813178 examples [12:05, 56341.66 examples/s]
Generating train split: 53830178 examples [12:05, 57500.10 examples/s]
Generating train split: 53847178 examples [12:05, 68087.22 examples/s]
Generating train split: 53860178 examples [12:06, 65853.99 examples/s]
Generating train split: 53872178 examples [12:06, 44002.23 examples/s]
Generating train split: 53895178 examples [12:06, 63004.61 examples/s]
Generating train split: 53910178 examples [12:06, 72728.43 examples/s]
Generating train split: 53928178 examples [12:07, 87346.14 examples/s]
Generating train split: 53950178 examples [12:07, 100058.59 examples/s]
Generating train split: 53966178 examples [12:07, 109301.59 examples/s]
Generating train split: 53982178 examples [12:07, 83556.73 examples/s]
Generating train split: 53998178 examples [12:07, 87604.75 examples/s]
Generating train split: 54009178 examples [12:07, 87735.93 examples/s]
Generating train split: 54024178 examples [12:08, 77033.96 examples/s]
Generating train split: 54075178 examples [12:08, 150099.42 examples/s]
Generating train split: 54097178 examples [12:08, 152320.00 examples/s]
Generating train split: 54118178 examples [12:08, 102036.78 examples/s]
Generating train split: 54140178 examples [12:08, 117170.43 examples/s]
Generating train split: 54156178 examples [12:09, 53099.16 examples/s]
Generating train split: 54174178 examples [12:10, 55352.44 examples/s]
Generating train split: 54185178 examples [12:10, 52113.64 examples/s]
Generating train split: 54202178 examples [12:10, 62707.84 examples/s]
Generating train split: 54212178 examples [12:10, 63313.86 examples/s]
Generating train split: 54223178 examples [12:10, 61792.02 examples/s]
Generating train split: 54235178 examples [12:11, 54836.24 examples/s]
Generating train split: 54253178 examples [12:11, 62803.23 examples/s]
Generating train split: 54286178 examples [12:11, 103366.87 examples/s]
Generating train split: 54302178 examples [12:11, 109810.79 examples/s]
Generating train split: 54317178 examples [12:12, 53677.25 examples/s]
Generating train split: 54333178 examples [12:12, 64491.37 examples/s]
Generating train split: 54350178 examples [12:12, 72026.54 examples/s]
Generating train split: 54362178 examples [12:12, 63447.55 examples/s]
Generating train split: 54393178 examples [12:12, 98754.83 examples/s]
Generating train split: 54424178 examples [12:13, 134836.74 examples/s]
Generating train split: 54449178 examples [12:13, 81952.37 examples/s]
Generating train split: 54465178 examples [12:13, 71320.90 examples/s]
Generating train split: 54479178 examples [12:14, 51123.38 examples/s]
Generating train split: 54495178 examples [12:14, 56067.12 examples/s]
Generating train split: 54517178 examples [12:15, 58148.78 examples/s]
Generating train split: 54529178 examples [12:15, 45716.77 examples/s]
Generating train split: 54547178 examples [12:15, 56858.78 examples/s]
Generating train split: 54556178 examples [12:15, 55494.95 examples/s]
Generating train split: 54583178 examples [12:15, 84110.50 examples/s]
Generating train split: 54598178 examples [12:16, 91757.15 examples/s]
Generating train split: 54626178 examples [12:16, 123240.99 examples/s]
Generating train split: 54644178 examples [12:16, 83641.83 examples/s]
Generating train split: 54661178 examples [12:16, 70411.38 examples/s]
Generating train split: 54690178 examples [12:17, 99373.43 examples/s]
Generating train split: 54708178 examples [12:17, 83270.93 examples/s]
Generating train split: 54724178 examples [12:17, 92729.48 examples/s]
Generating train split: 54751178 examples [12:17, 122621.58 examples/s]
Generating train split: 54786178 examples [12:17, 161686.75 examples/s]
Generating train split: 54810178 examples [12:18, 68719.09 examples/s]
Generating train split: 54826178 examples [12:18, 65629.33 examples/s]
Generating train split: 54844178 examples [12:19, 62838.58 examples/s]
Generating train split: 54859178 examples [12:19, 51028.94 examples/s]
Generating train split: 54870178 examples [12:19, 56952.21 examples/s]
Generating train split: 54881178 examples [12:20, 42788.55 examples/s]
Generating train split: 54910178 examples [12:20, 69503.49 examples/s]
Generating train split: 54929178 examples [12:20, 74633.51 examples/s]
Generating train split: 54961178 examples [12:20, 109116.75 examples/s]
Generating train split: 54979178 examples [12:21, 77779.38 examples/s]
Generating train split: 54994178 examples [12:21, 86891.73 examples/s]
Generating train split: 55009178 examples [12:21, 90813.46 examples/s]
Generating train split: 55022178 examples [12:21, 78562.65 examples/s]
Generating train split: 55048178 examples [12:21, 109065.66 examples/s]
Generating train split: 55069178 examples [12:21, 119570.04 examples/s]
Generating train split: 55086178 examples [12:21, 109034.04 examples/s]
Generating train split: 55104178 examples [12:22, 93442.05 examples/s]
Generating train split: 55122178 examples [12:22, 108100.78 examples/s]
Generating train split: 55138178 examples [12:23, 45368.93 examples/s]
Generating train split: 55151178 examples [12:23, 49314.66 examples/s]
Generating train split: 55162178 examples [12:23, 46177.70 examples/s]
Generating train split: 55180178 examples [12:24, 41473.08 examples/s]
Generating train split: 55206178 examples [12:24, 62886.75 examples/s]
Generating train split: 55222178 examples [12:24, 56577.87 examples/s]
Generating train split: 55233178 examples [12:24, 57327.75 examples/s]
Generating train split: 55264178 examples [12:25, 87919.44 examples/s]
Generating train split: 55280178 examples [12:25, 85342.33 examples/s]
Generating train split: 55296178 examples [12:25, 80507.00 examples/s]
Generating train split: 55310178 examples [12:25, 69643.89 examples/s]
Generating train split: 55336178 examples [12:25, 94898.47 examples/s]
Generating train split: 55357178 examples [12:25, 109470.38 examples/s]
Generating train split: 55374178 examples [12:26, 84642.08 examples/s]
Generating train split: 55401178 examples [12:26, 113273.03 examples/s]
Generating train split: 55419178 examples [12:26, 118868.79 examples/s]
Generating train split: 55438178 examples [12:26, 83145.27 examples/s]
Generating train split: 55455178 examples [12:27, 69071.91 examples/s]
Generating train split: 55465178 examples [12:27, 42430.79 examples/s]
Generating train split: 55481178 examples [12:28, 50869.97 examples/s]
Generating train split: 55493178 examples [12:28, 48132.25 examples/s]
Generating train split: 55509178 examples [12:28, 43694.54 examples/s]
Generating train split: 55539178 examples [12:28, 68709.24 examples/s]
Generating train split: 55552178 examples [12:29, 76644.14 examples/s]
Generating train split: 55567178 examples [12:29, 81813.70 examples/s]
Generating train split: 55584178 examples [12:29, 69299.88 examples/s]
Generating train split: 55595178 examples [12:29, 59575.87 examples/s]
Generating train split: 55606178 examples [12:30, 61150.52 examples/s]
Generating train split: 55617178 examples [12:30, 61413.63 examples/s]
Generating train split: 55638178 examples [12:30, 86170.30 examples/s]
Generating train split: 55650178 examples [12:30, 79492.33 examples/s]
Generating train split: 55672178 examples [12:30, 104638.93 examples/s]
Generating train split: 55702178 examples [12:30, 114759.92 examples/s]
Generating train split: 55718178 examples [12:30, 110544.12 examples/s]
Generating train split: 55739178 examples [12:31, 129749.05 examples/s]
Generating train split: 55762178 examples [12:31, 78749.41 examples/s]
Generating train split: 55778178 examples [12:31, 66755.95 examples/s]
Generating train split: 55794178 examples [12:32, 49316.88 examples/s]
Generating train split: 55804178 examples [12:32, 53520.29 examples/s]
Generating train split: 55830178 examples [12:32, 75373.25 examples/s]
Generating train split: 55844178 examples [12:33, 68479.73 examples/s]
Generating train split: 55854178 examples [12:33, 48425.78 examples/s]
Generating train split: 55877178 examples [12:33, 67518.98 examples/s]
Generating train split: 55896178 examples [12:33, 83190.55 examples/s]
Generating train split: 55917178 examples [12:33, 103028.92 examples/s]
Generating train split: 55949178 examples [12:33, 134581.45 examples/s]
Generating train split: 55970178 examples [12:34, 99924.79 examples/s]
Generating train split: 55987178 examples [12:34, 94321.48 examples/s]
Generating train split: 56003178 examples [12:34, 90702.77 examples/s]
Generating train split: 56018178 examples [12:35, 65094.41 examples/s]
Generating train split: 56052178 examples [12:35, 101405.49 examples/s]
Generating train split: 56075178 examples [12:35, 101439.33 examples/s]
Generating train split: 56091178 examples [12:35, 102178.19 examples/s]
Generating train split: 56107178 examples [12:36, 57513.46 examples/s]
Generating train split: 56118178 examples [12:36, 52664.08 examples/s]
Generating train split: 56135178 examples [12:36, 64872.19 examples/s]
Generating train split: 56149178 examples [12:37, 47331.93 examples/s]
Generating train split: 56160178 examples [12:37, 50908.62 examples/s]
Generating train split: 56171178 examples [12:37, 54555.61 examples/s]
Generating train split: 56182178 examples [12:37, 62465.82 examples/s]
Generating train split: 56203178 examples [12:37, 86743.72 examples/s]
Generating train split: 56219178 examples [12:37, 85016.77 examples/s]
Generating train split: 56234178 examples [12:38, 59657.68 examples/s]
Generating train split: 56250178 examples [12:38, 69652.28 examples/s]
Generating train split: 56271178 examples [12:38, 91915.74 examples/s]
Generating train split: 56284178 examples [12:38, 68996.54 examples/s]
Generating train split: 56310178 examples [12:39, 92526.76 examples/s]
Generating train split: 56323178 examples [12:39, 76842.58 examples/s]
Generating train split: 56339178 examples [12:39, 72819.72 examples/s]
Generating train split: 56351178 examples [12:39, 72778.07 examples/s]
Generating train split: 56383178 examples [12:39, 110793.27 examples/s]
Generating train split: 56411178 examples [12:40, 131555.12 examples/s]
Generating train split: 56429178 examples [12:40, 71184.78 examples/s]
Generating train split: 56446178 examples [12:40, 81410.82 examples/s]
Generating train split: 56463178 examples [12:41, 59704.03 examples/s]
Generating train split: 56478178 examples [12:41, 67209.39 examples/s]
Generating train split: 56491178 examples [12:41, 57026.06 examples/s]
Generating train split: 56502178 examples [12:42, 45765.45 examples/s]
Generating train split: 56519178 examples [12:42, 57491.19 examples/s]
Generating train split: 56535178 examples [12:42, 61936.18 examples/s]
Generating train split: 56558178 examples [12:42, 86188.77 examples/s]
Generating train split: 56579178 examples [12:42, 70575.57 examples/s]
Generating train split: 56590178 examples [12:43, 74881.87 examples/s]
Generating train split: 56604178 examples [12:43, 79966.34 examples/s]
Generating train split: 56615178 examples [12:43, 82935.79 examples/s]
Generating train split: 56627178 examples [12:43, 56878.91 examples/s]
Generating train split: 56638178 examples [12:43, 62850.77 examples/s]
Generating train split: 56665178 examples [12:43, 97708.98 examples/s]
Generating train split: 56685178 examples [12:44, 113825.28 examples/s]
Generating train split: 56700178 examples [12:44, 75530.19 examples/s]
Generating train split: 56725178 examples [12:44, 101166.28 examples/s]
Generating train split: 56741178 examples [12:44, 85046.68 examples/s]
Generating train split: 56757178 examples [12:45, 90287.74 examples/s]
Generating train split: 56773178 examples [12:45, 67028.67 examples/s]
Generating train split: 56784178 examples [12:45, 46917.47 examples/s]
Generating train split: 56806178 examples [12:46, 66700.08 examples/s]
Generating train split: 56821178 examples [12:46, 62248.09 examples/s]
Generating train split: 56838178 examples [12:46, 75965.20 examples/s]
Generating train split: 56853178 examples [12:47, 47834.41 examples/s]
Generating train split: 56863178 examples [12:47, 53227.57 examples/s]
Generating train split: 56880178 examples [12:47, 67576.41 examples/s]
Generating train split: 56898178 examples [12:47, 83775.07 examples/s]
Generating train split: 56913178 examples [12:47, 84694.37 examples/s]
Generating train split: 56925178 examples [12:47, 85625.27 examples/s]
Generating train split: 56940178 examples [12:47, 79636.78 examples/s]
Generating train split: 56961178 examples [12:47, 104055.14 examples/s]
Generating train split: 56976178 examples [12:48, 109244.88 examples/s]
Generating train split: 56999178 examples [12:48, 122144.74 examples/s]
Generating train split: 57016178 examples [12:48, 83946.30 examples/s]
Generating train split: 57055178 examples [12:48, 129301.07 examples/s]
Generating train split: 57073178 examples [12:49, 59336.51 examples/s]
Generating train split: 57088178 examples [12:49, 62477.38 examples/s]
Generating train split: 57100178 examples [12:49, 61520.90 examples/s]
Generating train split: 57111178 examples [12:50, 61339.01 examples/s]
Generating train split: 57122178 examples [12:50, 44608.64 examples/s]
Generating train split: 57137178 examples [12:50, 53629.77 examples/s]
Generating train split: 57148178 examples [12:50, 55751.64 examples/s]
Generating train split: 57164178 examples [12:51, 71040.69 examples/s]
Generating train split: 57192178 examples [12:51, 99933.67 examples/s]
Generating train split: 57210178 examples [12:51, 60127.97 examples/s]
Generating train split: 57222178 examples [12:51, 67182.43 examples/s]
Generating train split: 57254178 examples [12:52, 95117.16 examples/s]
Generating train split: 57271178 examples [12:52, 98395.60 examples/s]
Generating train split: 57298178 examples [12:52, 116512.16 examples/s]
Generating train split: 57314178 examples [12:52, 83728.12 examples/s]
Generating train split: 57342178 examples [12:52, 106910.99 examples/s]
Generating train split: 57358178 examples [12:53, 79733.68 examples/s]
Generating train split: 57393178 examples [12:53, 112384.04 examples/s]
Generating train split: 57411178 examples [12:54, 55960.50 examples/s]
Generating train split: 57428178 examples [12:54, 55537.22 examples/s]
Generating train split: 57445178 examples [12:54, 66994.17 examples/s]
Generating train split: 57461178 examples [12:54, 60576.86 examples/s]
Generating train split: 57472178 examples [12:55, 60059.38 examples/s]
Generating train split: 57483178 examples [12:55, 44807.81 examples/s]
Generating train split: 57493178 examples [12:55, 47308.62 examples/s]
Generating train split: 57515178 examples [12:56, 61158.21 examples/s]
Generating train split: 57536178 examples [12:56, 66437.41 examples/s]
Generating train split: 57559178 examples [12:56, 74794.08 examples/s]
Generating train split: 57580178 examples [12:56, 88106.82 examples/s]
Generating train split: 57602178 examples [12:56, 108270.07 examples/s]
Generating train split: 57618178 examples [12:56, 109172.10 examples/s]
Generating train split: 57634178 examples [12:57, 115108.38 examples/s]
Generating train split: 57650178 examples [12:57, 78140.88 examples/s]
Generating train split: 57667178 examples [12:57, 91974.26 examples/s]
Generating train split: 57688178 examples [12:57, 112259.03 examples/s]
Generating train split: 57703178 examples [12:57, 81234.20 examples/s]
Generating train split: 57720178 examples [12:58, 95443.37 examples/s]
Generating train split: 57741178 examples [12:58, 101927.60 examples/s]
Generating train split: 57758178 examples [12:58, 81851.27 examples/s]
Generating train split: 57770178 examples [12:58, 80884.45 examples/s]
Generating train split: 57781178 examples [12:59, 62156.38 examples/s]
Generating train split: 57792178 examples [12:59, 61032.24 examples/s]
Generating train split: 57805178 examples [12:59, 43907.39 examples/s]
Generating train split: 57815178 examples [12:59, 47317.95 examples/s]
Generating train split: 57826178 examples [13:00, 53898.68 examples/s]
Generating train split: 57838178 examples [13:00, 49835.21 examples/s]
Generating train split: 57849178 examples [13:00, 51826.54 examples/s]
Generating train split: 57878178 examples [13:00, 77255.82 examples/s]
Generating train split: 57903178 examples [13:00, 103654.29 examples/s]
Generating train split: 57920178 examples [13:00, 98358.02 examples/s]
Generating train split: 57936178 examples [13:01, 86378.38 examples/s]
Generating train split: 57956178 examples [13:01, 87405.78 examples/s]
Generating train split: 57973178 examples [13:01, 99693.75 examples/s]
Generating train split: 57995178 examples [13:01, 116167.66 examples/s]
Generating train split: 58011178 examples [13:02, 77325.18 examples/s]
Generating train split: 58024178 examples [13:02, 70773.84 examples/s]
Generating train split: 58034178 examples [13:02, 54773.16 examples/s]
Generating train split: 58057178 examples [13:02, 78609.02 examples/s]
Generating train split: 58076178 examples [13:02, 84223.51 examples/s]
Generating train split: 58092178 examples [13:03, 89290.86 examples/s]
Generating train split: 58103178 examples [13:03, 68144.39 examples/s]
Generating train split: 58115178 examples [13:03, 47328.63 examples/s]
Generating train split: 58126178 examples [13:04, 43350.92 examples/s]
Generating train split: 58141178 examples [13:04, 53521.33 examples/s]
Generating train split: 58151178 examples [13:04, 52481.85 examples/s]
Generating train split: 58169178 examples [13:04, 69520.92 examples/s]
Generating train split: 58180178 examples [13:05, 48364.14 examples/s]
Generating train split: 58198178 examples [13:05, 62375.11 examples/s]
Generating train split: 58210178 examples [13:05, 70170.97 examples/s]
Generating train split: 58240178 examples [13:05, 107225.33 examples/s]
Generating train split: 58256178 examples [13:05, 103668.35 examples/s]
Generating train split: 58271178 examples [13:05, 88147.30 examples/s]
Generating train split: 58285178 examples [13:06, 97120.61 examples/s]
Generating train split: 58301178 examples [13:06, 94288.72 examples/s]
Generating train split: 58330178 examples [13:06, 130140.37 examples/s]
Generating train split: 58350178 examples [13:06, 123922.34 examples/s]
Generating train split: 58366178 examples [13:06, 93536.76 examples/s]
Generating train split: 58396178 examples [13:07, 78189.55 examples/s]
Generating train split: 58407178 examples [13:07, 64840.84 examples/s]
Generating train split: 58418178 examples [13:07, 69370.10 examples/s]
Generating train split: 58429178 examples [13:07, 74484.50 examples/s]
Generating train split: 58441178 examples [13:08, 61039.92 examples/s]
Generating train split: 58452178 examples [13:08, 48914.28 examples/s]
Generating train split: 58462178 examples [13:08, 55553.88 examples/s]
Generating train split: 58474178 examples [13:08, 45018.94 examples/s]
Generating train split: 58485178 examples [13:09, 53264.87 examples/s]
Generating train split: 58497178 examples [13:09, 56062.10 examples/s]
Generating train split: 58508178 examples [13:09, 57703.22 examples/s]
Generating train split: 58518178 examples [13:09, 46335.82 examples/s]
Generating train split: 58530178 examples [13:09, 57056.07 examples/s]
Generating train split: 58556178 examples [13:09, 91777.10 examples/s]
Generating train split: 58573178 examples [13:10, 98830.64 examples/s]
Generating train split: 58589178 examples [13:10, 99948.29 examples/s]
Generating train split: 58607178 examples [13:10, 108915.16 examples/s]
Generating train split: 58620178 examples [13:10, 95893.68 examples/s]
Generating train split: 58641178 examples [13:10, 116397.93 examples/s]
Generating train split: 58658178 examples [13:10, 108256.08 examples/s]
Generating train split: 58691178 examples [13:11, 82905.05 examples/s]
Generating train split: 58707178 examples [13:11, 90286.47 examples/s]
Generating train split: 58722178 examples [13:11, 97855.98 examples/s]
Generating train split: 58734178 examples [13:12, 53083.52 examples/s]
Generating train split: 58746178 examples [13:12, 56736.52 examples/s]
Generating train split: 58756178 examples [13:12, 48663.33 examples/s]
Generating train split: 58777178 examples [13:13, 49409.43 examples/s]
Generating train split: 58788178 examples [13:13, 54303.17 examples/s]
Generating train split: 58807178 examples [13:13, 66012.41 examples/s]
Generating train split: 58817178 examples [13:13, 69849.69 examples/s]
Generating train split: 58829178 examples [13:13, 69932.24 examples/s]
Generating train split: 58844178 examples [13:13, 80277.02 examples/s]
Generating train split: 58857178 examples [13:14, 72942.38 examples/s]
Generating train split: 58867178 examples [13:14, 63958.91 examples/s]
Generating train split: 58879178 examples [13:14, 57215.95 examples/s]
Generating train split: 58901178 examples [13:14, 77533.66 examples/s]
Generating train split: 58934178 examples [13:14, 101737.82 examples/s]
Generating train split: 58955178 examples [13:14, 120429.88 examples/s]
Generating train split: 58970178 examples [13:15, 74845.61 examples/s]
Generating train split: 58995178 examples [13:15, 97629.87 examples/s]
Generating train split: 59023178 examples [13:16, 72192.47 examples/s]
Generating train split: 59036178 examples [13:16, 75068.06 examples/s]
Generating train split: 59057178 examples [13:16, 70657.35 examples/s]
Generating train split: 59067178 examples [13:16, 73827.54 examples/s]
Generating train split: 59077178 examples [13:16, 73012.67 examples/s]
Generating train split: 59093178 examples [13:17, 74556.40 examples/s]
Generating train split: 59104178 examples [13:17, 39706.17 examples/s]
Generating train split: 59121178 examples [13:17, 51015.30 examples/s]
Generating train split: 59137178 examples [13:18, 63164.31 examples/s]
Generating train split: 59159178 examples [13:18, 85176.42 examples/s]
Generating train split: 59174178 examples [13:18, 63813.33 examples/s]
Generating train split: 59196178 examples [13:18, 79537.99 examples/s]
Generating train split: 59211178 examples [13:18, 81579.94 examples/s]
Generating train split: 59231178 examples [13:18, 100122.97 examples/s]
Generating train split: 59247178 examples [13:19, 87570.35 examples/s]
Generating train split: 59270178 examples [13:19, 105199.91 examples/s]
Generating train split: 59286178 examples [13:19, 113743.92 examples/s]
Generating train split: 59303178 examples [13:19, 116250.79 examples/s]
Generating train split: 59338178 examples [13:19, 160165.01 examples/s]
Generating train split: 59358178 examples [13:20, 97489.73 examples/s]
Generating train split: 59373178 examples [13:20, 86794.50 examples/s]
Generating train split: 59389178 examples [13:20, 62630.92 examples/s]
Generating train split: 59410178 examples [13:20, 80324.81 examples/s]
Generating train split: 59426178 examples [13:21, 81690.44 examples/s]
Generating train split: 59442178 examples [13:21, 61906.63 examples/s]
Generating train split: 59456178 examples [13:22, 47518.05 examples/s]
Generating train split: 59468178 examples [13:22, 42267.08 examples/s]
Generating train split: 59484178 examples [13:22, 53765.91 examples/s]
Generating train split: 59498178 examples [13:22, 57779.58 examples/s]
Generating train split: 59515178 examples [13:22, 62856.72 examples/s]
Generating train split: 59525178 examples [13:23, 56245.71 examples/s]
Generating train split: 59546178 examples [13:23, 77364.80 examples/s]
Generating train split: 59557178 examples [13:23, 76450.44 examples/s]
Generating train split: 59573178 examples [13:23, 91158.19 examples/s]
Generating train split: 59588178 examples [13:23, 91927.67 examples/s]
Generating train split: 59603178 examples [13:24, 77910.43 examples/s]
Generating train split: 59622178 examples [13:24, 96833.01 examples/s]
Generating train split: 59638178 examples [13:24, 87613.15 examples/s]
Generating train split: 59665178 examples [13:24, 121617.12 examples/s]
Generating train split: 59685178 examples [13:24, 84175.88 examples/s]
Generating train split: 59703178 examples [13:24, 94411.59 examples/s]
Generating train split: 59720178 examples [13:25, 62101.43 examples/s]
Generating train split: 59734178 examples [13:25, 68350.55 examples/s]
Generating train split: 59746178 examples [13:25, 70514.41 examples/s]
Generating train split: 59757178 examples [13:26, 49435.04 examples/s]
Generating train split: 59767178 examples [13:26, 38168.05 examples/s]
Generating train split: 59785178 examples [13:26, 53437.66 examples/s]
Generating train split: 59801178 examples [13:26, 67170.29 examples/s]
Generating train split: 59813178 examples [13:27, 47017.09 examples/s]
Generating train split: 59835178 examples [13:27, 51181.82 examples/s]
Generating train split: 59844178 examples [13:27, 55290.22 examples/s]
Generating train split: 59854178 examples [13:28, 56048.06 examples/s]
Generating train split: 59871178 examples [13:28, 70675.83 examples/s]
Generating train split: 59899178 examples [13:28, 105891.69 examples/s]
Generating train split: 59916178 examples [13:28, 114551.83 examples/s]
Generating train split: 59933178 examples [13:28, 122588.28 examples/s]
Generating train split: 59948178 examples [13:28, 87288.89 examples/s]
Generating train split: 59963178 examples [13:29, 81199.55 examples/s]
Generating train split: 59985178 examples [13:29, 104189.38 examples/s]
Generating train split: 60006178 examples [13:29, 80424.90 examples/s]
Generating train split: 60028178 examples [13:29, 100353.41 examples/s]
Generating train split: 60050178 examples [13:29, 117650.31 examples/s]
Generating train split: 60070178 examples [13:30, 73813.06 examples/s]
Generating train split: 60086178 examples [13:30, 82064.04 examples/s]
Generating train split: 60102178 examples [13:31, 38998.14 examples/s]
Generating train split: 60130178 examples [13:31, 58939.75 examples/s]
Generating train split: 60147178 examples [13:31, 57120.20 examples/s]
Generating train split: 60159178 examples [13:32, 59364.79 examples/s]
Generating train split: 60171178 examples [13:32, 62424.96 examples/s]
Generating train split: 60186178 examples [13:32, 74349.24 examples/s]
Generating train split: 60201178 examples [13:32, 68310.03 examples/s]
Generating train split: 60222178 examples [13:32, 90370.86 examples/s]
Generating train split: 60238178 examples [13:32, 82691.87 examples/s]
Generating train split: 60265178 examples [13:32, 114211.81 examples/s]
Generating train split: 60283178 examples [13:33, 105735.54 examples/s]
Generating train split: 60301178 examples [13:33, 83407.97 examples/s]
Generating train split: 60316178 examples [13:33, 79887.20 examples/s]
Generating train split: 60337178 examples [13:33, 97907.07 examples/s]
Generating train split: 60353686 examples [13:34, 74469.43 examples/s]
Generating train split: 60374686 examples [13:34, 94008.39 examples/s]
Generating train split: 60398686 examples [13:34, 67838.59 examples/s]
Generating train split: 60410686 examples [13:34, 70539.35 examples/s]
Generating train split: 60420686 examples [13:35, 46983.74 examples/s]
Generating train split: 60436686 examples [13:35, 59502.93 examples/s]
Generating train split: 60447686 examples [13:35, 50831.33 examples/s]
Generating train split: 60458686 examples [13:36, 55300.59 examples/s]
Generating train split: 60484686 examples [13:36, 82652.25 examples/s]
Generating train split: 60497686 examples [13:36, 64094.25 examples/s]
Generating train split: 60507686 examples [13:36, 63834.13 examples/s]
Generating train split: 60520686 examples [13:36, 61578.62 examples/s]
Generating train split: 60537686 examples [13:37, 73220.52 examples/s]
Generating train split: 60559686 examples [13:37, 84997.26 examples/s]
Generating train split: 60570686 examples [13:37, 79924.12 examples/s]
Generating train split: 60593686 examples [13:37, 101244.52 examples/s]
Generating train split: 60614686 examples [13:37, 119398.17 examples/s]
Generating train split: 60631686 examples [13:37, 107153.32 examples/s]
Generating train split: 60647686 examples [13:38, 89659.98 examples/s]
Generating train split: 60664686 examples [13:38, 98431.04 examples/s]
Generating train split: 60680686 examples [13:38, 109564.13 examples/s]
Generating train split: 60696686 examples [13:38, 68194.46 examples/s]
Generating train split: 60724686 examples [13:38, 97544.86 examples/s]
Generating train split: 60742686 examples [13:40, 40433.38 examples/s]
Generating train split: 60760686 examples [13:40, 50364.58 examples/s]
Generating train split: 60776686 examples [13:40, 53509.49 examples/s]
Generating train split: 60789686 examples [13:40, 61871.93 examples/s]
Generating train split: 60805686 examples [13:40, 61062.06 examples/s]
Generating train split: 60816686 examples [13:41, 46127.34 examples/s]
Generating train split: 60833686 examples [13:41, 58653.86 examples/s]
Generating train split: 60853686 examples [13:41, 77741.66 examples/s]
Generating train split: 60889686 examples [13:41, 115936.03 examples/s]
Generating train split: 60911686 examples [13:41, 131744.84 examples/s]
Generating train split: 60932686 examples [13:42, 102425.03 examples/s]
Generating train split: 60947686 examples [13:42, 103288.91 examples/s]
Generating train split: 60974686 examples [13:42, 87961.65 examples/s]
Generating train split: 60986686 examples [13:42, 84095.77 examples/s]
Generating train split: 61019686 examples [13:42, 121301.21 examples/s]
Generating train split: 61035686 examples [13:43, 70540.37 examples/s]
Generating train split: 61052686 examples [13:44, 38368.64 examples/s]
Generating train split: 61084686 examples [13:44, 58783.94 examples/s]
Generating train split: 61098686 examples [13:44, 55191.07 examples/s]
Generating train split: 61113686 examples [13:45, 64449.79 examples/s]
Generating train split: 61127686 examples [13:45, 59985.03 examples/s]
Generating train split: 61140686 examples [13:45, 65564.41 examples/s]
Generating train split: 61156686 examples [13:45, 77617.22 examples/s]
Generating train split: 61170686 examples [13:45, 80476.14 examples/s]
Generating train split: 61186686 examples [13:45, 76077.73 examples/s]
Generating train split: 61214686 examples [13:46, 109015.81 examples/s]
Generating train split: 61234686 examples [13:46, 122057.47 examples/s]
Generating train split: 61252686 examples [13:46, 121931.87 examples/s]
Generating train split: 61269686 examples [13:46, 81376.50 examples/s]
Generating train split: 61282686 examples [13:47, 62654.51 examples/s]
Generating train split: 61293686 examples [13:47, 67916.84 examples/s]
Generating train split: 61311686 examples [13:47, 85866.62 examples/s]
Generating train split: 61333686 examples [13:47, 109883.00 examples/s]
Generating train split: 61350686 examples [13:48, 41800.55 examples/s]
Generating train split: 61373686 examples [13:48, 58579.71 examples/s]
Generating train split: 61387686 examples [13:48, 51367.80 examples/s]
Generating train split: 61400686 examples [13:49, 43301.08 examples/s]
Generating train split: 61416686 examples [13:49, 54592.28 examples/s]
Generating train split: 61433686 examples [13:49, 66650.99 examples/s]
Generating train split: 61447686 examples [13:49, 73624.29 examples/s]
Generating train split: 61469686 examples [13:49, 92871.84 examples/s]
Generating train split: 61486686 examples [13:50, 101807.88 examples/s]
Generating train split: 61515686 examples [13:50, 136656.71 examples/s]
Generating train split: 61532686 examples [13:50, 87899.57 examples/s]
Generating train split: 61547686 examples [13:50, 96771.32 examples/s]
Generating train split: 61576686 examples [13:50, 131299.80 examples/s]
Generating train split: 61594686 examples [13:50, 138530.49 examples/s]
Generating train split: 61616686 examples [13:51, 103382.57 examples/s]
Generating train split: 61637686 examples [13:51, 112612.71 examples/s]
Generating train split: 61652686 examples [13:51, 67898.79 examples/s]
Generating train split: 61667686 examples [13:51, 76974.31 examples/s]
Generating train split: 61680686 examples [13:52, 37137.32 examples/s]
Generating train split: 61713686 examples [13:53, 61292.65 examples/s]
Generating train split: 61731686 examples [13:53, 53647.66 examples/s]
Generating train split: 61747686 examples [13:53, 54274.78 examples/s]
Generating train split: 61759686 examples [13:53, 58517.26 examples/s]
Generating train split: 61776686 examples [13:54, 60016.09 examples/s]
Generating train split: 61787686 examples [13:54, 60098.01 examples/s]
Generating train split: 61809686 examples [13:54, 80450.03 examples/s]
Generating train split: 61826686 examples [13:54, 92469.07 examples/s]
Generating train split: 61843686 examples [13:54, 105922.76 examples/s]
Generating train split: 61863686 examples [13:54, 117475.41 examples/s]
Generating train split: 61879686 examples [13:54, 124638.93 examples/s]
Generating train split: 61896686 examples [13:55, 111281.10 examples/s]
Generating train split: 61929686 examples [13:55, 150143.11 examples/s]
Generating train split: 61949686 examples [13:55, 76868.28 examples/s]
Generating train split: 61972686 examples [13:55, 95436.27 examples/s]
Generating train split: 61989686 examples [13:56, 63245.05 examples/s]
Generating train split: 62005686 examples [13:57, 38965.04 examples/s]
Generating train split: 62021686 examples [13:57, 48225.87 examples/s]
Generating train split: 62048686 examples [13:57, 56743.83 examples/s]
Generating train split: 62058686 examples [13:58, 51349.76 examples/s]
Generating train split: 62069686 examples [13:58, 54302.49 examples/s]
Generating train split: 62079686 examples [13:58, 59733.46 examples/s]
Generating train split: 62091686 examples [13:58, 64890.31 examples/s]
Generating train split: 62103686 examples [13:58, 58128.93 examples/s]
Generating train split: 62114686 examples [13:58, 62036.91 examples/s]
Generating train split: 62127686 examples [13:59, 73890.93 examples/s]
Generating train split: 62154686 examples [13:59, 108401.89 examples/s]
Generating train split: 62183686 examples [13:59, 136037.11 examples/s]
Generating train split: 62202686 examples [13:59, 131172.07 examples/s]
Generating train split: 62218686 examples [13:59, 103955.46 examples/s]
Generating train split: 62246686 examples [13:59, 136237.52 examples/s]
Generating train split: 62267686 examples [14:00, 94935.29 examples/s]
Generating train split: 62283686 examples [14:00, 92808.47 examples/s]
Generating train split: 62310686 examples [14:00, 121481.58 examples/s]
Generating train split: 62328686 examples [14:00, 73505.67 examples/s]
Generating train split: 62345686 examples [14:01, 58784.51 examples/s]
Generating train split: 62362686 examples [14:01, 69904.94 examples/s]
Generating train split: 62377686 examples [14:02, 53277.31 examples/s]
Generating train split: 62393686 examples [14:02, 56804.93 examples/s]
Generating train split: 62404686 examples [14:02, 47731.06 examples/s]
Generating train split: 62422686 examples [14:02, 61723.66 examples/s]
Generating train split: 62434686 examples [14:02, 63325.92 examples/s]
Generating train split: 62456686 examples [14:03, 85945.18 examples/s]
Generating train split: 62469686 examples [14:03, 86104.28 examples/s]
Generating train split: 62487686 examples [14:03, 102401.79 examples/s]
Generating train split: 62503686 examples [14:03, 93516.14 examples/s]
Generating train split: 62528686 examples [14:03, 114770.00 examples/s]
Generating train split: 62544686 examples [14:03, 116132.35 examples/s]
Generating train split: 62561686 examples [14:03, 119681.18 examples/s]
Generating train split: 62579686 examples [14:04, 86952.55 examples/s]
Generating train split: 62593686 examples [14:04, 49924.48 examples/s]
Generating train split: 62615686 examples [14:04, 68395.24 examples/s]
Generating train split: 62632686 examples [14:05, 54856.56 examples/s]
Generating train split: 62648686 examples [14:05, 52426.10 examples/s]
Generating train split: 62660686 examples [14:05, 59346.34 examples/s]
Generating train split: 62681686 examples [14:06, 44734.73 examples/s]
Generating train split: 62699686 examples [14:06, 53368.42 examples/s]
Generating train split: 62712686 examples [14:06, 58898.39 examples/s]
Generating train split: 62722686 examples [14:07, 56710.68 examples/s]
Generating train split: 62733686 examples [14:07, 61626.15 examples/s]
Generating train split: 62745686 examples [14:07, 66106.96 examples/s]
Generating train split: 62764686 examples [14:07, 88114.03 examples/s]
Generating train split: 62777686 examples [14:07, 91690.05 examples/s]
Generating train split: 62796686 examples [14:07, 108128.58 examples/s]
Generating train split: 62811686 examples [14:07, 116810.38 examples/s]
Generating train split: 62836686 examples [14:07, 135952.32 examples/s]
Generating train split: 62852686 examples [14:08, 137642.88 examples/s]
Generating train split: 62872686 examples [14:08, 146067.15 examples/s]
Generating train split: 62898686 examples [14:08, 160831.31 examples/s]
Generating train split: 62918686 examples [14:09, 69897.27 examples/s]
Generating train split: 62933686 examples [14:09, 67553.98 examples/s]
Generating train split: 62949686 examples [14:09, 78296.50 examples/s]
Generating train split: 62972686 examples [14:09, 95989.45 examples/s]
Generating train split: 62987686 examples [14:09, 70329.65 examples/s]
Generating train split: 62998686 examples [14:10, 50176.42 examples/s]
Generating train split: 63022686 examples [14:11, 44998.23 examples/s]
Generating train split: 63044686 examples [14:11, 59343.94 examples/s]
Generating train split: 63059686 examples [14:11, 69458.91 examples/s]
Generating train split: 63078686 examples [14:11, 81768.16 examples/s]
Generating train split: 63090686 examples [14:11, 53127.10 examples/s]
Generating train split: 63101686 examples [14:12, 59467.20 examples/s]
Generating train split: 63123686 examples [14:12, 78434.82 examples/s]
Generating train split: 63158686 examples [14:12, 114909.61 examples/s]
Generating train split: 63178686 examples [14:12, 118064.57 examples/s]
Generating train split: 63194686 examples [14:12, 125540.56 examples/s]
Generating train split: 63216686 examples [14:12, 131825.04 examples/s]
Generating train split: 63233686 examples [14:12, 128309.40 examples/s]
Generating train split: 63251686 examples [14:13, 105313.77 examples/s]
Generating train split: 63268686 examples [14:13, 97156.57 examples/s]
Generating train split: 63284686 examples [14:13, 107943.62 examples/s]
Generating train split: 63302686 examples [14:14, 53002.84 examples/s]
Generating train split: 63313686 examples [14:14, 49866.82 examples/s]
Generating train split: 63329686 examples [14:14, 61805.10 examples/s]
Generating train split: 63339686 examples [14:14, 45365.41 examples/s]
Generating train split: 63360686 examples [14:15, 52014.14 examples/s]
Generating train split: 63377686 examples [14:15, 56988.94 examples/s]
Generating train split: 63397686 examples [14:15, 72567.61 examples/s]
Generating train split: 63413686 examples [14:15, 79738.67 examples/s]
Generating train split: 63425686 examples [14:16, 61901.95 examples/s]
Generating train split: 63445686 examples [14:16, 77402.90 examples/s]
Generating train split: 63455686 examples [14:16, 80922.01 examples/s]
Generating train split: 63476686 examples [14:16, 105411.86 examples/s]
Generating train split: 63501686 examples [14:16, 125857.81 examples/s]
Generating train split: 63520686 examples [14:16, 126549.56 examples/s]
Generating train split: 63536686 examples [14:16, 130444.26 examples/s]
Generating train split: 63556686 examples [14:17, 134039.51 examples/s]
Generating train split: 63576686 examples [14:17, 77095.71 examples/s]
Generating train split: 63589686 examples [14:17, 67761.83 examples/s]
Generating train split: 63615686 examples [14:18, 48473.10 examples/s]
Generating train split: 63625686 examples [14:18, 46402.59 examples/s]
Generating train split: 63634686 examples [14:18, 50643.06 examples/s]
Generating train split: 63651686 examples [14:19, 65840.56 examples/s]
Generating train split: 63664686 examples [14:19, 43213.15 examples/s]
Generating train split: 63674686 examples [14:19, 46089.16 examples/s]
Generating train split: 63697686 examples [14:19, 69485.90 examples/s]
Generating train split: 63719686 examples [14:19, 90991.98 examples/s]
Generating train split: 63736686 examples [14:20, 68602.30 examples/s]
Generating train split: 63749686 examples [14:20, 66356.48 examples/s]
Generating train split: 63761686 examples [14:20, 74148.28 examples/s]
Generating train split: 63773686 examples [14:20, 80718.32 examples/s]
Generating train split: 63788686 examples [14:20, 92061.06 examples/s]
Generating train split: 63811686 examples [14:21, 112867.55 examples/s]
Generating train split: 63824686 examples [14:21, 98298.47 examples/s]
Generating train split: 63845686 examples [14:21, 118853.62 examples/s]
Generating train split: 63863686 examples [14:21, 113497.95 examples/s]
Generating train split: 63880686 examples [14:21, 111265.93 examples/s]
Generating train split: 63896686 examples [14:22, 58951.93 examples/s]
Generating train split: 63913686 examples [14:22, 72350.44 examples/s]
Generating train split: 63930686 examples [14:23, 38166.86 examples/s]
Generating train split: 63959686 examples [14:23, 59624.82 examples/s]
Generating train split: 63976686 examples [14:23, 52583.44 examples/s]
Generating train split: 63994686 examples [14:24, 63814.90 examples/s]
Generating train split: 64009686 examples [14:24, 72623.30 examples/s]
Generating train split: 64029686 examples [14:24, 91030.44 examples/s]
Generating train split: 64046686 examples [14:24, 75530.45 examples/s]
Generating train split: 64062686 examples [14:24, 65717.28 examples/s]
Generating train split: 64081686 examples [14:24, 82577.17 examples/s]
Generating train split: 64099686 examples [14:25, 81788.15 examples/s]
Generating train split: 64122686 examples [14:25, 105170.93 examples/s]
Generating train split: 64140686 examples [14:25, 117880.44 examples/s]
Generating train split: 64160686 examples [14:25, 134243.23 examples/s]
Generating train split: 64180686 examples [14:25, 106834.91 examples/s]
Generating train split: 64202686 examples [14:25, 126688.66 examples/s]
Generating train split: 64218686 examples [14:26, 97173.30 examples/s]
Generating train split: 64239686 examples [14:26, 62003.25 examples/s]
Generating train split: 64250686 examples [14:26, 67031.84 examples/s]
Generating train split: 64264686 examples [14:27, 55499.08 examples/s]
Generating train split: 64274686 examples [14:27, 54871.51 examples/s]
Generating train split: 64284686 examples [14:27, 44787.93 examples/s]
Generating train split: 64294686 examples [14:27, 51848.29 examples/s]
Generating train split: 64304686 examples [14:28, 58740.59 examples/s]
Generating train split: 64314686 examples [14:28, 61463.17 examples/s]
Generating train split: 64326686 examples [14:28, 64724.69 examples/s]
Generating train split: 64339686 examples [14:28, 74532.66 examples/s]
Generating train split: 64352686 examples [14:28, 76571.41 examples/s]
Generating train split: 64362686 examples [14:28, 79401.92 examples/s]
Generating train split: 64383686 examples [14:28, 108649.34 examples/s]
Generating train split: 64399686 examples [14:29, 95392.15 examples/s]
Generating train split: 64411686 examples [14:29, 55748.03 examples/s]
Generating train split: 64423686 examples [14:29, 60841.87 examples/s]
Generating train split: 64450686 examples [14:29, 94078.77 examples/s]
Generating train split: 64482686 examples [14:29, 128133.22 examples/s]
Generating train split: 64501686 examples [14:30, 123762.81 examples/s]
Generating train split: 64518686 examples [14:30, 82019.65 examples/s]
Generating train split: 64530686 examples [14:30, 73035.79 examples/s]
Generating train split: 64546686 examples [14:30, 85488.03 examples/s]
Generating train split: 64561686 examples [14:30, 91545.77 examples/s]
Generating train split: 64577686 examples [14:31, 42966.45 examples/s]
Generating train split: 64589686 examples [14:31, 47061.13 examples/s]
Generating train split: 64607686 examples [14:32, 62753.36 examples/s]
Generating train split: 64619686 examples [14:32, 49435.05 examples/s]
Generating train split: 64630686 examples [14:32, 50861.18 examples/s]
Generating train split: 64649686 examples [14:32, 70038.01 examples/s]
Generating train split: 64670686 examples [14:32, 85106.08 examples/s]
Generating train split: 64682686 examples [14:33, 66755.66 examples/s]
Generating train split: 64693686 examples [14:33, 63914.57 examples/s]
Generating train split: 64714686 examples [14:33, 80072.15 examples/s]
Generating train split: 64730686 examples [14:33, 89531.77 examples/s]
Generating train split: 64750686 examples [14:33, 105245.66 examples/s]
Generating train split: 64765686 examples [14:34, 102022.96 examples/s]
Generating train split: 64778686 examples [14:34, 89427.00 examples/s]
Generating train split: 64795686 examples [14:34, 84639.74 examples/s]
Generating train split: 64807686 examples [14:34, 77047.46 examples/s]
Generating train split: 64817686 examples [14:34, 80622.53 examples/s]
Generating train split: 64833686 examples [14:34, 95836.32 examples/s]
Generating train split: 64847686 examples [14:34, 98026.35 examples/s]
Generating train split: 64859686 examples [14:35, 80969.15 examples/s]
Generating train split: 64875686 examples [14:35, 95205.22 examples/s]
Generating train split: 64891686 examples [14:35, 89845.06 examples/s]
Generating train split: 64902686 examples [14:35, 55657.59 examples/s]
Generating train split: 64914686 examples [14:36, 47406.85 examples/s]
Generating train split: 64930686 examples [14:36, 59118.55 examples/s]
Generating train split: 64941686 examples [14:36, 57935.06 examples/s]
Generating train split: 64952686 examples [14:36, 52272.39 examples/s]
Generating train split: 64964686 examples [14:37, 53169.19 examples/s]
Generating train split: 64985686 examples [14:37, 73908.59 examples/s]
Generating train split: 65005686 examples [14:37, 89784.19 examples/s]
Generating train split: 65017686 examples [14:37, 62093.23 examples/s]
Generating train split: 65039686 examples [14:38, 70602.65 examples/s]
Generating train split: 65056686 examples [14:38, 85063.02 examples/s]
Generating train split: 65085686 examples [14:38, 119885.84 examples/s]
Generating train split: 65102686 examples [14:38, 72892.65 examples/s]
Generating train split: 65118686 examples [14:38, 75209.92 examples/s]
Generating train split: 65132686 examples [14:39, 70385.69 examples/s]
Generating train split: 65157686 examples [14:39, 96703.08 examples/s]
Generating train split: 65182686 examples [14:39, 122297.50 examples/s]
Generating train split: 65202686 examples [14:40, 57714.87 examples/s]
Generating train split: 65218686 examples [14:40, 60526.38 examples/s]
Generating train split: 65234686 examples [14:40, 57670.74 examples/s]
Generating train split: 65244686 examples [14:40, 51450.12 examples/s]
Generating train split: 65279686 examples [14:41, 77153.16 examples/s]
Generating train split: 65296686 examples [14:41, 83477.25 examples/s]
Generating train split: 65312686 examples [14:41, 64389.24 examples/s]
Generating train split: 65338686 examples [14:41, 88505.93 examples/s]
Generating train split: 65354686 examples [14:42, 54206.26 examples/s]
Generating train split: 65408686 examples [14:42, 104958.94 examples/s]
Generating train split: 65434686 examples [14:42, 106898.85 examples/s]
Generating train split: 65455686 examples [14:43, 82305.03 examples/s]
Generating train split: 65470686 examples [14:43, 72158.39 examples/s]
Generating train split: 65492686 examples [14:43, 85756.90 examples/s]
Generating train split: 65508686 examples [14:44, 52088.46 examples/s]
Generating train split: 65528686 examples [14:44, 65104.75 examples/s]
Generating train split: 65544686 examples [14:44, 63979.59 examples/s]
Generating train split: 65559686 examples [14:45, 58012.75 examples/s]
Generating train split: 65570686 examples [14:45, 56805.77 examples/s]
Generating train split: 65587686 examples [14:45, 70629.17 examples/s]
Generating train split: 65613686 examples [14:45, 75154.49 examples/s]
Generating train split: 65623686 examples [14:46, 61473.52 examples/s]
Generating train split: 65634686 examples [14:46, 65872.04 examples/s]
Generating train split: 65646686 examples [14:46, 72675.74 examples/s]
Generating train split: 65666686 examples [14:46, 93081.85 examples/s]
Generating train split: 65696686 examples [14:46, 124362.30 examples/s]
Generating train split: 65711686 examples [14:46, 86858.68 examples/s]
Generating train split: 65726686 examples [14:47, 76255.37 examples/s]
Generating train split: 65742686 examples [14:47, 87149.69 examples/s]
Generating train split: 65753686 examples [14:47, 77901.60 examples/s]
Generating train split: 65776686 examples [14:47, 101353.76 examples/s]
Generating train split: 65791686 examples [14:48, 60567.47 examples/s]
Generating train split: 65801686 examples [14:48, 48822.79 examples/s]
Generating train split: 65812686 examples [14:48, 52020.36 examples/s]
Generating train split: 65823686 examples [14:48, 57586.89 examples/s]
Generating train split: 65834686 examples [14:48, 64951.54 examples/s]
Generating train split: 65845686 examples [14:49, 49036.40 examples/s]
Generating train split: 65855686 examples [14:49, 56320.03 examples/s]
Generating train split: 65885686 examples [14:49, 94122.15 examples/s]
Generating train split: 65900686 examples [14:49, 86183.27 examples/s]
Generating train split: 65915686 examples [14:49, 80694.51 examples/s]
Generating train split: 65926686 examples [14:50, 69184.17 examples/s]
Generating train split: 65941686 examples [14:50, 82216.83 examples/s]
Generating train split: 65956686 examples [14:50, 93971.16 examples/s]
Generating train split: 65970686 examples [14:50, 96226.79 examples/s]
Generating train split: 65981686 examples [14:50, 91894.44 examples/s]
Generating train split: 65997686 examples [14:50, 78366.25 examples/s]
Generating train split: 66017686 examples [14:51, 100637.06 examples/s]
Generating train split: 66029686 examples [14:51, 91232.53 examples/s]
Generating train split: 66040686 examples [14:51, 61510.34 examples/s]
Generating train split: 66054686 examples [14:51, 73455.58 examples/s]
Generating train split: 66065686 examples [14:51, 62770.29 examples/s]
Generating train split: 66090686 examples [14:52, 93979.78 examples/s]
Generating train split: 66120686 examples [14:52, 65170.02 examples/s]
Generating train split: 66135686 examples [14:52, 71132.88 examples/s]
Generating train split: 66160686 examples [14:52, 93492.50 examples/s]
Generating train split: 66177686 examples [14:53, 91758.35 examples/s]
Generating train split: 66190686 examples [14:53, 68016.25 examples/s]
Generating train split: 66202686 examples [14:53, 66337.50 examples/s]
Generating train split: 66220686 examples [14:53, 83496.99 examples/s]
Generating train split: 66240686 examples [14:53, 103397.42 examples/s]
Generating train split: 66254686 examples [14:54, 51621.94 examples/s]
Generating train split: 66280686 examples [14:54, 73972.71 examples/s]
Generating train split: 66295686 examples [14:54, 82058.62 examples/s]
Generating train split: 66313686 examples [14:55, 88090.21 examples/s]
Generating train split: 66330686 examples [14:55, 66292.71 examples/s]
Generating train split: 66340686 examples [14:55, 69776.20 examples/s]
Generating train split: 66351686 examples [14:55, 55696.56 examples/s]
Generating train split: 66380686 examples [14:55, 85724.77 examples/s]
Generating train split: 66402686 examples [14:56, 106109.89 examples/s]
Generating train split: 66418686 examples [14:56, 72290.23 examples/s]
Generating train split: 66433686 examples [14:56, 69845.49 examples/s]
Generating train split: 66445686 examples [14:57, 46520.93 examples/s]
Generating train split: 66456686 examples [14:57, 51714.65 examples/s]
Generating train split: 66471686 examples [14:57, 63754.45 examples/s]
Generating train split: 66484686 examples [14:57, 60941.91 examples/s]
Generating train split: 66494686 examples [14:57, 60411.95 examples/s]
Generating train split: 66507686 examples [14:58, 66273.61 examples/s]
Generating train split: 66523686 examples [14:58, 80862.64 examples/s]
Generating train split: 66534686 examples [14:58, 85420.30 examples/s]
Generating train split: 66549686 examples [14:58, 82810.33 examples/s]
Generating train split: 66563686 examples [14:58, 94604.45 examples/s]
Generating train split: 66578686 examples [14:58, 78860.37 examples/s]
Generating train split: 66589686 examples [14:59, 79672.42 examples/s]
Generating train split: 66600686 examples [14:59, 72175.41 examples/s]
Generating train split: 66611686 examples [14:59, 78656.82 examples/s]
Generating train split: 66627686 examples [14:59, 90526.22 examples/s]
Generating train split: 66643686 examples [14:59, 98687.26 examples/s]
Generating train split: 66654686 examples [15:00, 55636.90 examples/s]
Generating train split: 66665686 examples [15:00, 55298.84 examples/s]
Generating train split: 66676686 examples [15:00, 57249.31 examples/s]
Generating train split: 66693686 examples [15:00, 75689.93 examples/s]
Generating train split: 66723686 examples [15:00, 117067.26 examples/s]
Generating train split: 66739686 examples [15:01, 75573.28 examples/s]
Generating train split: 66751686 examples [15:01, 79692.76 examples/s]
Generating train split: 66768686 examples [15:01, 69514.23 examples/s]
Generating train split: 66779686 examples [15:01, 54714.54 examples/s]
Generating train split: 66788686 examples [15:01, 59137.74 examples/s]
Generating train split: 66799686 examples [15:02, 66908.15 examples/s]
Generating train split: 66811686 examples [15:02, 59228.62 examples/s]
Generating train split: 66823686 examples [15:02, 60640.77 examples/s]
Generating train split: 66838686 examples [15:02, 74468.24 examples/s]
Generating train split: 66852686 examples [15:02, 77413.05 examples/s]
Generating train split: 66867686 examples [15:02, 78967.73 examples/s]
Generating train split: 66890686 examples [15:03, 107097.70 examples/s]
Generating train split: 66905686 examples [15:03, 87772.67 examples/s]
Generating train split: 66923686 examples [15:03, 103041.49 examples/s]
Generating train split: 66952686 examples [15:03, 137174.68 examples/s]
Generating train split: 66988686 examples [15:03, 186700.41 examples/s]
Generating train split: 67018686 examples [15:03, 188810.96 examples/s]
Generating train split: 67044686 examples [15:03, 176757.12 examples/s]
Generating train split: 67066686 examples [15:04, 128631.71 examples/s]
Generating train split: 67086686 examples [15:04, 138460.45 examples/s]
Generating train split: 67106686 examples [15:04, 128443.92 examples/s]
Generating train split: 67125686 examples [15:04, 134025.91 examples/s]
Generating train split: 67148686 examples [15:04, 149279.13 examples/s]
Generating train split: 67168686 examples [15:05, 104610.24 examples/s]
Generating train split: 67185686 examples [15:05, 106939.88 examples/s]
Generating train split: 67202686 examples [15:05, 101442.24 examples/s]
Generating train split: 67219686 examples [15:05, 104973.53 examples/s]
Generating train split: 67235686 examples [15:05, 94848.95 examples/s]
Generating train split: 67246686 examples [15:05, 94345.96 examples/s]
Generating train split: 67261686 examples [15:06, 90806.59 examples/s]
Generating train split: 67282686 examples [15:06, 107669.20 examples/s]
Generating train split: 67300686 examples [15:06, 117072.12 examples/s]
Generating train split: 67323686 examples [15:06, 134867.01 examples/s]
Generating train split: 67343686 examples [15:06, 145935.17 examples/s]
Generating train split: 67360686 examples [15:06, 140731.54 examples/s]
Generating train split: 67377686 examples [15:07, 57879.37 examples/s]
Generating train split: 67395686 examples [15:07, 67142.05 examples/s]
Generating train split: 67412686 examples [15:07, 77289.94 examples/s]
Generating train split: 67429686 examples [15:08, 79380.08 examples/s]
Generating train split: 67440686 examples [15:08, 70337.56 examples/s]
Generating train split: 67451686 examples [15:08, 51327.62 examples/s]
Generating train split: 67463686 examples [15:09, 40627.86 examples/s]
Generating train split: 67473686 examples [15:09, 44976.00 examples/s]
Generating train split: 67481686 examples [15:09, 36260.55 examples/s]
Generating train split: 67491686 examples [15:09, 43265.45 examples/s]
Generating train split: 67502686 examples [15:10, 42861.44 examples/s]
Generating train split: 67513686 examples [15:10, 52005.66 examples/s]
Generating train split: 67534686 examples [15:10, 71980.56 examples/s]
Generating train split: 67544686 examples [15:10, 72758.53 examples/s]
Generating train split: 67558686 examples [15:10, 78389.01 examples/s]
Generating train split: 67568686 examples [15:10, 81494.27 examples/s]
Generating train split: 67584686 examples [15:10, 92764.78 examples/s]
Generating train split: 67599686 examples [15:10, 84726.85 examples/s]
Generating train split: 67631686 examples [15:11, 131155.11 examples/s]
Generating train split: 67647686 examples [15:11, 116880.48 examples/s]
Generating train split: 67662686 examples [15:11, 64238.86 examples/s]
Generating train split: 67676686 examples [15:12, 65898.60 examples/s]
Generating train split: 67687686 examples [15:12, 59394.90 examples/s]
Generating train split: 67703686 examples [15:12, 73105.23 examples/s]
Generating train split: 67738686 examples [15:12, 115923.51 examples/s]
Generating train split: 67754686 examples [15:12, 98418.02 examples/s]
Generating train split: 67769686 examples [15:13, 76516.38 examples/s]
Generating train split: 67785686 examples [15:13, 86828.01 examples/s]
Generating train split: 67799686 examples [15:13, 53825.40 examples/s]
Generating train split: 67815686 examples [15:13, 64916.20 examples/s]
Generating train split: 67825686 examples [15:14, 60592.65 examples/s]
Generating train split: 67846686 examples [15:14, 76954.23 examples/s]
Generating train split: 67856686 examples [15:14, 69508.32 examples/s]
Generating train split: 67867686 examples [15:14, 72250.24 examples/s]
Generating train split: 67884686 examples [15:14, 84240.81 examples/s]
Generating train split: 67905686 examples [15:14, 109192.56 examples/s]
Generating train split: 67933686 examples [15:14, 120888.76 examples/s]
Generating train split: 67948686 examples [15:15, 126446.77 examples/s]
Generating train split: 67964686 examples [15:15, 80514.19 examples/s]
Generating train split: 67994686 examples [15:15, 94314.84 examples/s]
Generating train split: 68014686 examples [15:15, 109110.60 examples/s]
Generating train split: 68030686 examples [15:15, 110695.47 examples/s]
Generating train split: 68047686 examples [15:16, 93651.32 examples/s]
Generating train split: 68059686 examples [15:16, 88587.85 examples/s]
Generating train split: 68070686 examples [15:16, 62795.50 examples/s]
Generating train split: 68085686 examples [15:16, 75551.86 examples/s]
Generating train split: 68101686 examples [15:16, 85591.68 examples/s]
Generating train split: 68112686 examples [15:17, 78764.32 examples/s]
Generating train split: 68123686 examples [15:17, 83776.09 examples/s]
Generating train split: 68136686 examples [15:17, 57857.43 examples/s]
Generating train split: 68148686 examples [15:18, 37416.54 examples/s]
Generating train split: 68165686 examples [15:18, 50229.84 examples/s]
Generating train split: 68177686 examples [15:18, 54029.97 examples/s]
Generating train split: 68189686 examples [15:18, 63502.67 examples/s]
Generating train split: 68201686 examples [15:18, 64236.20 examples/s]
Generating train split: 68212686 examples [15:18, 71113.65 examples/s]
Generating train split: 68223686 examples [15:19, 75870.23 examples/s]
Generating train split: 68235686 examples [15:19, 83504.39 examples/s]
Generating train split: 68257686 examples [15:19, 95722.78 examples/s]
Generating train split: 68270686 examples [15:19, 100938.68 examples/s]
Generating train split: 68283686 examples [15:19, 82411.22 examples/s]
Generating train split: 68294686 examples [15:19, 80800.50 examples/s]
Generating train split: 68304686 examples [15:19, 82269.01 examples/s]
Generating train split: 68315686 examples [15:20, 76060.50 examples/s]
Generating train split: 68331686 examples [15:20, 86695.43 examples/s]
Generating train split: 68341686 examples [15:20, 69589.65 examples/s]
Generating train split: 68353686 examples [15:20, 78852.06 examples/s]
Generating train split: 68364686 examples [15:20, 82445.60 examples/s]
Generating train split: 68376686 examples [15:20, 79216.17 examples/s]
Generating train split: 68387686 examples [15:21, 62741.43 examples/s]
Generating train split: 68396686 examples [15:21, 60959.58 examples/s]
Generating train split: 68406686 examples [15:21, 60609.39 examples/s]
Generating train split: 68414686 examples [15:21, 59615.90 examples/s]
Generating train split: 68426686 examples [15:22, 44188.75 examples/s]
Generating train split: 68444686 examples [15:22, 63981.51 examples/s]
Generating train split: 68456686 examples [15:22, 50967.16 examples/s]
Generating train split: 68464686 examples [15:22, 40745.74 examples/s]
Generating train split: 68485686 examples [15:22, 60890.33 examples/s]
Generating train split: 68507686 examples [15:23, 81556.70 examples/s]
Generating train split: 68529686 examples [15:23, 88542.56 examples/s]
Generating train split: 68541686 examples [15:23, 89758.45 examples/s]
Generating train split: 68554686 examples [15:23, 80461.81 examples/s]
Generating train split: 68571686 examples [15:23, 92793.00 examples/s]
Generating train split: 68582686 examples [15:23, 96201.38 examples/s]
Generating train split: 68605686 examples [15:24, 119370.29 examples/s]
Generating train split: 68618686 examples [15:24, 82429.35 examples/s]
Generating train split: 68634686 examples [15:24, 86108.56 examples/s]
Generating train split: 68646686 examples [15:24, 91637.54 examples/s]
Generating train split: 68658686 examples [15:24, 83762.53 examples/s]
Generating train split: 68670686 examples [15:25, 56657.27 examples/s]
Generating train split: 68697686 examples [15:25, 86199.05 examples/s]
Generating train split: 68709686 examples [15:25, 78304.39 examples/s]
Generating train split: 68719686 examples [15:25, 56498.90 examples/s]
Generating train split: 68736686 examples [15:26, 67794.19 examples/s]
Generating train split: 68747686 examples [15:26, 68061.31 examples/s]
Generating train split: 68758686 examples [15:26, 74889.80 examples/s]
Generating train split: 68769686 examples [15:26, 54703.26 examples/s]
Generating train split: 68784686 examples [15:26, 66947.92 examples/s]
Generating train split: 68801686 examples [15:26, 82758.20 examples/s]
Generating train split: 68813686 examples [15:27, 53174.05 examples/s]
Generating train split: 68830686 examples [15:27, 66822.55 examples/s]
Generating train split: 68848686 examples [15:27, 82941.29 examples/s]
Generating train split: 68878686 examples [15:27, 113523.25 examples/s]
Generating train split: 68896686 examples [15:27, 103986.06 examples/s]
Generating train split: 68914686 examples [15:28, 78044.96 examples/s]
Generating train split: 68926686 examples [15:28, 73174.85 examples/s]
Generating train split: 68954686 examples [15:28, 104055.81 examples/s]
Generating train split: 68970686 examples [15:28, 102478.14 examples/s]
Generating train split: 68989686 examples [15:28, 108531.05 examples/s]
Generating train split: 69011686 examples [15:29, 124365.27 examples/s]
Generating train split: 69028686 examples [15:29, 108397.46 examples/s]
Generating train split: 69042686 examples [15:29, 73744.96 examples/s]
Generating train split: 69055686 examples [15:30, 58045.05 examples/s]
Generating train split: 69067686 examples [15:30, 48418.52 examples/s]
Generating train split: 69076686 examples [15:30, 51106.39 examples/s]
Generating train split: 69083686 examples [15:30, 51476.01 examples/s]
Generating train split: 69090686 examples [15:31, 33604.00 examples/s]
Generating train split: 69102686 examples [15:31, 43263.66 examples/s]
Generating train split: 69115686 examples [15:31, 54710.01 examples/s]
Generating train split: 69130686 examples [15:31, 47442.79 examples/s]
Generating train split: 69154686 examples [15:31, 65399.56 examples/s]
Generating train split: 69177686 examples [15:32, 89388.73 examples/s]
Generating train split: 69215686 examples [15:32, 136628.48 examples/s]
Generating train split: 69238686 examples [15:32, 111244.05 examples/s]
Generating train split: 69255686 examples [15:32, 76495.46 examples/s]
Generating train split: 69273686 examples [15:33, 89481.52 examples/s]
Generating train split: 69291686 examples [15:33, 86378.33 examples/s]
Generating train split: 69306686 examples [15:33, 92405.01 examples/s]
Generating train split: 69319686 examples [15:33, 81528.24 examples/s]
Generating train split: 69329686 examples [15:33, 80718.20 examples/s]
Generating train split: 69340686 examples [15:34, 55774.83 examples/s]
Generating train split: 69361686 examples [15:34, 63835.17 examples/s]
Generating train split: 69392686 examples [15:34, 94129.17 examples/s]
Generating train split: 69404686 examples [15:34, 89329.73 examples/s]
Generating train split: 69419686 examples [15:35, 40172.35 examples/s]
Generating train split: 69431686 examples [15:35, 47506.82 examples/s]
Generating train split: 69464686 examples [15:35, 80117.34 examples/s]
Generating train split: 69485686 examples [15:36, 85918.03 examples/s]
Generating train split: 69502686 examples [15:36, 77170.03 examples/s]
Generating train split: 69519686 examples [15:36, 85019.69 examples/s]
Generating train split: 69533686 examples [15:36, 89721.22 examples/s]
Generating train split: 69556686 examples [15:36, 92869.60 examples/s]
Generating train split: 69574686 examples [15:36, 107512.06 examples/s]
Generating train split: 69591686 examples [15:37, 117912.62 examples/s]
Generating train split: 69608686 examples [15:37, 93329.19 examples/s]
Generating train split: 69625686 examples [15:37, 90791.70 examples/s]
Generating train split: 69638686 examples [15:37, 75912.91 examples/s]
Generating train split: 69655686 examples [15:38, 64386.57 examples/s]
Generating train split: 69665686 examples [15:38, 48507.50 examples/s]
Generating train split: 69672686 examples [15:38, 50640.59 examples/s]
Generating train split: 69684686 examples [15:38, 52503.93 examples/s]
Generating train split: 69709686 examples [15:38, 82569.43 examples/s]
Generating train split: 69721686 examples [15:39, 75722.49 examples/s]
Generating train split: 69738686 examples [15:39, 84088.78 examples/s]
Generating train split: 69753686 examples [15:39, 60521.08 examples/s]
Generating train split: 69766686 examples [15:40, 46946.29 examples/s]
Generating train split: 69792686 examples [15:40, 71049.63 examples/s]
Generating train split: 69810686 examples [15:40, 81394.76 examples/s]
Generating train split: 69836686 examples [15:40, 104432.28 examples/s]
Generating train split: 69857686 examples [15:41, 73857.65 examples/s]
Generating train split: 69878686 examples [15:41, 77737.37 examples/s]
Generating train split: 69895686 examples [15:41, 89264.54 examples/s]
Generating train split: 69907686 examples [15:41, 72133.95 examples/s]
Generating train split: 69937686 examples [15:41, 102423.83 examples/s]
Generating train split: 69954686 examples [15:41, 103748.47 examples/s]
Generating train split: 69971686 examples [15:42, 78492.26 examples/s]
Generating train split: 69982686 examples [15:42, 71559.75 examples/s]
Generating train split: 69993686 examples [15:42, 72831.25 examples/s]
Generating train split: 70007686 examples [15:42, 81882.29 examples/s]
Generating train split: 70017686 examples [15:43, 67581.63 examples/s]
Generating train split: 70027686 examples [15:43, 62646.60 examples/s]
Generating train split: 70059686 examples [15:43, 105118.56 examples/s]
Generating train split: 70072686 examples [15:44, 48148.36 examples/s]
Generating train split: 70084686 examples [15:44, 55824.42 examples/s]
Generating train split: 70105686 examples [15:44, 75365.93 examples/s]
Generating train split: 70120686 examples [15:44, 55427.82 examples/s]
Generating train split: 70138686 examples [15:44, 67683.32 examples/s]
Generating train split: 70153686 examples [15:45, 79398.43 examples/s]
Generating train split: 70185686 examples [15:45, 120422.57 examples/s]
Generating train split: 70206686 examples [15:45, 99653.18 examples/s]
Generating train split: 70221686 examples [15:45, 92084.76 examples/s]
Generating train split: 70236686 examples [15:45, 86473.44 examples/s]
Generating train split: 70262686 examples [15:45, 111163.00 examples/s]
Generating train split: 70279686 examples [15:46, 98035.86 examples/s]
Generating train split: 70303686 examples [15:46, 122651.43 examples/s]
Generating train split: 70319686 examples [15:46, 90645.23 examples/s]
Generating train split: 70334686 examples [15:47, 63819.74 examples/s]
Generating train split: 70346686 examples [15:47, 63181.25 examples/s]
Generating train split: 70361686 examples [15:47, 74324.21 examples/s]
Generating train split: 70372686 examples [15:47, 77185.43 examples/s]
Generating train split: 70387686 examples [15:47, 90606.81 examples/s]
Generating train split: 70403686 examples [15:47, 89882.95 examples/s]
Generating train split: 70415686 examples [15:48, 39361.54 examples/s]
Generating train split: 70448686 examples [15:48, 70369.17 examples/s]
Generating train split: 70466686 examples [15:49, 63331.98 examples/s]
Generating train split: 70484686 examples [15:49, 62661.43 examples/s]
Generating train split: 70496686 examples [15:49, 66955.98 examples/s]
Generating train split: 70523686 examples [15:49, 95524.35 examples/s]
Generating train split: 70543686 examples [15:49, 91858.09 examples/s]
Generating train split: 70559686 examples [15:50, 80649.70 examples/s]
Generating train split: 70570686 examples [15:50, 79633.92 examples/s]
Generating train split: 70582686 examples [15:50, 84715.69 examples/s]
Generating train split: 70599686 examples [15:50, 86261.28 examples/s]
Generating train split: 70610686 examples [15:50, 82873.58 examples/s]
Generating train split: 70635686 examples [15:50, 87089.32 examples/s]
Generating train split: 70648686 examples [15:51, 77098.03 examples/s]
Generating train split: 70660686 examples [15:51, 64874.22 examples/s]
Generating train split: 70672686 examples [15:51, 56981.31 examples/s]
Generating train split: 70683686 examples [15:51, 61307.77 examples/s]
Generating train split: 70690686 examples [15:52, 51883.49 examples/s]
Generating train split: 70697686 examples [15:52, 48670.13 examples/s]
Generating train split: 70707686 examples [15:52, 56365.54 examples/s]
Generating train split: 70719686 examples [15:52, 54882.62 examples/s]
Generating train split: 70729686 examples [15:53, 36585.05 examples/s]
Generating train split: 70757686 examples [15:53, 67497.33 examples/s]
Generating train split: 70772686 examples [15:53, 71431.04 examples/s]
Generating train split: 70788686 examples [15:53, 83538.59 examples/s]
Generating train split: 70803686 examples [15:53, 73090.91 examples/s]
Generating train split: 70815686 examples [15:53, 69722.58 examples/s]
Generating train split: 70835686 examples [15:54, 89262.30 examples/s]
Generating train split: 70857686 examples [15:54, 91676.09 examples/s]
Generating train split: 70869686 examples [15:54, 83633.21 examples/s]
Generating train split: 70880686 examples [15:54, 73036.25 examples/s]
Generating train split: 70903686 examples [15:54, 89918.24 examples/s]
Generating train split: 70917686 examples [15:55, 69892.56 examples/s]
Generating train split: 70939686 examples [15:55, 74303.21 examples/s]
Generating train split: 70957686 examples [15:55, 87753.18 examples/s]
Generating train split: 70968686 examples [15:55, 90202.78 examples/s]
Generating train split: 70984686 examples [15:55, 85585.26 examples/s]
Generating train split: 71001686 examples [15:56, 87111.38 examples/s]
Generating train split: 71011686 examples [15:56, 88654.95 examples/s]
Generating train split: 71022686 examples [15:56, 53041.06 examples/s]
Generating train split: 71033686 examples [15:56, 59927.97 examples/s]
Generating train split: 71045686 examples [15:57, 48588.19 examples/s]
Generating train split: 71057686 examples [15:57, 43709.87 examples/s]
Generating train split: 71068686 examples [15:57, 51034.81 examples/s]
Generating train split: 71089686 examples [15:57, 73547.45 examples/s]
Generating train split: 71103686 examples [15:57, 77701.05 examples/s]
Generating train split: 71118686 examples [15:58, 89632.38 examples/s]
Generating train split: 71129686 examples [15:58, 70797.54 examples/s]
Generating train split: 71142686 examples [15:58, 70275.80 examples/s]
Generating train split: 71161686 examples [15:58, 91711.52 examples/s]
Generating train split: 71178686 examples [15:58, 80633.18 examples/s]
Generating train split: 71189686 examples [15:58, 78208.83 examples/s]
Generating train split: 71199686 examples [15:59, 81151.20 examples/s]
Generating train split: 71215686 examples [15:59, 95463.18 examples/s]
Generating train split: 71248686 examples [15:59, 127996.62 examples/s]
Generating train split: 71264686 examples [15:59, 63405.82 examples/s]
Generating train split: 71276686 examples [16:00, 65739.44 examples/s]
Generating train split: 71292686 examples [16:00, 67941.15 examples/s]
Generating train split: 71307686 examples [16:00, 79806.11 examples/s]
Generating train split: 71322686 examples [16:00, 87936.32 examples/s]
Generating train split: 71338686 examples [16:00, 93580.32 examples/s]
Generating train split: 71353686 examples [16:01, 75101.10 examples/s]
Generating train split: 71365686 examples [16:01, 51306.73 examples/s]
Generating train split: 71381686 examples [16:01, 61680.30 examples/s]
Generating train split: 71392686 examples [16:01, 56890.17 examples/s]
Generating train split: 71413686 examples [16:02, 70981.66 examples/s]
Generating train split: 71425686 examples [16:02, 72366.78 examples/s]
Generating train split: 71435686 examples [16:02, 62842.01 examples/s]
Generating train split: 71446686 examples [16:02, 54549.07 examples/s]
Generating train split: 71462686 examples [16:02, 70965.80 examples/s]
Generating train split: 71472686 examples [16:02, 71456.27 examples/s]
Generating train split: 71497686 examples [16:03, 102949.30 examples/s]
Generating train split: 71512686 examples [16:03, 96530.60 examples/s]
Generating train split: 71533686 examples [16:03, 117475.67 examples/s]
Generating train split: 71548686 examples [16:03, 86890.44 examples/s]
Generating train split: 71570686 examples [16:03, 101367.84 examples/s]
Generating train split: 71588686 examples [16:03, 106558.44 examples/s]
Generating train split: 71605686 examples [16:04, 74272.40 examples/s]
Generating train split: 71616686 examples [16:04, 62944.05 examples/s]
Generating train split: 71628686 examples [16:04, 70169.70 examples/s]
Generating train split: 71642686 examples [16:05, 54712.26 examples/s]
Generating train split: 71650686 examples [16:05, 49849.00 examples/s]
Generating train split: 71660686 examples [16:05, 56360.28 examples/s]
Generating train split: 71670686 examples [16:05, 63669.43 examples/s]
Generating train split: 71687686 examples [16:06, 48393.47 examples/s]
Generating train split: 71695686 examples [16:06, 51733.59 examples/s]
Generating train split: 71709686 examples [16:06, 65923.73 examples/s]
Generating train split: 71731686 examples [16:06, 91813.10 examples/s]
Generating train split: 71745686 examples [16:06, 85582.81 examples/s]
Generating train split: 71762686 examples [16:06, 101525.94 examples/s]
Generating train split: 71777686 examples [16:06, 84931.19 examples/s]
Generating train split: 71788686 examples [16:07, 68340.82 examples/s]
Generating train split: 71800686 examples [16:07, 76617.32 examples/s]
Generating train split: 71810686 examples [16:07, 78746.05 examples/s]
Generating train split: 71827686 examples [16:07, 94600.22 examples/s]
Generating train split: 71849686 examples [16:07, 111362.66 examples/s]
Generating train split: 71869686 examples [16:07, 120405.61 examples/s]
Generating train split: 71885686 examples [16:08, 56542.72 examples/s]
Generating train split: 71910686 examples [16:08, 80117.32 examples/s]
Generating train split: 71929686 examples [16:08, 90907.15 examples/s]
Generating train split: 71945686 examples [16:08, 93272.55 examples/s]
Generating train split: 71961686 examples [16:09, 51976.23 examples/s]
Generating train split: 71977686 examples [16:09, 62593.42 examples/s]
Generating train split: 71989686 examples [16:10, 50698.33 examples/s]
Generating train split: 72002686 examples [16:10, 41499.31 examples/s]
Generating train split: 72018686 examples [16:10, 53452.55 examples/s]
Generating train split: 72039686 examples [16:10, 73947.52 examples/s]
Generating train split: 72059686 examples [16:10, 88503.74 examples/s]
Generating train split: 72074686 examples [16:11, 75002.17 examples/s]
Generating train split: 72096686 examples [16:11, 85970.38 examples/s]
Generating train split: 72119686 examples [16:11, 100619.09 examples/s]
Generating train split: 72136686 examples [16:11, 71365.70 examples/s]
Generating train split: 72146686 examples [16:12, 66646.11 examples/s]
Generating train split: 72174686 examples [16:12, 97803.31 examples/s]
Generating train split: 72189686 examples [16:12, 81970.02 examples/s]
Generating train split: 72201686 examples [16:12, 57801.29 examples/s]
Generating train split: 72210686 examples [16:13, 61191.55 examples/s]
Generating train split: 72237686 examples [16:13, 90698.23 examples/s]
Generating train split: 72254686 examples [16:13, 83349.66 examples/s]
Generating train split: 72266686 examples [16:13, 51424.74 examples/s]
Generating train split: 72284686 examples [16:14, 66480.91 examples/s]
Generating train split: 72297686 examples [16:14, 62585.71 examples/s]
Generating train split: 72308686 examples [16:14, 68591.46 examples/s]
Generating train split: 72319686 examples [16:14, 75221.01 examples/s]
Generating train split: 72334686 examples [16:14, 73458.53 examples/s]
Generating train split: 72348686 examples [16:14, 68728.07 examples/s]
Generating train split: 72358686 examples [16:15, 73942.23 examples/s]
Generating train split: 72370686 examples [16:15, 42537.61 examples/s]
Generating train split: 72391686 examples [16:15, 63819.25 examples/s]
Generating train split: 72406686 examples [16:15, 76479.73 examples/s]
Generating train split: 72426686 examples [16:16, 95005.00 examples/s]
Generating train split: 72441686 examples [16:16, 59254.59 examples/s]
Generating train split: 72452686 examples [16:16, 63344.78 examples/s]
Generating train split: 72475686 examples [16:16, 87263.78 examples/s]
Generating train split: 72490686 examples [16:16, 96030.24 examples/s]
Generating train split: 72507686 examples [16:17, 101163.39 examples/s]
Generating train split: 72523686 examples [16:17, 55127.81 examples/s]
Generating train split: 72535686 examples [16:17, 62013.33 examples/s]
Generating train split: 72545686 examples [16:17, 63989.62 examples/s]
Generating train split: 72574686 examples [16:18, 54590.19 examples/s]
Generating train split: 72583686 examples [16:18, 50638.16 examples/s]
Generating train split: 72593686 examples [16:18, 55018.87 examples/s]
Generating train split: 72603686 examples [16:18, 61250.60 examples/s]
Generating train split: 72631686 examples [16:19, 96591.31 examples/s]
Generating train split: 72663686 examples [16:19, 134691.56 examples/s]
Generating train split: 72683686 examples [16:19, 98168.72 examples/s]
Generating train split: 72699686 examples [16:19, 108127.08 examples/s]
Generating train split: 72717686 examples [16:20, 57127.35 examples/s]
Generating train split: 72742686 examples [16:20, 78931.10 examples/s]
Generating train split: 72762686 examples [16:20, 94056.08 examples/s]
Generating train split: 72781686 examples [16:20, 82156.25 examples/s]
Generating train split: 72796686 examples [16:20, 87466.40 examples/s]
Generating train split: 72813686 examples [16:21, 95073.79 examples/s]
Generating train split: 72829686 examples [16:21, 100808.79 examples/s]
Generating train split: 72845686 examples [16:21, 104466.79 examples/s]
Generating train split: 72860686 examples [16:21, 98158.88 examples/s]
Generating train split: 72881686 examples [16:21, 110997.31 examples/s]
Generating train split: 72897686 examples [16:22, 56502.50 examples/s]
Generating train split: 72927686 examples [16:22, 84917.37 examples/s]
Generating train split: 72945686 examples [16:22, 63922.17 examples/s]
Generating train split: 72962686 examples [16:23, 74301.73 examples/s]
Generating train split: 72978686 examples [16:23, 76813.68 examples/s]
Generating train split: 72998686 examples [16:23, 95231.53 examples/s]
Generating train split: 73014686 examples [16:23, 56282.76 examples/s]
Generating train split: 73025686 examples [16:24, 62187.87 examples/s]
Generating train split: 73038686 examples [16:24, 61685.90 examples/s]
Generating train split: 73061686 examples [16:24, 83418.10 examples/s]
Generating train split: 73073686 examples [16:24, 76333.18 examples/s]
Generating train split: 73086686 examples [16:24, 67799.96 examples/s]
Generating train split: 73097686 examples [16:24, 73162.22 examples/s]
Generating train split: 73108686 examples [16:25, 78990.03 examples/s]
Generating train split: 73124686 examples [16:25, 95342.33 examples/s]
Generating train split: 73136686 examples [16:25, 99853.81 examples/s]
Generating train split: 73152686 examples [16:25, 100416.10 examples/s]
Generating train split: 73163686 examples [16:25, 72549.21 examples/s]
Generating train split: 73174686 examples [16:26, 56491.99 examples/s]
Generating train split: 73187686 examples [16:26, 64760.61 examples/s]
Generating train split: 73199686 examples [16:26, 39775.29 examples/s]
Generating train split: 73228686 examples [16:26, 67068.74 examples/s]
Generating train split: 73239686 examples [16:27, 54314.27 examples/s]
Generating train split: 73250686 examples [16:27, 61540.67 examples/s]
Generating train split: 73270686 examples [16:27, 82668.67 examples/s]
Generating train split: 73307686 examples [16:27, 112492.55 examples/s]
Generating train split: 73325686 examples [16:27, 115104.91 examples/s]
Generating train split: 73340686 examples [16:28, 86900.46 examples/s]
Generating train split: 73351686 examples [16:28, 77354.72 examples/s]
Generating train split: 73363686 examples [16:28, 54412.67 examples/s]
Generating train split: 73378686 examples [16:28, 63508.24 examples/s]
Generating train split: 73391686 examples [16:29, 70537.32 examples/s]
Generating train split: 73402686 examples [16:29, 75254.76 examples/s]
Generating train split: 73423686 examples [16:29, 98322.16 examples/s]
Generating train split: 73436686 examples [16:29, 102569.77 examples/s]
Generating train split: 73451686 examples [16:29, 76847.68 examples/s]
Generating train split: 73465686 examples [16:29, 88023.84 examples/s]
Generating train split: 73476686 examples [16:30, 70195.12 examples/s]
Generating train split: 73487686 examples [16:30, 61927.20 examples/s]
Generating train split: 73498686 examples [16:30, 57909.42 examples/s]
Generating train split: 73509686 examples [16:30, 56181.91 examples/s]
Generating train split: 73519686 examples [16:31, 39027.46 examples/s]
Generating train split: 73541686 examples [16:31, 61087.20 examples/s]
Generating train split: 73567686 examples [16:31, 70602.20 examples/s]
Generating train split: 73580686 examples [16:31, 74612.27 examples/s]
Generating train split: 73598686 examples [16:31, 89960.23 examples/s]
Generating train split: 73614686 examples [16:32, 73916.25 examples/s]
Generating train split: 73624686 examples [16:32, 64445.87 examples/s]
Generating train split: 73641686 examples [16:32, 75200.16 examples/s]
Generating train split: 73653686 examples [16:32, 74179.50 examples/s]
Generating train split: 73664686 examples [16:33, 44000.34 examples/s]
Generating train split: 73676686 examples [16:33, 51755.73 examples/s]
Generating train split: 73687686 examples [16:33, 58143.89 examples/s]
Generating train split: 73714686 examples [16:33, 92849.31 examples/s]
Generating train split: 73735686 examples [16:33, 113177.50 examples/s]
Generating train split: 73752686 examples [16:33, 118200.01 examples/s]
Generating train split: 73769686 examples [16:34, 65787.67 examples/s]
Generating train split: 73786686 examples [16:34, 66118.23 examples/s]
Generating train split: 73796686 examples [16:34, 55875.46 examples/s]
Generating train split: 73812686 examples [16:35, 69791.65 examples/s]
Generating train split: 73833686 examples [16:35, 91109.77 examples/s]
Generating train split: 73850686 examples [16:35, 59380.50 examples/s]
Generating train split: 73868686 examples [16:35, 74818.02 examples/s]
Generating train split: 73888686 examples [16:35, 88084.64 examples/s]
Generating train split: 73903686 examples [16:36, 86787.11 examples/s]
Generating train split: 73919686 examples [16:36, 98347.12 examples/s]
Generating train split: 73937686 examples [16:36, 114546.56 examples/s]
Generating train split: 73954686 examples [16:36, 107681.96 examples/s]
Generating train split: 73975686 examples [16:37, 66126.79 examples/s]
Generating train split: 73990686 examples [16:37, 63437.05 examples/s]
Generating train split: 74001686 examples [16:37, 68971.01 examples/s]
Generating train split: 74012686 examples [16:37, 58087.32 examples/s]
Generating train split: 74023686 examples [16:37, 64512.60 examples/s]
Generating train split: 74034686 examples [16:37, 65740.94 examples/s]
Generating train split: 74053686 examples [16:38, 86243.80 examples/s]
Generating train split: 74086686 examples [16:38, 125060.22 examples/s]
Generating train split: 74101686 examples [16:38, 95729.16 examples/s]
Generating train split: 74113686 examples [16:38, 81508.25 examples/s]
Generating train split: 74124686 examples [16:38, 84278.66 examples/s]
Generating train split: 74135686 examples [16:39, 46582.49 examples/s]
Generating train split: 74151686 examples [16:39, 60103.33 examples/s]
Generating train split: 74168686 examples [16:39, 52301.83 examples/s]
Generating train split: 74181686 examples [16:40, 59668.26 examples/s]
Generating train split: 74215686 examples [16:40, 99667.10 examples/s]
Generating train split: 74232686 examples [16:40, 94290.19 examples/s]
Generating train split: 74250686 examples [16:40, 105959.45 examples/s]
Generating train split: 74264686 examples [16:40, 78862.68 examples/s]
Generating train split: 74275686 examples [16:40, 81215.70 examples/s]
Generating train split: 74287686 examples [16:41, 43516.87 examples/s]
Generating train split: 74298686 examples [16:41, 47877.07 examples/s]
Generating train split: 74316686 examples [16:41, 63118.42 examples/s]
Generating train split: 74333686 examples [16:42, 77908.85 examples/s]
Generating train split: 74346686 examples [16:42, 55399.90 examples/s]
Generating train split: 74356686 examples [16:42, 60950.69 examples/s]
Generating train split: 74394686 examples [16:42, 109035.06 examples/s]
Generating train split: 74415686 examples [16:42, 122760.71 examples/s]
Generating train split: 74431686 examples [16:43, 86239.92 examples/s]
Generating train split: 74446686 examples [16:43, 90697.13 examples/s]
Generating train split: 74469686 examples [16:43, 112794.67 examples/s]
Generating train split: 74484686 examples [16:43, 65131.89 examples/s]
Generating train split: 74506686 examples [16:44, 82506.45 examples/s]
Generating train split: 74522686 examples [16:44, 63897.39 examples/s]
Generating train split: 74554686 examples [16:44, 96924.21 examples/s]
Generating train split: 74585686 examples [16:44, 106284.44 examples/s]
Generating train split: 74601686 examples [16:44, 114031.66 examples/s]
Generating train split: 74617686 examples [16:45, 99401.09 examples/s]
Generating train split: 74639686 examples [16:45, 100228.96 examples/s]
Generating train split: 74652686 examples [16:45, 104161.02 examples/s]
Generating train split: 74668686 examples [16:46, 49558.30 examples/s]
Generating train split: 74681686 examples [16:46, 54485.99 examples/s]
Generating train split: 74713686 examples [16:46, 86935.28 examples/s]
Generating train split: 74735686 examples [16:46, 104276.46 examples/s]
Generating train split: 74752686 examples [16:47, 76939.65 examples/s]
Generating train split: 74765686 examples [16:47, 71017.46 examples/s]
Generating train split: 74779686 examples [16:47, 73762.69 examples/s]
Generating train split: 74792686 examples [16:47, 67303.16 examples/s]
Generating train split: 74803686 examples [16:47, 72382.33 examples/s]
Generating train split: 74828686 examples [16:47, 98896.03 examples/s]
Generating train split: 74842686 examples [16:48, 51772.35 examples/s]
Generating train split: 74852686 examples [16:48, 45802.26 examples/s]
Generating train split: 74875686 examples [16:48, 66973.64 examples/s]
Generating train split: 74909686 examples [16:49, 74196.34 examples/s]
Generating train split: 74921686 examples [16:49, 69989.20 examples/s]
Generating train split: 74933686 examples [16:49, 66140.63 examples/s]
Generating train split: 74949686 examples [16:49, 76434.77 examples/s]
Generating train split: 74959686 examples [16:50, 38573.17 examples/s]
Generating train split: 74980686 examples [16:50, 55812.66 examples/s]
Generating train split: 74994686 examples [16:50, 64398.77 examples/s]
Generating train split: 75020686 examples [16:51, 93051.40 examples/s]
Generating train split: 75053686 examples [16:51, 133842.87 examples/s]
Generating train split: 75078686 examples [16:51, 73397.54 examples/s]
Generating train split: 75095686 examples [16:51, 81314.02 examples/s]
Generating train split: 75110686 examples [16:52, 82822.66 examples/s]
Generating train split: 75126686 examples [16:52, 88822.93 examples/s]
Generating train split: 75142686 examples [16:52, 99318.93 examples/s]
Generating train split: 75159686 examples [16:52, 74115.57 examples/s]
Generating train split: 75177686 examples [16:52, 89078.88 examples/s]
Generating train split: 75198686 examples [16:52, 103620.02 examples/s]
Generating train split: 75212686 examples [16:53, 93791.37 examples/s]
Generating train split: 75228686 examples [16:53, 76658.15 examples/s]
Generating train split: 75249686 examples [16:53, 96602.29 examples/s]
Generating train split: 75265686 examples [16:53, 89448.66 examples/s]
Generating train split: 75280686 examples [16:53, 99025.96 examples/s]
Generating train split: 75293686 examples [16:54, 43601.70 examples/s]
Generating train split: 75316686 examples [16:54, 58774.14 examples/s]
Generating train split: 75331686 examples [16:55, 56290.61 examples/s]
Generating train split: 75349686 examples [16:55, 60106.49 examples/s]
Generating train split: 75365686 examples [16:55, 71211.45 examples/s]
Generating train split: 75382686 examples [16:55, 84009.73 examples/s]
Generating train split: 75394686 examples [16:56, 65257.14 examples/s]
Generating train split: 75409686 examples [16:56, 77130.76 examples/s]
Generating train split: 75424686 examples [16:56, 82718.74 examples/s]
Generating train split: 75448686 examples [16:56, 71124.34 examples/s]
Generating train split: 75466686 examples [16:56, 81680.47 examples/s]
Generating train split: 75478686 examples [16:56, 82778.66 examples/s]
Generating train split: 75488686 examples [16:57, 69468.81 examples/s]
Generating train split: 75500686 examples [16:57, 74918.89 examples/s]
Generating train split: 75518686 examples [16:57, 88915.24 examples/s]
Generating train split: 75530686 examples [16:57, 80944.27 examples/s]
Generating train split: 75544686 examples [16:57, 75961.50 examples/s]
Generating train split: 75560686 examples [16:57, 87124.95 examples/s]
Generating train split: 75582686 examples [16:58, 91694.83 examples/s]
Generating train split: 75598686 examples [16:58, 104127.34 examples/s]
Generating train split: 75614686 examples [16:59, 46422.45 examples/s]
Generating train split: 75638686 examples [16:59, 65778.27 examples/s]
Generating train split: 75654686 examples [16:59, 77525.72 examples/s]
Generating train split: 75671686 examples [16:59, 64201.85 examples/s]
Generating train split: 75682686 examples [16:59, 63498.42 examples/s]
Generating train split: 75694686 examples [17:00, 64525.33 examples/s]
Generating train split: 75721686 examples [17:00, 95007.80 examples/s]
Generating train split: 75735686 examples [17:00, 87863.95 examples/s]
Generating train split: 75752686 examples [17:00, 97429.32 examples/s]
Generating train split: 75769686 examples [17:00, 111117.73 examples/s]
Generating train split: 75790686 examples [17:00, 122933.04 examples/s]
Generating train split: 75806686 examples [17:01, 83717.82 examples/s]
Generating train split: 75823686 examples [17:01, 72017.33 examples/s]
Generating train split: 75834686 examples [17:01, 77428.92 examples/s]
Generating train split: 75847686 examples [17:01, 69365.53 examples/s]
Generating train split: 75863686 examples [17:01, 81115.93 examples/s]
Generating train split: 75882686 examples [17:02, 75498.17 examples/s]
Generating train split: 75893686 examples [17:02, 79136.57 examples/s]
Generating train split: 75904686 examples [17:02, 74623.05 examples/s]
Generating train split: 75917686 examples [17:02, 59675.48 examples/s]
Generating train split: 75928686 examples [17:03, 47520.63 examples/s]
Generating train split: 75938686 examples [17:03, 51749.13 examples/s]
Generating train split: 75949686 examples [17:03, 53318.61 examples/s]
Generating train split: 75976686 examples [17:03, 82944.23 examples/s]
Generating train split: 76005686 examples [17:03, 82782.73 examples/s]
Generating train split: 76022686 examples [17:04, 93499.62 examples/s]
Generating train split: 76038686 examples [17:04, 67310.58 examples/s]
Generating train split: 76057686 examples [17:04, 76931.03 examples/s]
Generating train split: 76069686 examples [17:04, 67077.94 examples/s]
Generating train split: 76080686 examples [17:05, 68060.17 examples/s]
Generating train split: 76096686 examples [17:05, 81845.29 examples/s]
Generating train split: 76106686 examples [17:05, 82110.71 examples/s]
Generating train split: 76119686 examples [17:05, 72088.58 examples/s]
Generating train split: 76130686 examples [17:05, 78399.21 examples/s]
Generating train split: 76144686 examples [17:05, 89772.16 examples/s]
Generating train split: 76159686 examples [17:05, 90214.48 examples/s]
Generating train split: 76171686 examples [17:06, 83989.90 examples/s]
Generating train split: 76180686 examples [17:06, 59473.89 examples/s]
Generating train split: 76196686 examples [17:06, 59437.19 examples/s]
Generating train split: 76207686 examples [17:06, 66043.98 examples/s]
Generating train split: 76218686 examples [17:06, 71247.97 examples/s]
Generating train split: 76228686 examples [17:07, 74541.81 examples/s]
Generating train split: 76242686 examples [17:07, 43638.49 examples/s]
Generating train split: 76253686 examples [17:07, 43988.87 examples/s]
Generating train split: 76265686 examples [17:07, 51854.36 examples/s]
Generating train split: 76287686 examples [17:08, 76283.20 examples/s]
Generating train split: 76314686 examples [17:08, 109121.56 examples/s]
Generating train split: 76331686 examples [17:08, 85699.84 examples/s]
Generating train split: 76358686 examples [17:08, 83488.55 examples/s]
Generating train split: 76378686 examples [17:08, 96953.30 examples/s]
Generating train split: 76395686 examples [17:09, 103376.85 examples/s]
Generating train split: 76412686 examples [17:09, 75113.26 examples/s]
Generating train split: 76430686 examples [17:09, 74111.60 examples/s]
Generating train split: 76442686 examples [17:10, 62137.46 examples/s]
Generating train split: 76459686 examples [17:10, 75100.06 examples/s]
Generating train split: 76479686 examples [17:10, 93315.71 examples/s]
Generating train split: 76504686 examples [17:10, 111589.97 examples/s]
Generating train split: 76521686 examples [17:10, 120487.92 examples/s]
Generating train split: 76539686 examples [17:11, 76101.20 examples/s]
Generating train split: 76553686 examples [17:11, 68954.51 examples/s]
Generating train split: 76564686 examples [17:11, 43454.07 examples/s]
Generating train split: 76581686 examples [17:12, 54597.43 examples/s]
Generating train split: 76597686 examples [17:12, 66311.57 examples/s]
Generating train split: 76608686 examples [17:12, 51323.08 examples/s]
Generating train split: 76625686 examples [17:12, 65604.94 examples/s]
Generating train split: 76636686 examples [17:12, 61704.38 examples/s]
Generating train split: 76646686 examples [17:12, 67437.41 examples/s]
Generating train split: 76679686 examples [17:13, 115765.22 examples/s]
Generating train split: 76700686 examples [17:13, 104865.66 examples/s]
Generating train split: 76715686 examples [17:13, 112819.14 examples/s]
Generating train split: 76730686 examples [17:13, 108829.22 examples/s]
Generating train split: 76745686 examples [17:13, 106242.73 examples/s]
Generating train split: 76760686 examples [17:13, 91794.69 examples/s]
Generating train split: 76775686 examples [17:14, 78210.43 examples/s]
Generating train split: 76786686 examples [17:14, 74543.04 examples/s]
Generating train split: 76796686 examples [17:14, 69635.46 examples/s]
Generating train split: 76815686 examples [17:14, 91344.30 examples/s]
Generating train split: 76837686 examples [17:14, 114788.56 examples/s]
Generating train split: 76854686 examples [17:15, 89642.66 examples/s]
Generating train split: 76870686 examples [17:15, 76084.22 examples/s]
Generating train split: 76883686 examples [17:15, 61097.03 examples/s]
Generating train split: 76894686 examples [17:16, 44452.35 examples/s]
Generating train split: 76904686 examples [17:16, 48189.67 examples/s]
Generating train split: 76921686 examples [17:16, 64733.61 examples/s]
Generating train split: 76942686 examples [17:16, 80402.88 examples/s]
Generating train split: 76955686 examples [17:16, 69944.23 examples/s]
Generating train split: 76966686 examples [17:16, 74547.37 examples/s]
Generating train split: 76977686 examples [17:17, 71880.59 examples/s]
Generating train split: 76988686 examples [17:17, 65969.62 examples/s]
Generating train split: 77004686 examples [17:17, 80769.98 examples/s]
Generating train split: 77029686 examples [17:17, 103704.34 examples/s]
Generating train split: 77042686 examples [17:17, 102935.09 examples/s]
Generating train split: 77058686 examples [17:17, 102510.10 examples/s]
Generating train split: 77070686 examples [17:18, 75816.68 examples/s]
Generating train split: 77082686 examples [17:18, 65778.82 examples/s]
Generating train split: 77100686 examples [17:18, 79329.24 examples/s]
Generating train split: 77110686 examples [17:18, 77188.70 examples/s]
Generating train split: 77121686 examples [17:18, 79554.45 examples/s]
Generating train split: 77132686 examples [17:18, 77477.43 examples/s]
Generating train split: 77142686 examples [17:19, 74873.70 examples/s]
Generating train split: 77163686 examples [17:19, 102279.59 examples/s]
Generating train split: 77177686 examples [17:19, 105234.91 examples/s]
Generating train split: 77189686 examples [17:19, 85990.16 examples/s]
Generating train split: 77200686 examples [17:19, 82321.64 examples/s]
Generating train split: 77210686 examples [17:20, 47508.00 examples/s]
Generating train split: 77217686 examples [17:20, 49822.92 examples/s]
Generating train split: 77228686 examples [17:20, 54877.90 examples/s]
Generating train split: 77239686 examples [17:20, 63508.40 examples/s]
Generating train split: 77250686 examples [17:20, 59762.14 examples/s]
Generating train split: 77265686 examples [17:20, 74702.42 examples/s]
Generating train split: 77276686 examples [17:21, 68090.24 examples/s]
Generating train split: 77297686 examples [17:21, 92226.14 examples/s]
Generating train split: 77312686 examples [17:21, 82857.25 examples/s]
Generating train split: 77322686 examples [17:21, 76318.36 examples/s]
Generating train split: 77344686 examples [17:21, 95596.07 examples/s]
Generating train split: 77366686 examples [17:21, 114834.39 examples/s]
Generating train split: 77379686 examples [17:22, 79323.30 examples/s]
Generating train split: 77390686 examples [17:22, 62731.34 examples/s]
Generating train split: 77401686 examples [17:22, 69669.50 examples/s]
Generating train split: 77413686 examples [17:22, 73414.94 examples/s]
Generating train split: 77428686 examples [17:22, 83076.20 examples/s]
Generating train split: 77439686 examples [17:22, 87632.57 examples/s]
Generating train split: 77450686 examples [17:23, 81640.19 examples/s]
Generating train split: 77460686 examples [17:23, 58447.38 examples/s]
Generating train split: 77482686 examples [17:23, 83611.48 examples/s]
Generating train split: 77499686 examples [17:23, 60952.56 examples/s]
Generating train split: 77509686 examples [17:24, 36844.55 examples/s]
Generating train split: 77523686 examples [17:24, 47027.45 examples/s]
Generating train split: 77549686 examples [17:24, 70792.65 examples/s]
Generating train split: 77566686 examples [17:25, 70877.23 examples/s]
Generating train split: 77582686 examples [17:25, 83282.47 examples/s]
Generating train split: 77597686 examples [17:25, 69433.76 examples/s]
Generating train split: 77614686 examples [17:25, 83504.51 examples/s]
Generating train split: 77643686 examples [17:25, 119412.95 examples/s]
Generating train split: 77664686 examples [17:25, 123636.50 examples/s]
Generating train split: 77681686 examples [17:26, 92840.67 examples/s]
Generating train split: 77699686 examples [17:26, 64110.73 examples/s]
Generating train split: 77715686 examples [17:26, 74415.43 examples/s]
Generating train split: 77731686 examples [17:26, 86557.54 examples/s]
Generating train split: 77748686 examples [17:27, 95444.50 examples/s]
Generating train split: 77765686 examples [17:27, 64277.60 examples/s]
Generating train split: 77776686 examples [17:27, 61565.56 examples/s]
Generating train split: 77790686 examples [17:27, 73022.06 examples/s]
Generating train split: 77812686 examples [17:28, 87666.99 examples/s]
Generating train split: 77824686 examples [17:28, 60670.33 examples/s]
Generating train split: 77837686 examples [17:28, 64954.22 examples/s]
Generating train split: 77852686 examples [17:28, 76678.17 examples/s]
Generating train split: 77863686 examples [17:29, 57031.88 examples/s]
Generating train split: 77882686 examples [17:29, 75468.14 examples/s]
Generating train split: 77893686 examples [17:29, 73292.94 examples/s]
Generating train split: 77903686 examples [17:29, 60512.52 examples/s]
Generating train split: 77913686 examples [17:29, 50497.93 examples/s]
Generating train split: 77934686 examples [17:30, 69200.74 examples/s]
Generating train split: 77951686 examples [17:30, 84273.08 examples/s]
Generating train split: 77962686 examples [17:30, 85894.76 examples/s]
Generating train split: 77988686 examples [17:30, 105711.89 examples/s]
Generating train split: 78005686 examples [17:30, 104346.00 examples/s]
Generating train split: 78020686 examples [17:31, 53597.10 examples/s]
Generating train split: 78052686 examples [17:31, 81261.96 examples/s]
Generating train split: 78074686 examples [17:31, 100632.25 examples/s]
Generating train split: 78094686 examples [17:31, 78892.16 examples/s]
Generating train split: 78111686 examples [17:32, 69542.24 examples/s]
Generating train split: 78122686 examples [17:32, 68389.15 examples/s]
Generating train split: 78138686 examples [17:32, 81491.63 examples/s]
Generating train split: 78155686 examples [17:32, 65174.14 examples/s]
Generating train split: 78170686 examples [17:33, 72549.08 examples/s]
Generating train split: 78180686 examples [17:33, 67557.15 examples/s]
Generating train split: 78191686 examples [17:33, 57416.53 examples/s]
Generating train split: 78205686 examples [17:33, 69015.69 examples/s]
Generating train split: 78229686 examples [17:33, 97385.46 examples/s]
Generating train split: 78242686 examples [17:33, 93870.25 examples/s]
Generating train split: 78258686 examples [17:34, 84701.48 examples/s]
Generating train split: 78281686 examples [17:34, 105148.21 examples/s]
Generating train split: 78298686 examples [17:34, 94255.89 examples/s]
Generating train split: 78313686 examples [17:34, 92021.23 examples/s]
Generating train split: 78329686 examples [17:34, 82632.61 examples/s]
Generating train split: 78343686 examples [17:35, 78352.46 examples/s]
Generating train split: 78354686 examples [17:35, 80690.08 examples/s]
Generating train split: 78363686 examples [17:35, 79842.33 examples/s]
Generating train split: 78374686 examples [17:35, 51914.73 examples/s]
Generating train split: 78390686 examples [17:35, 67429.42 examples/s]
Generating train split: 78406686 examples [17:35, 82886.09 examples/s]
Generating train split: 78417686 examples [17:36, 81484.62 examples/s]
Generating train split: 78429686 examples [17:36, 82465.55 examples/s]
Generating train split: 78445686 examples [17:36, 98969.94 examples/s]
Generating train split: 78461686 examples [17:36, 63809.94 examples/s]
Generating train split: 78473686 examples [17:36, 66907.62 examples/s]
Generating train split: 78490686 examples [17:37, 82410.20 examples/s]
Generating train split: 78501686 examples [17:37, 67051.00 examples/s]
Generating train split: 78526686 examples [17:37, 95837.13 examples/s]
Generating train split: 78541686 examples [17:37, 100432.56 examples/s]
Generating train split: 78557686 examples [17:37, 76209.63 examples/s]
Generating train split: 78567686 examples [17:38, 77013.64 examples/s]
Generating train split: 78577686 examples [17:38, 78061.84 examples/s]
Generating train split: 78597686 examples [17:38, 102203.29 examples/s]
Generating train split: 78612686 examples [17:38, 94921.79 examples/s]
Generating train split: 78625686 examples [17:38, 100267.05 examples/s]
Generating train split: 78646686 examples [17:38, 123432.77 examples/s]
Generating train split: 78662686 examples [17:39, 70964.62 examples/s]
Generating train split: 78677686 examples [17:39, 64988.91 examples/s]
Generating train split: 78688686 examples [17:39, 68779.64 examples/s]
Generating train split: 78708686 examples [17:39, 89566.01 examples/s]
Generating train split: 78720686 examples [17:40, 56733.01 examples/s]
Generating train split: 78732686 examples [17:40, 61560.55 examples/s]
Generating train split: 78743686 examples [17:40, 69083.30 examples/s]
Generating train split: 78754686 examples [17:40, 62035.90 examples/s]
Generating train split: 78777686 examples [17:40, 91157.09 examples/s]
Generating train split: 78794686 examples [17:40, 97565.58 examples/s]
Generating train split: 78811686 examples [17:41, 53761.48 examples/s]
Generating train split: 78833686 examples [17:41, 66595.13 examples/s]
Generating train split: 78849686 examples [17:41, 76255.86 examples/s]
Generating train split: 78865686 examples [17:41, 89313.43 examples/s]
Generating train split: 78878686 examples [17:42, 85405.52 examples/s]
Generating train split: 78896686 examples [17:42, 95479.27 examples/s]
Generating train split: 78912686 examples [17:42, 78882.37 examples/s]
Generating train split: 78929686 examples [17:42, 92477.53 examples/s]
Generating train split: 78943686 examples [17:42, 62998.00 examples/s]
Generating train split: 78955686 examples [17:43, 69038.79 examples/s]
Generating train split: 78969686 examples [17:43, 67177.89 examples/s]
Generating train split: 78979686 examples [17:43, 72411.72 examples/s]
Generating train split: 78993686 examples [17:43, 54703.96 examples/s]
Generating train split: 79003686 examples [17:43, 60478.93 examples/s]
Generating train split: 79016686 examples [17:44, 65639.11 examples/s]
Generating train split: 79028686 examples [17:44, 56637.19 examples/s]
Generating train split: 79040686 examples [17:44, 64654.98 examples/s]
Generating train split: 79051686 examples [17:44, 72641.09 examples/s]
Generating train split: 79075686 examples [17:44, 70948.49 examples/s]
Generating train split: 79092686 examples [17:45, 85817.96 examples/s]
Generating train split: 79103686 examples [17:45, 40303.11 examples/s]
Generating train split: 79117686 examples [17:45, 50883.76 examples/s]
Generating train split: 79130686 examples [17:46, 57008.79 examples/s]
Generating train split: 79157686 examples [17:46, 87440.10 examples/s]
Generating train split: 79179686 examples [17:46, 101042.93 examples/s]
Generating train split: 79193686 examples [17:46, 98830.39 examples/s]
Generating train split: 79209686 examples [17:46, 79700.67 examples/s]
Generating train split: 79226686 examples [17:46, 91683.35 examples/s]
Generating train split: 79258686 examples [17:47, 132245.32 examples/s]
Generating train split: 79278686 examples [17:47, 110736.43 examples/s]
Generating train split: 79294686 examples [17:47, 105927.90 examples/s]
Generating train split: 79310686 examples [17:47, 113357.34 examples/s]
Generating train split: 79324686 examples [17:47, 100991.78 examples/s]
Generating train split: 79336686 examples [17:48, 52028.53 examples/s]
Generating train split: 79346686 examples [17:48, 57227.24 examples/s]
Generating train split: 79356686 examples [17:48, 58497.11 examples/s]
Generating train split: 79366686 examples [17:48, 58915.44 examples/s]
Generating train split: 79383686 examples [17:48, 71110.14 examples/s]
Generating train split: 79403686 examples [17:49, 92718.37 examples/s]
Generating train split: 79428686 examples [17:49, 115215.63 examples/s]
Generating train split: 79446686 examples [17:49, 82550.41 examples/s]
Generating train split: 79461686 examples [17:50, 51073.42 examples/s]
Generating train split: 79474686 examples [17:50, 55092.30 examples/s]
Generating train split: 79484686 examples [17:50, 60042.07 examples/s]
Generating train split: 79512686 examples [17:50, 91104.69 examples/s]
Generating train split: 79531686 examples [17:50, 103641.63 examples/s]
Generating train split: 79547686 examples [17:50, 98357.69 examples/s]
Generating train split: 79568686 examples [17:51, 85720.45 examples/s]
Generating train split: 79584686 examples [17:51, 96011.55 examples/s]
Generating train split: 79602686 examples [17:51, 82242.51 examples/s]
Generating train split: 79612686 examples [17:51, 74489.84 examples/s]
Generating train split: 79624686 examples [17:51, 76633.79 examples/s]
Generating train split: 79635686 examples [17:52, 71736.29 examples/s]
Generating train split: 79647686 examples [17:52, 79906.06 examples/s]
Generating train split: 79660686 examples [17:52, 41513.47 examples/s]
Generating train split: 79667686 examples [17:52, 44316.57 examples/s]
Generating train split: 79679686 examples [17:53, 51526.00 examples/s]
Generating train split: 79695686 examples [17:53, 65555.04 examples/s]
Generating train split: 79708686 examples [17:53, 72660.92 examples/s]
Generating train split: 79728686 examples [17:53, 94421.00 examples/s]
Generating train split: 79742686 examples [17:54, 39584.55 examples/s]
Generating train split: 79764686 examples [17:54, 55079.08 examples/s]
Generating train split: 79776686 examples [17:54, 58130.95 examples/s]
Generating train split: 79800686 examples [17:54, 81404.88 examples/s]
Generating train split: 79816686 examples [17:55, 82529.25 examples/s]
Generating train split: 79838686 examples [17:55, 96308.85 examples/s]
Generating train split: 79855686 examples [17:55, 105437.19 examples/s]
Generating train split: 79872686 examples [17:55, 74219.88 examples/s]
Generating train split: 79896686 examples [17:55, 96304.68 examples/s]
Generating train split: 79912686 examples [17:56, 94078.65 examples/s]
Generating train split: 79929686 examples [17:56, 94582.32 examples/s]
Generating train split: 79945686 examples [17:56, 62711.69 examples/s]
Generating train split: 79956686 examples [17:57, 51212.98 examples/s]
Generating train split: 79968686 examples [17:57, 52708.25 examples/s]
Generating train split: 79998686 examples [17:57, 84788.79 examples/s]
Generating train split: 80016686 examples [17:57, 96970.28 examples/s]
Generating train split: 80033686 examples [17:57, 68226.06 examples/s]
Generating train split: 80045686 examples [17:58, 66276.76 examples/s]
Generating train split: 80059686 examples [17:58, 60253.65 examples/s]
Generating train split: 80071686 examples [17:58, 59495.40 examples/s]
Generating train split: 80086686 examples [17:58, 69999.95 examples/s]
Generating train split: 80108686 examples [17:58, 90108.18 examples/s]
Generating train split: 80122686 examples [17:59, 53581.27 examples/s]
Generating train split: 80149686 examples [17:59, 80317.55 examples/s]
Generating train split: 80165686 examples [17:59, 91604.12 examples/s]
Generating train split: 80198686 examples [17:59, 132926.29 examples/s]
Generating train split: 80218686 examples [18:00, 92605.81 examples/s]
Generating train split: 80235686 examples [18:00, 86607.22 examples/s]
Generating train split: 80252686 examples [18:00, 91073.40 examples/s]
Generating train split: 80266686 examples [18:00, 92191.90 examples/s]
Generating train split: 80282686 examples [18:01, 57717.59 examples/s]
Generating train split: 80294686 examples [18:01, 63330.23 examples/s]
Generating train split: 80304686 examples [18:01, 68349.15 examples/s]
Generating train split: 80315686 examples [18:01, 61823.69 examples/s]
Generating train split: 80327686 examples [18:01, 70731.84 examples/s]
Generating train split: 80362686 examples [18:01, 120309.30 examples/s]
Generating train split: 80381686 examples [18:02, 128716.14 examples/s]
Generating train split: 80398686 examples [18:02, 59774.38 examples/s]
Generating train split: 80419686 examples [18:03, 60219.44 examples/s]
Generating train split: 80437686 examples [18:03, 73737.67 examples/s]
Generating train split: 80453686 examples [18:03, 83643.42 examples/s]
Generating train split: 80481686 examples [18:03, 105177.25 examples/s]
Generating train split: 80498686 examples [18:03, 69772.76 examples/s]
Generating train split: 80510686 examples [18:04, 73475.75 examples/s]
Generating train split: 80529686 examples [18:04, 91104.83 examples/s]
Generating train split: 80550686 examples [18:04, 96465.04 examples/s]
Generating train split: 80566686 examples [18:04, 84666.53 examples/s]
Generating train split: 80581686 examples [18:04, 70171.92 examples/s]
Generating train split: 80591686 examples [18:05, 73332.54 examples/s]
Generating train split: 80602686 examples [18:05, 41972.03 examples/s]
Generating train split: 80626686 examples [18:05, 64170.83 examples/s]
Generating train split: 80640686 examples [18:05, 73075.29 examples/s]
Generating train split: 80652686 examples [18:06, 65418.36 examples/s]
Generating train split: 80663686 examples [18:06, 70766.59 examples/s]
Generating train split: 80683686 examples [18:06, 92565.51 examples/s]
Generating train split: 80697686 examples [18:06, 51372.92 examples/s]
Generating train split: 80708686 examples [18:07, 51510.23 examples/s]
Generating train split: 80727686 examples [18:07, 52979.11 examples/s]
Generating train split: 80738686 examples [18:07, 56539.95 examples/s]
Generating train split: 80759686 examples [18:07, 77510.78 examples/s]
Generating train split: 80776686 examples [18:07, 91715.41 examples/s]
Generating train split: 80791686 examples [18:07, 101383.19 examples/s]
Generating train split: 80808686 examples [18:08, 111889.93 examples/s]
Generating train split: 80824686 examples [18:08, 78717.51 examples/s]
Generating train split: 80852686 examples [18:08, 105172.63 examples/s]
Generating train split: 80868686 examples [18:08, 83989.34 examples/s]
Generating train split: 80885686 examples [18:09, 62963.04 examples/s]
Generating train split: 80897686 examples [18:09, 67376.50 examples/s]
Generating train split: 80910686 examples [18:10, 43107.14 examples/s]
Generating train split: 80947686 examples [18:10, 77576.02 examples/s]
Generating train split: 80966686 examples [18:10, 92222.51 examples/s]
Generating train split: 80985686 examples [18:10, 74093.49 examples/s]
Generating train split: 81004686 examples [18:11, 64214.71 examples/s]
Generating train split: 81015686 examples [18:11, 57727.78 examples/s]
Generating train split: 81036686 examples [18:11, 76251.32 examples/s]
Generating train split: 81051686 examples [18:11, 81325.56 examples/s]
Generating train split: 81074686 examples [18:12, 68464.69 examples/s]
Generating train split: 81087686 examples [18:12, 74620.67 examples/s]
Generating train split: 81108686 examples [18:12, 94807.33 examples/s]
Generating train split: 81125686 examples [18:12, 106694.33 examples/s]
Generating train split: 81149686 examples [18:12, 119902.64 examples/s]
Generating train split: 81164686 examples [18:12, 98492.15 examples/s]
Generating train split: 81191686 examples [18:12, 117591.94 examples/s]
Generating train split: 81207686 examples [18:13, 53225.72 examples/s]
Generating train split: 81218686 examples [18:13, 58373.79 examples/s]
Generating train split: 81229686 examples [18:13, 59705.41 examples/s]
Generating train split: 81261686 examples [18:14, 97003.38 examples/s]
Generating train split: 81282686 examples [18:14, 78175.37 examples/s]
Generating train split: 81302686 examples [18:14, 91986.65 examples/s]
Generating train split: 81316686 examples [18:15, 66988.43 examples/s]
Generating train split: 81328686 examples [18:15, 51144.64 examples/s]
Generating train split: 81345686 examples [18:15, 63313.37 examples/s]
Generating train split: 81356686 examples [18:15, 61967.52 examples/s]
Generating train split: 81367686 examples [18:15, 63325.56 examples/s]
Generating train split: 81388686 examples [18:16, 84000.06 examples/s]
Generating train split: 81411686 examples [18:16, 56011.36 examples/s]
Generating train split: 81432686 examples [18:16, 73704.68 examples/s]
Generating train split: 81455686 examples [18:16, 88846.05 examples/s]
Generating train split: 81473686 examples [18:17, 81886.18 examples/s]
Generating train split: 81497686 examples [18:17, 102417.06 examples/s]
Generating train split: 81514686 examples [18:17, 83463.82 examples/s]
Generating train split: 81528686 examples [18:17, 87818.77 examples/s]
Generating train split: 81543686 examples [18:17, 93335.30 examples/s]
Generating train split: 81560686 examples [18:18, 56466.77 examples/s]
Generating train split: 81592686 examples [18:18, 88756.58 examples/s]
Generating train split: 81608686 examples [18:19, 61696.40 examples/s]
Generating train split: 81624686 examples [18:19, 56448.38 examples/s]
Generating train split: 81640686 examples [18:19, 67737.39 examples/s]
Generating train split: 81656686 examples [18:19, 64358.42 examples/s]
Generating train split: 81684686 examples [18:19, 90950.40 examples/s]
Generating train split: 81698686 examples [18:20, 64169.97 examples/s]
Generating train split: 81729686 examples [18:20, 90809.84 examples/s]
Generating train split: 81743686 examples [18:20, 78011.44 examples/s]
Generating train split: 81759686 examples [18:21, 71842.64 examples/s]
Generating train split: 81781686 examples [18:21, 92652.99 examples/s]
Generating train split: 81795686 examples [18:21, 69573.21 examples/s]
Generating train split: 81811686 examples [18:21, 81387.04 examples/s]
Generating train split: 81828686 examples [18:22, 65599.69 examples/s]
Generating train split: 81839686 examples [18:22, 54655.53 examples/s]
Generating train split: 81860686 examples [18:22, 54133.58 examples/s]
Generating train split: 81880686 examples [18:22, 70777.58 examples/s]
Generating train split: 81913686 examples [18:23, 59765.75 examples/s]
Generating train split: 81925686 examples [18:23, 49934.32 examples/s]
Generating train split: 81947686 examples [18:24, 62506.80 examples/s]
Generating train split: 81958686 examples [18:24, 61486.01 examples/s]
Generating train split: 81998686 examples [18:24, 105282.71 examples/s]
Generating train split: 82019686 examples [18:24, 92403.06 examples/s]
Generating train split: 82034686 examples [18:25, 78824.22 examples/s]
Generating train split: 82055686 examples [18:25, 96915.28 examples/s]
Generating train split: 82083686 examples [18:25, 126386.06 examples/s]
Generating train split: 82105686 examples [18:25, 98667.44 examples/s]
Generating train split: 82122686 examples [18:25, 107061.45 examples/s]
Generating train split: 82137686 examples [18:26, 74709.30 examples/s]
Generating train split: 82156686 examples [18:26, 90632.28 examples/s]
Generating train split: 82173686 examples [18:26, 57372.43 examples/s]
Generating train split: 82193686 examples [18:26, 72517.81 examples/s]
Generating train split: 82209686 examples [18:27, 59479.80 examples/s]
Generating train split: 82226686 examples [18:27, 72270.20 examples/s]
Generating train split: 82238686 examples [18:27, 45928.07 examples/s]
Generating train split: 82249686 examples [18:28, 52843.93 examples/s]
Generating train split: 82259686 examples [18:28, 50808.92 examples/s]
Generating train split: 82269686 examples [18:28, 57141.09 examples/s]
Generating train split: 82280686 examples [18:28, 59594.55 examples/s]
Generating train split: 82290686 examples [18:28, 63589.29 examples/s]
Generating train split: 82299686 examples [18:28, 67812.32 examples/s]
Generating train split: 82326686 examples [18:28, 109235.58 examples/s]
Generating train split: 82341686 examples [18:29, 93763.20 examples/s]
Generating train split: 82358686 examples [18:29, 79629.50 examples/s]
Generating train split: 82379686 examples [18:29, 100384.94 examples/s]
Generating train split: 82411686 examples [18:29, 143115.78 examples/s]
Generating train split: 82431686 examples [18:29, 104340.71 examples/s]
Generating train split: 82449686 examples [18:30, 113722.11 examples/s]
Generating train split: 82468686 examples [18:30, 71369.83 examples/s]
Generating train split: 82483686 examples [18:30, 78581.18 examples/s]
Generating train split: 82500686 examples [18:30, 90548.48 examples/s]
Generating train split: 82514686 examples [18:31, 79051.27 examples/s]
Generating train split: 82529686 examples [18:31, 50490.83 examples/s]
Generating train split: 82540686 examples [18:31, 57324.04 examples/s]
Generating train split: 82552686 examples [18:31, 58190.78 examples/s]
Generating train split: 82562686 examples [18:32, 47824.90 examples/s]
Generating train split: 82573686 examples [18:32, 55480.96 examples/s]
Generating train split: 82584686 examples [18:32, 54280.56 examples/s]
Generating train split: 82596686 examples [18:32, 63246.68 examples/s]
Generating train split: 82613686 examples [18:32, 77276.82 examples/s]
Generating train split: 82627686 examples [18:32, 87110.03 examples/s]
Generating train split: 82643686 examples [18:33, 74071.19 examples/s]
Generating train split: 82660686 examples [18:33, 91358.69 examples/s]
Generating train split: 82677686 examples [18:33, 99735.06 examples/s]
Generating train split: 82694686 examples [18:33, 85892.63 examples/s]
Generating train split: 82709686 examples [18:33, 80732.07 examples/s]
Generating train split: 82727686 examples [18:34, 72283.51 examples/s]
Generating train split: 82738686 examples [18:34, 77222.14 examples/s]
Generating train split: 82759686 examples [18:34, 99693.33 examples/s]
Generating train split: 82776686 examples [18:34, 67602.29 examples/s]
Generating train split: 82798686 examples [18:35, 89861.01 examples/s]
Generating train split: 82821686 examples [18:35, 112094.76 examples/s]
Generating train split: 82837686 examples [18:35, 76895.73 examples/s]
Generating train split: 82852686 examples [18:36, 47990.62 examples/s]
Generating train split: 82863686 examples [18:36, 48302.07 examples/s]
Generating train split: 82874686 examples [18:36, 46843.50 examples/s]
Generating train split: 82912686 examples [18:36, 84442.83 examples/s]
Generating train split: 82927686 examples [18:37, 78308.07 examples/s]
Generating train split: 82947686 examples [18:37, 90385.03 examples/s]
Generating train split: 82964686 examples [18:37, 94383.26 examples/s]
Generating train split: 82981686 examples [18:37, 83988.95 examples/s]
Generating train split: 82998686 examples [18:37, 94207.54 examples/s]
Generating train split: 83031686 examples [18:37, 131485.74 examples/s]
Generating train split: 83047686 examples [18:38, 82370.51 examples/s]
Generating train split: 83065686 examples [18:38, 93600.01 examples/s]
Generating train split: 83079686 examples [18:38, 75014.78 examples/s]
Generating train split: 83091686 examples [18:38, 71252.76 examples/s]
Generating train split: 83108686 examples [18:39, 81035.91 examples/s]
Generating train split: 83123686 examples [18:39, 71924.60 examples/s]
Generating train split: 83136686 examples [18:39, 66164.08 examples/s]
Generating train split: 83154686 examples [18:40, 53983.43 examples/s]
Generating train split: 83171686 examples [18:40, 66214.91 examples/s]
Generating train split: 83183686 examples [18:40, 52525.01 examples/s]
Generating train split: 83195686 examples [18:40, 41790.41 examples/s]
Generating train split: 83205686 examples [18:41, 48052.18 examples/s]
Generating train split: 83226686 examples [18:41, 65895.42 examples/s]
Generating train split: 83241686 examples [18:41, 75849.04 examples/s]
Generating train split: 83253686 examples [18:41, 82965.64 examples/s]
Generating train split: 83294686 examples [18:41, 144814.88 examples/s]
Generating train split: 83317686 examples [18:41, 146868.38 examples/s]
Generating train split: 83339686 examples [18:42, 95800.28 examples/s]
Generating train split: 83360686 examples [18:42, 90074.02 examples/s]
Generating train split: 83377686 examples [18:42, 78907.53 examples/s]
Generating train split: 83404686 examples [18:42, 104530.12 examples/s]
Generating train split: 83419686 examples [18:43, 83863.98 examples/s]
Generating train split: 83446686 examples [18:43, 106712.28 examples/s]
Generating train split: 83462686 examples [18:43, 110328.67 examples/s]
Generating train split: 83478686 examples [18:44, 58400.08 examples/s]
Generating train split: 83489686 examples [18:44, 49258.82 examples/s]
Generating train split: 83504686 examples [18:44, 60374.67 examples/s]
Generating train split: 83515686 examples [18:44, 52829.45 examples/s]
Generating train split: 83539686 examples [18:44, 73129.27 examples/s]
Generating train split: 83550686 examples [18:45, 68656.32 examples/s]
Generating train split: 83560686 examples [18:45, 58757.34 examples/s]
Generating train split: 83570686 examples [18:45, 60819.26 examples/s]
Generating train split: 83588686 examples [18:45, 67701.01 examples/s]
Generating train split: 83605686 examples [18:45, 82463.64 examples/s]
Generating train split: 83649686 examples [18:46, 141049.52 examples/s]
Generating train split: 83666686 examples [18:46, 146095.09 examples/s]
Generating train split: 83684686 examples [18:46, 78754.45 examples/s]
Generating train split: 83702686 examples [18:46, 70380.27 examples/s]
Generating train split: 83724686 examples [18:47, 85185.76 examples/s]
Generating train split: 83742686 examples [18:47, 76429.67 examples/s]
Generating train split: 83763686 examples [18:47, 94060.78 examples/s]
Generating train split: 83779686 examples [18:47, 83496.87 examples/s]
Generating train split: 83792686 examples [18:48, 51637.46 examples/s]
Generating train split: 83810686 examples [18:48, 52898.00 examples/s]
Generating train split: 83823686 examples [18:48, 58690.22 examples/s]
Generating train split: 83833686 examples [18:49, 45354.84 examples/s]
Generating train split: 83843686 examples [18:49, 51208.18 examples/s]
Generating train split: 83864686 examples [18:49, 71062.47 examples/s]
Generating train split: 83875686 examples [18:49, 62471.85 examples/s]
Generating train split: 83886686 examples [18:49, 66021.23 examples/s]
Generating train split: 83906686 examples [18:49, 87904.34 examples/s]
Generating train split: 83920686 examples [18:50, 81368.68 examples/s]
Generating train split: 83931686 examples [18:50, 85875.10 examples/s]
Generating train split: 83971686 examples [18:50, 148724.85 examples/s]
Generating train split: 83993686 examples [18:51, 72271.33 examples/s]
Generating train split: 84008686 examples [18:51, 70203.40 examples/s]
Generating train split: 84023686 examples [18:51, 79985.19 examples/s]
Generating train split: 84050686 examples [18:51, 108605.91 examples/s]
Generating train split: 84067686 examples [18:51, 109327.15 examples/s]
Generating train split: 84086686 examples [18:51, 100838.52 examples/s]
Generating train split: 84101686 examples [18:52, 108360.20 examples/s]
Generating train split: 84117686 examples [18:52, 88405.75 examples/s]
Generating train split: 84133686 examples [18:52, 63202.13 examples/s]
Generating train split: 84144686 examples [18:52, 66840.20 examples/s]
Generating train split: 84155686 examples [18:53, 58926.49 examples/s]
Generating train split: 84166686 examples [18:53, 55553.66 examples/s]
Generating train split: 84175686 examples [18:53, 45013.49 examples/s]
Generating train split: 84193686 examples [18:53, 61868.11 examples/s]
Generating train split: 84208686 examples [18:53, 69639.53 examples/s]
Generating train split: 84220686 examples [18:54, 71169.73 examples/s]
Generating train split: 84236686 examples [18:54, 83645.00 examples/s]
Generating train split: 84258686 examples [18:54, 109546.14 examples/s]
Generating train split: 84280686 examples [18:54, 107580.71 examples/s]
Generating train split: 84304686 examples [18:54, 125332.46 examples/s]
Generating train split: 84322686 examples [18:55, 77236.83 examples/s]
Generating train split: 84345686 examples [18:55, 93931.46 examples/s]
Generating train split: 84362686 examples [18:55, 72577.14 examples/s]
Generating train split: 84373686 examples [18:55, 77401.87 examples/s]
Generating train split: 84386686 examples [18:55, 84502.15 examples/s]
Generating train split: 84397686 examples [18:56, 82387.92 examples/s]
Generating train split: 84415686 examples [18:56, 95967.54 examples/s]
Generating train split: 84426686 examples [18:56, 98659.66 examples/s]
Generating train split: 84437686 examples [18:56, 89450.48 examples/s]
Generating train split: 84449686 examples [18:56, 68780.31 examples/s]
Generating train split: 84459686 examples [18:57, 50105.00 examples/s]
Generating train split: 84469686 examples [18:57, 53074.89 examples/s]
Generating train split: 84480686 examples [18:57, 52389.20 examples/s]
Generating train split: 84487686 examples [18:57, 41194.67 examples/s]
Generating train split: 84498686 examples [18:57, 42637.00 examples/s]
Generating train split: 84520686 examples [18:58, 68300.27 examples/s]
Generating train split: 84546686 examples [18:58, 97450.96 examples/s]
Generating train split: 84564686 examples [18:58, 109447.37 examples/s]
Generating train split: 84579686 examples [18:58, 101800.28 examples/s]
Generating train split: 84608686 examples [18:58, 136258.41 examples/s]
Generating train split: 84624686 examples [18:58, 109712.39 examples/s]
Generating train split: 84642686 examples [18:59, 97102.80 examples/s]
Generating train split: 84654686 examples [18:59, 70391.61 examples/s]
Generating train split: 84670686 examples [18:59, 82276.03 examples/s]
Generating train split: 84687686 examples [18:59, 62582.40 examples/s]
Generating train split: 84700686 examples [19:00, 58745.43 examples/s]
Generating train split: 84734686 examples [19:00, 97153.32 examples/s]
Generating train split: 84749686 examples [19:00, 88125.91 examples/s]
Generating train split: 84776686 examples [19:01, 65664.19 examples/s]
Generating train split: 84789686 examples [19:01, 63602.63 examples/s]
Generating train split: 84800686 examples [19:01, 43160.72 examples/s]
Generating train split: 84821686 examples [19:02, 56617.57 examples/s]
Generating train split: 84838686 examples [19:02, 60360.03 examples/s]
Generating train split: 84849686 examples [19:02, 66531.93 examples/s]
Generating train split: 84860686 examples [19:02, 72470.28 examples/s]
Generating train split: 84872686 examples [19:02, 80245.92 examples/s]
Generating train split: 84893686 examples [19:02, 87204.62 examples/s]
Generating train split: 84922686 examples [19:02, 124358.63 examples/s]
Generating train split: 84939686 examples [19:03, 125010.42 examples/s]
Generating train split: 84958686 examples [19:03, 89660.88 examples/s]
Generating train split: 84970686 examples [19:03, 84338.21 examples/s]
Generating train split: 84981686 examples [19:03, 76848.71 examples/s]
Generating train split: 85005686 examples [19:04, 85194.17 examples/s]
Generating train split: 85030686 examples [19:04, 111155.15 examples/s]
Generating train split: 85046686 examples [19:04, 86860.38 examples/s]
Generating train split: 85057686 examples [19:04, 86029.09 examples/s]
Generating train split: 85068686 examples [19:04, 65961.05 examples/s]
Generating train split: 85079686 examples [19:05, 70836.69 examples/s]
Generating train split: 85090686 examples [19:05, 48032.69 examples/s]
Generating train split: 85101686 examples [19:05, 49394.60 examples/s]
Generating train split: 85117686 examples [19:05, 63929.19 examples/s]
Generating train split: 85128686 examples [19:06, 40356.67 examples/s]
Generating train split: 85138686 examples [19:06, 47226.26 examples/s]
Generating train split: 85171686 examples [19:06, 85972.23 examples/s]
Generating train split: 85185686 examples [19:06, 72285.22 examples/s]
Generating train split: 85215686 examples [19:06, 104875.49 examples/s]
Generating train split: 85232686 examples [19:07, 104874.97 examples/s]
Generating train split: 85253686 examples [19:07, 117029.34 examples/s]
Generating train split: 85270686 examples [19:07, 79329.83 examples/s]
Generating train split: 85285686 examples [19:07, 86637.48 examples/s]
Generating train split: 85302686 examples [19:07, 88275.41 examples/s]
Generating train split: 85317686 examples [19:08, 61613.83 examples/s]
Generating train split: 85339686 examples [19:08, 81300.02 examples/s]
Generating train split: 85355686 examples [19:08, 76429.21 examples/s]
Generating train split: 85371686 examples [19:08, 80346.10 examples/s]
Generating train split: 85381686 examples [19:09, 82811.02 examples/s]
Generating train split: 85393686 examples [19:09, 69302.89 examples/s]
Generating train split: 85409686 examples [19:09, 83679.60 examples/s]
Generating train split: 85420686 examples [19:09, 80283.37 examples/s]
Generating train split: 85431686 examples [19:10, 43583.11 examples/s]
Generating train split: 85441686 examples [19:10, 48025.68 examples/s]
Generating train split: 85451686 examples [19:10, 37254.08 examples/s]
Generating train split: 85476686 examples [19:10, 59985.23 examples/s]
Generating train split: 85487686 examples [19:11, 61807.93 examples/s]
Generating train split: 85497686 examples [19:11, 65940.76 examples/s]
Generating train split: 85523686 examples [19:11, 100841.95 examples/s]
Generating train split: 85549686 examples [19:11, 122463.36 examples/s]
Generating train split: 85566686 examples [19:11, 117691.43 examples/s]
Generating train split: 85586686 examples [19:11, 130862.84 examples/s]
Generating train split: 85602686 examples [19:12, 90000.71 examples/s]
Generating train split: 85617686 examples [19:12, 81398.84 examples/s]
Generating train split: 85633686 examples [19:12, 94243.17 examples/s]
Generating train split: 85649686 examples [19:12, 58855.35 examples/s]
Generating train split: 85661686 examples [19:13, 57656.19 examples/s]
Generating train split: 85678686 examples [19:13, 69441.80 examples/s]
Generating train split: 85689686 examples [19:13, 60644.92 examples/s]
Generating train split: 85699686 examples [19:13, 56673.01 examples/s]
Generating train split: 85723686 examples [19:13, 82075.17 examples/s]
Generating train split: 85734686 examples [19:14, 80462.72 examples/s]
Generating train split: 85745686 examples [19:14, 42144.11 examples/s]
Generating train split: 85772686 examples [19:14, 67937.82 examples/s]
Generating train split: 85788686 examples [19:15, 62287.81 examples/s]
Generating train split: 85820686 examples [19:15, 95388.66 examples/s]
Generating train split: 85843686 examples [19:15, 92691.29 examples/s]
Generating train split: 85861686 examples [19:15, 104909.79 examples/s]
Generating train split: 85878686 examples [19:15, 98786.31 examples/s]
Generating train split: 85896686 examples [19:15, 104102.00 examples/s]
Generating train split: 85918686 examples [19:16, 120737.27 examples/s]
Generating train split: 85935686 examples [19:16, 109874.25 examples/s]
Generating train split: 85950686 examples [19:16, 76582.23 examples/s]
Generating train split: 85963686 examples [19:16, 83246.12 examples/s]
Generating train split: 85974686 examples [19:16, 82483.53 examples/s]
Generating train split: 85985686 examples [19:17, 52120.63 examples/s]
Generating train split: 85996686 examples [19:17, 49979.45 examples/s]
Generating train split: 86016686 examples [19:17, 70484.08 examples/s]
Generating train split: 86042686 examples [19:17, 95809.44 examples/s]
Generating train split: 86058686 examples [19:18, 78201.87 examples/s]
Generating train split: 86079686 examples [19:18, 98295.70 examples/s]
Generating train split: 86093686 examples [19:18, 73639.00 examples/s]
Generating train split: 86110686 examples [19:18, 61344.23 examples/s]
Generating train split: 86120686 examples [19:19, 65982.35 examples/s]
Generating train split: 86142686 examples [19:19, 69005.57 examples/s]
Generating train split: 86153686 examples [19:19, 74483.36 examples/s]
Generating train split: 86164686 examples [19:19, 79739.44 examples/s]
Generating train split: 86180686 examples [19:19, 93252.52 examples/s]
Generating train split: 86196686 examples [19:19, 97158.41 examples/s]
Generating train split: 86214686 examples [19:19, 108123.76 examples/s]
Generating train split: 86233686 examples [19:20, 114891.34 examples/s]
Generating train split: 86247686 examples [19:20, 82833.49 examples/s]
Generating train split: 86259686 examples [19:20, 60828.14 examples/s]
Generating train split: 86269686 examples [19:20, 54567.13 examples/s]
Generating train split: 86280686 examples [19:21, 59875.43 examples/s]
Generating train split: 86291686 examples [19:21, 60410.62 examples/s]
Generating train split: 86298686 examples [19:21, 41836.94 examples/s]
Generating train split: 86309686 examples [19:21, 41394.63 examples/s]
Generating train split: 86317686 examples [19:22, 45597.04 examples/s]
Generating train split: 86348686 examples [19:22, 84305.98 examples/s]
Generating train split: 86359686 examples [19:22, 78996.88 examples/s]
Generating train split: 86373686 examples [19:22, 72161.44 examples/s]
Generating train split: 86387686 examples [19:22, 59886.07 examples/s]
Generating train split: 86403686 examples [19:23, 73825.31 examples/s]
Generating train split: 86417686 examples [19:23, 56562.44 examples/s]
Generating train split: 86449686 examples [19:23, 78649.93 examples/s]
Generating train split: 86460686 examples [19:23, 72928.05 examples/s]
Generating train split: 86488686 examples [19:24, 98380.22 examples/s]
Generating train split: 86508686 examples [19:24, 110805.75 examples/s]
Generating train split: 86524686 examples [19:24, 110956.39 examples/s]
Generating train split: 86546686 examples [19:24, 131775.39 examples/s]
Generating train split: 86566686 examples [19:24, 73883.74 examples/s]
Generating train split: 86578686 examples [19:25, 65400.13 examples/s]
Generating train split: 86588686 examples [19:25, 67535.43 examples/s]
Generating train split: 86599686 examples [19:25, 49079.83 examples/s]
Generating train split: 86609686 examples [19:25, 53319.99 examples/s]
Generating train split: 86625686 examples [19:26, 66043.14 examples/s]
Generating train split: 86642686 examples [19:26, 79565.99 examples/s]
Generating train split: 86653686 examples [19:26, 52336.47 examples/s]
Generating train split: 86663686 examples [19:26, 58656.73 examples/s]
Generating train split: 86673686 examples [19:26, 60922.58 examples/s]
Generating train split: 86693686 examples [19:26, 83551.00 examples/s]
Generating train split: 86712686 examples [19:27, 74697.05 examples/s]
Generating train split: 86734686 examples [19:27, 94888.62 examples/s]
Generating train split: 86750686 examples [19:27, 63821.01 examples/s]
Generating train split: 86767686 examples [19:28, 68575.20 examples/s]
Generating train split: 86777686 examples [19:28, 62431.95 examples/s]
Generating train split: 86788686 examples [19:28, 67913.30 examples/s]
Generating train split: 86803686 examples [19:28, 78866.03 examples/s]
Generating train split: 86821686 examples [19:28, 92160.54 examples/s]
Generating train split: 86836686 examples [19:28, 94949.20 examples/s]
Generating train split: 86847686 examples [19:28, 95320.19 examples/s]
Generating train split: 86858686 examples [19:29, 50976.50 examples/s]
Generating train split: 86871686 examples [19:29, 51397.90 examples/s]
Generating train split: 86899686 examples [19:30, 45124.31 examples/s]
Generating train split: 86929686 examples [19:30, 67549.29 examples/s]
Generating train split: 86955686 examples [19:30, 89053.99 examples/s]
Generating train split: 86970686 examples [19:30, 76287.22 examples/s]
Generating train split: 86985686 examples [19:31, 59900.08 examples/s]
Generating train split: 87021686 examples [19:31, 93234.42 examples/s]
Generating train split: 87040686 examples [19:31, 82028.32 examples/s]
Generating train split: 87063686 examples [19:31, 92585.42 examples/s]
Generating train split: 87079686 examples [19:32, 68178.37 examples/s]
Generating train split: 87114686 examples [19:32, 97867.68 examples/s]
Generating train split: 87129686 examples [19:32, 78880.76 examples/s]
Generating train split: 87145686 examples [19:33, 77801.78 examples/s]
Generating train split: 87162686 examples [19:33, 88505.02 examples/s]
Generating train split: 87176686 examples [19:33, 73880.11 examples/s]
Generating train split: 87186686 examples [19:33, 76817.58 examples/s]
Generating train split: 87202686 examples [19:33, 61705.80 examples/s]
Generating train split: 87213686 examples [19:34, 52953.15 examples/s]
Generating train split: 87244686 examples [19:34, 65583.15 examples/s]
Generating train split: 87255686 examples [19:34, 59997.24 examples/s]
Generating train split: 87282686 examples [19:34, 87552.84 examples/s]
Generating train split: 87298686 examples [19:35, 72055.20 examples/s]
Generating train split: 87326686 examples [19:35, 66880.91 examples/s]
Generating train split: 87338686 examples [19:36, 63874.79 examples/s]
Generating train split: 87347686 examples [19:36, 57949.86 examples/s]
Generating train split: 87369686 examples [19:36, 77037.62 examples/s]
Generating train split: 87380686 examples [19:36, 73400.64 examples/s]
Generating train split: 87389686 examples [19:36, 65439.75 examples/s]
Generating train split: 87400686 examples [19:36, 71632.78 examples/s]
Generating train split: 87410686 examples [19:36, 71766.48 examples/s]
Generating train split: 87437686 examples [19:37, 110953.37 examples/s]
Generating train split: 87455686 examples [19:37, 61059.60 examples/s]
Generating train split: 87466686 examples [19:37, 57005.15 examples/s]
Generating train split: 87482686 examples [19:38, 69377.58 examples/s]
Generating train split: 87493686 examples [19:38, 74484.09 examples/s]
Generating train split: 87509686 examples [19:38, 45742.09 examples/s]
Generating train split: 87541686 examples [19:38, 77949.90 examples/s]
Generating train split: 87557686 examples [19:38, 87873.44 examples/s]
Generating train split: 87576686 examples [19:39, 85843.21 examples/s]
Generating train split: 87602686 examples [19:39, 108100.22 examples/s]
Generating train split: 87617686 examples [19:39, 73695.33 examples/s]
Generating train split: 87637686 examples [19:39, 89594.78 examples/s]
Generating train split: 87651686 examples [19:40, 51464.60 examples/s]
Generating train split: 87662686 examples [19:40, 50467.26 examples/s]
Generating train split: 87680686 examples [19:40, 64282.99 examples/s]
Generating train split: 87691686 examples [19:41, 69283.33 examples/s]
Generating train split: 87702686 examples [19:41, 69546.50 examples/s]
Generating train split: 87718686 examples [19:41, 77998.85 examples/s]
Generating train split: 87729686 examples [19:41, 74445.12 examples/s]
Generating train split: 87755686 examples [19:41, 109641.30 examples/s]
Generating train split: 87775686 examples [19:41, 113860.29 examples/s]
Generating train split: 87792686 examples [19:41, 104182.37 examples/s]
Generating train split: 87804686 examples [19:42, 48398.18 examples/s]
Generating train split: 87836686 examples [19:42, 78873.48 examples/s]
Generating train split: 87857686 examples [19:43, 56772.69 examples/s]
Generating train split: 87879686 examples [19:43, 70136.86 examples/s]
Generating train split: 87894686 examples [19:43, 78670.86 examples/s]
Generating train split: 87909686 examples [19:43, 84489.48 examples/s]
Generating train split: 87927686 examples [19:44, 61491.46 examples/s]
Generating train split: 87937686 examples [19:44, 63071.90 examples/s]
Generating train split: 87949686 examples [19:44, 67214.81 examples/s]
Generating train split: 87966686 examples [19:44, 80917.94 examples/s]
Generating train split: 87980686 examples [19:44, 84078.85 examples/s]
Generating train split: 87990686 examples [19:44, 86490.07 examples/s]
Generating train split: 88000686 examples [19:45, 82509.58 examples/s]
Generating train split: 88010686 examples [19:45, 61001.84 examples/s]
Generating train split: 88028686 examples [19:45, 81508.22 examples/s]
Generating train split: 88048686 examples [19:45, 101677.52 examples/s]
Generating train split: 88076686 examples [19:45, 139353.99 examples/s]
Generating train split: 88097686 examples [19:45, 155342.40 examples/s]
Generating train split: 88120686 examples [19:45, 167145.91 examples/s]
Generating train split: 88143686 examples [19:46, 174459.44 examples/s]
Generating train split: 88168686 examples [19:46, 191567.97 examples/s]
Generating train split: 88194686 examples [19:46, 196577.92 examples/s]
Generating train split: 88218686 examples [19:46, 145814.23 examples/s]
Generating train split: 88237686 examples [19:46, 143560.00 examples/s]
Generating train split: 88267686 examples [19:46, 165088.79 examples/s]
Generating train split: 88288686 examples [19:46, 163804.47 examples/s]
Generating train split: 88311686 examples [19:47, 150003.57 examples/s]
Generating train split: 88338686 examples [19:47, 173033.81 examples/s]
Generating train split: 88390686 examples [19:47, 251087.68 examples/s]
Generating train split: 88432686 examples [19:47, 290578.87 examples/s]
Generating train split: 88482686 examples [19:47, 310306.13 examples/s]
Generating train split: 88516686 examples [19:47, 260809.06 examples/s]
Generating train split: 88555686 examples [19:47, 268565.20 examples/s]
Generating train split: 88587686 examples [19:48, 209637.95 examples/s]
Generating train split: 88615686 examples [19:48, 162977.03 examples/s]
Generating train split: 88635686 examples [19:48, 134769.27 examples/s]
Generating train split: 88653686 examples [19:48, 131006.55 examples/s]
Generating train split: 88679686 examples [19:48, 141658.93 examples/s]
Generating train split: 88713686 examples [19:49, 173748.11 examples/s]
Generating train split: 88735686 examples [19:49, 97127.30 examples/s]
Generating train split: 88751686 examples [19:49, 104062.42 examples/s]
Generating train split: 88767686 examples [19:50, 73192.42 examples/s]
Generating train split: 88796686 examples [19:50, 98206.84 examples/s]
Generating train split: 88812686 examples [19:50, 106639.04 examples/s]
Generating train split: 88837686 examples [19:50, 99567.44 examples/s]
Generating train split: 88864686 examples [19:50, 117045.02 examples/s]
Generating train split: 88883686 examples [19:50, 127167.62 examples/s]
Generating train split: 88899686 examples [19:51, 107888.17 examples/s]
Generating train split: 88920686 examples [19:51, 80986.29 examples/s]
Generating train split: 88931686 examples [19:51, 80116.77 examples/s]
Generating train split: 88964686 examples [19:51, 120239.29 examples/s]
Generating train split: 88988686 examples [19:51, 129714.12 examples/s]
Generating train split: 89007686 examples [19:52, 128456.93 examples/s]
Generating train split: 89024686 examples [19:52, 129724.70 examples/s]
Generating train split: 89040686 examples [19:52, 116957.13 examples/s]
Generating train split: 89065686 examples [19:53, 65840.03 examples/s]
Generating train split: 89088686 examples [19:53, 80262.93 examples/s]
Generating train split: 89104686 examples [19:53, 91021.98 examples/s]
Generating train split: 89117686 examples [19:53, 96683.80 examples/s]
Generating train split: 89135686 examples [19:53, 78963.56 examples/s]
Generating train split: 89194686 examples [19:53, 155130.51 examples/s]
Generating train split: 89216686 examples [19:54, 74234.09 examples/s]
Generating train split: 89233686 examples [19:54, 76473.79 examples/s]
Generating train split: 89248686 examples [19:55, 82998.59 examples/s]
Generating train split: 89266686 examples [19:55, 90115.67 examples/s]
Generating train split: 89283686 examples [19:55, 83534.76 examples/s]
Generating train split: 89310686 examples [19:55, 106971.81 examples/s]
Generating train split: 89351686 examples [19:56, 82292.03 examples/s]
Generating train split: 89371686 examples [19:56, 95566.38 examples/s]
Generating train split: 89396686 examples [19:56, 111867.75 examples/s]
Generating train split: 89414686 examples [19:56, 107004.47 examples/s]
Generating train split: 89433686 examples [19:56, 90564.34 examples/s]
Generating train split: 89464686 examples [19:57, 120486.92 examples/s]
Generating train split: 89513686 examples [19:57, 123073.41 examples/s]
Generating train split: 89531686 examples [19:57, 102431.44 examples/s]
Generating train split: 89548686 examples [19:57, 109914.08 examples/s]
Generating train split: 89571686 examples [19:58, 110694.58 examples/s]
Generating train split: 89614686 examples [19:58, 155818.71 examples/s]
Generating train split: 89638686 examples [19:58, 118261.19 examples/s]
Generating train split: 89655686 examples [19:58, 122002.41 examples/s]
Generating train split: 89670686 examples [19:58, 112841.09 examples/s]
Generating train split: 89688686 examples [19:59, 84536.60 examples/s]
Generating train split: 89703686 examples [19:59, 90741.81 examples/s]
Generating train split: 89722686 examples [19:59, 107081.49 examples/s]
Generating train split: 89739686 examples [19:59, 105499.67 examples/s]
Generating train split: 89756686 examples [19:59, 78785.83 examples/s]
Generating train split: 89768686 examples [20:00, 82260.53 examples/s]
Generating train split: 89801686 examples [20:00, 121173.29 examples/s]
Generating train split: 89838686 examples [20:00, 127737.08 examples/s]
Generating train split: 89855686 examples [20:00, 105852.89 examples/s]
Generating train split: 89873686 examples [20:00, 102439.25 examples/s]
Generating train split: 89885686 examples [20:01, 76689.70 examples/s]
Generating train split: 89903686 examples [20:01, 91079.73 examples/s]
Generating train split: 89921686 examples [20:01, 98227.42 examples/s]
Generating train split: 89937686 examples [20:01, 104068.91 examples/s]
Generating train split: 89985686 examples [20:01, 177929.65 examples/s]
Generating train split: 90012686 examples [20:02, 101613.08 examples/s]
Generating train split: 90035686 examples [20:02, 98988.80 examples/s]
Generating train split: 90050686 examples [20:02, 83149.71 examples/s]
Generating train split: 90077686 examples [20:02, 106596.56 examples/s]
Generating train split: 90116686 examples [20:03, 132251.68 examples/s]
Generating train split: 90137686 examples [20:03, 128052.68 examples/s]
Generating train split: 90159686 examples [20:03, 111652.15 examples/s]
Generating train split: 90176686 examples [20:03, 81411.75 examples/s]
Generating train split: 90191686 examples [20:04, 56472.90 examples/s]
Generating train split: 90234686 examples [20:04, 95347.84 examples/s]
Generating train split: 90267686 examples [20:04, 118704.80 examples/s]
Generating train split: 90288686 examples [20:04, 126076.05 examples/s]
Generating train split: 90310686 examples [20:05, 119663.72 examples/s]
Generating train split: 90326686 examples [20:05, 61302.33 examples/s]
Generating train split: 90359686 examples [20:05, 83043.27 examples/s]
Generating train split: 90374686 examples [20:06, 90486.40 examples/s]
Generating train split: 90417686 examples [20:06, 136673.94 examples/s]
Generating train split: 90441686 examples [20:07, 56997.11 examples/s]
Generating train split: 90468686 examples [20:07, 74340.02 examples/s]
Generating train split: 90489686 examples [20:07, 80161.56 examples/s]
Generating train split: 90536686 examples [20:07, 124069.26 examples/s]
Generating train split: 90563686 examples [20:07, 121845.38 examples/s]
Generating train split: 90583686 examples [20:08, 60683.50 examples/s]
Generating train split: 90598686 examples [20:08, 67286.45 examples/s]
Generating train split: 90640686 examples [20:09, 105541.46 examples/s]
Generating train split: 90667686 examples [20:09, 123735.22 examples/s]
Generating train split: 90694686 examples [20:09, 127217.76 examples/s]
Generating train split: 90719686 examples [20:09, 94232.92 examples/s]
Generating train split: 90738686 examples [20:10, 81445.21 examples/s]
Generating train split: 90754686 examples [20:10, 84342.67 examples/s]
Generating train split: 90770686 examples [20:10, 88745.22 examples/s]
Generating train split: 90786686 examples [20:10, 99424.64 examples/s]
Generating train split: 90801686 examples [20:11, 62530.58 examples/s]
Generating train split: 90837686 examples [20:11, 98502.63 examples/s]
Generating train split: 90856686 examples [20:11, 61983.47 examples/s]
Generating train split: 90894686 examples [20:12, 92510.62 examples/s]
Generating train split: 90911686 examples [20:12, 95949.93 examples/s]
Generating train split: 90927686 examples [20:12, 87438.30 examples/s]
Generating train split: 90943686 examples [20:12, 96748.32 examples/s]
Generating train split: 90964686 examples [20:12, 73860.62 examples/s]
Generating train split: 90991686 examples [20:13, 95028.82 examples/s]
Generating train split: 91006686 examples [20:13, 92881.53 examples/s]
Generating train split: 91031686 examples [20:13, 91924.52 examples/s]
Generating train split: 91053686 examples [20:13, 104594.65 examples/s]
Generating train split: 91068686 examples [20:13, 87979.58 examples/s]
Generating train split: 91093686 examples [20:14, 111320.63 examples/s]
Generating train split: 91115686 examples [20:14, 110189.95 examples/s]
Generating train split: 91130686 examples [20:14, 86444.54 examples/s]
Generating train split: 91147686 examples [20:14, 98657.37 examples/s]
Generating train split: 91182686 examples [20:14, 135431.07 examples/s]
Generating train split: 91205686 examples [20:15, 79027.38 examples/s]
Generating train split: 91225686 examples [20:15, 88725.23 examples/s]
Generating train split: 91252686 examples [20:15, 111097.19 examples/s]
Generating train split: 91282686 examples [20:15, 135564.96 examples/s]
Generating train split: 91300686 examples [20:16, 89199.71 examples/s]
Generating train split: 91323686 examples [20:16, 97741.64 examples/s]
Generating train split: 91337686 examples [20:16, 81884.90 examples/s]
Generating train split: 91351686 examples [20:16, 86217.89 examples/s]
Generating train split: 91362686 examples [20:16, 89197.63 examples/s]
Generating train split: 91373686 examples [20:17, 84059.99 examples/s]
Generating train split: 91408686 examples [20:17, 133086.68 examples/s]
Generating train split: 91436686 examples [20:17, 112440.92 examples/s]
Generating train split: 91454686 examples [20:17, 85437.68 examples/s]
Generating train split: 91475686 examples [20:18, 97863.90 examples/s]
Generating train split: 91491686 examples [20:18, 107011.07 examples/s]
Generating train split: 91505686 examples [20:18, 103256.96 examples/s]
Generating train split: 91524686 examples [20:18, 119348.98 examples/s]
Generating train split: 91540686 examples [20:18, 106759.94 examples/s]
Generating train split: 91557686 examples [20:18, 86437.75 examples/s]
Generating train split: 91573686 examples [20:19, 80948.28 examples/s]
Generating train split: 91583686 examples [20:19, 71791.05 examples/s]
Generating train split: 91615686 examples [20:19, 72014.55 examples/s]
Generating train split: 91633686 examples [20:19, 85519.35 examples/s]
Generating train split: 91652686 examples [20:19, 96302.88 examples/s]
Generating train split: 91671686 examples [20:20, 100474.08 examples/s]
Generating train split: 91693686 examples [20:20, 122136.55 examples/s]
Generating train split: 91714686 examples [20:20, 132815.42 examples/s]
Generating train split: 91730686 examples [20:20, 93861.88 examples/s]
Generating train split: 91747686 examples [20:21, 65870.49 examples/s]
Generating train split: 91759686 examples [20:21, 68189.49 examples/s]
Generating train split: 91769686 examples [20:21, 62925.54 examples/s]
Generating train split: 91785686 examples [20:21, 77714.42 examples/s]
Generating train split: 91835686 examples [20:21, 147918.19 examples/s]
Generating train split: 91856686 examples [20:22, 102327.16 examples/s]
Generating train split: 91878686 examples [20:22, 66177.23 examples/s]
Generating train split: 91923686 examples [20:22, 107177.50 examples/s]
Generating train split: 91945686 examples [20:23, 113804.03 examples/s]
Generating train split: 91969686 examples [20:23, 129027.64 examples/s]
Generating train split: 91992686 examples [20:23, 112376.83 examples/s]
Generating train split: 92010686 examples [20:23, 121646.83 examples/s]
Generating train split: 92028686 examples [20:23, 101756.24 examples/s]
Generating train split: 92045686 examples [20:24, 64393.07 examples/s]
Generating train split: 92073686 examples [20:24, 81479.56 examples/s]
Generating train split: 92106686 examples [20:24, 114738.04 examples/s]
Generating train split: 92128686 examples [20:24, 122152.47 examples/s]
Generating train split: 92151686 examples [20:24, 136530.13 examples/s]
Generating train split: 92169686 examples [20:25, 67037.93 examples/s]
Generating train split: 92199686 examples [20:25, 89080.16 examples/s]
Generating train split: 92217686 examples [20:25, 93729.02 examples/s]
Generating train split: 92239686 examples [20:26, 111668.11 examples/s]
Generating train split: 92261686 examples [20:26, 95902.22 examples/s]
Generating train split: 92278686 examples [20:26, 97764.01 examples/s]
Generating train split: 92295686 examples [20:26, 98610.85 examples/s]
Generating train split: 92311686 examples [20:27, 67300.09 examples/s]
Generating train split: 92333686 examples [20:27, 84966.45 examples/s]
Generating train split: 92349686 examples [20:27, 93190.74 examples/s]
Generating train split: 92388686 examples [20:27, 142037.41 examples/s]
Generating train split: 92417686 examples [20:27, 148171.74 examples/s]
Generating train split: 92436686 examples [20:28, 79901.92 examples/s]
Generating train split: 92453686 examples [20:28, 80607.36 examples/s]
Generating train split: 92470686 examples [20:28, 85682.07 examples/s]
Generating train split: 92486686 examples [20:28, 82545.92 examples/s]
Generating train split: 92520686 examples [20:28, 118799.74 examples/s]
Generating train split: 92537686 examples [20:29, 124328.40 examples/s]
Generating train split: 92553686 examples [20:29, 117061.09 examples/s]
Generating train split: 92569686 examples [20:29, 62503.21 examples/s]
Generating train split: 92586686 examples [20:29, 75640.64 examples/s]
Generating train split: 92602686 examples [20:30, 82666.42 examples/s]
Generating train split: 92633686 examples [20:30, 120418.03 examples/s]
Generating train split: 92665686 examples [20:30, 145022.24 examples/s]
Generating train split: 92687686 examples [20:30, 84844.02 examples/s]
Generating train split: 92703686 examples [20:31, 85528.77 examples/s]
Generating train split: 92726686 examples [20:31, 106362.37 examples/s]
Generating train split: 92749686 examples [20:31, 81682.41 examples/s]
Generating train split: 92767686 examples [20:31, 79003.53 examples/s]
Generating train split: 92795686 examples [20:31, 101504.71 examples/s]
Generating train split: 92813686 examples [20:32, 105188.32 examples/s]
Generating train split: 92829686 examples [20:32, 88292.79 examples/s]
Generating train split: 92844686 examples [20:32, 72566.42 examples/s]
Generating train split: 92871686 examples [20:32, 96929.36 examples/s]
Generating train split: 92888686 examples [20:33, 77615.09 examples/s]
Generating train split: 92904686 examples [20:33, 83883.61 examples/s]
Generating train split: 92925686 examples [20:33, 104149.57 examples/s]
Generating train split: 92958686 examples [20:34, 70620.02 examples/s]
Generating train split: 92974686 examples [20:34, 80428.74 examples/s]
Generating train split: 93001686 examples [20:34, 103369.91 examples/s]
Generating train split: 93023686 examples [20:34, 112147.48 examples/s]
Generating train split: 93039686 examples [20:34, 88592.32 examples/s]
Generating train split: 93070686 examples [20:34, 121814.77 examples/s]
Generating train split: 93091686 examples [20:35, 102938.08 examples/s]
Generating train split: 93107686 examples [20:35, 105486.96 examples/s]
Generating train split: 93126686 examples [20:35, 95539.46 examples/s]
Generating train split: 93142686 examples [20:36, 66652.93 examples/s]
Generating train split: 93152686 examples [20:36, 70646.49 examples/s]
Generating train split: 93172686 examples [20:36, 84380.06 examples/s]
Generating train split: 93183686 examples [20:36, 85461.34 examples/s]
Generating train split: 93224686 examples [20:36, 145281.11 examples/s]
Generating train split: 93245686 examples [20:37, 84662.05 examples/s]
Generating train split: 93264686 examples [20:37, 80715.62 examples/s]
Generating train split: 93282686 examples [20:37, 85128.66 examples/s]
Generating train split: 93319686 examples [20:37, 90975.04 examples/s]
Generating train split: 93334686 examples [20:38, 85352.83 examples/s]
Generating train split: 93357686 examples [20:38, 101802.15 examples/s]
Generating train split: 93373686 examples [20:38, 78772.36 examples/s]
Generating train split: 93386686 examples [20:38, 78701.52 examples/s]
Generating train split: 93396686 examples [20:38, 78767.43 examples/s]
Generating train split: 93439686 examples [20:38, 136930.84 examples/s]
Generating train split: 93461686 examples [20:39, 121034.65 examples/s]
Generating train split: 93477686 examples [20:39, 86606.68 examples/s]
Generating train split: 93494686 examples [20:39, 96627.46 examples/s]
Generating train split: 93510686 examples [20:40, 65506.22 examples/s]
Generating train split: 93543686 examples [20:40, 95640.37 examples/s]
Generating train split: 93574686 examples [20:40, 127629.27 examples/s]
Generating train split: 93595686 examples [20:40, 119606.49 examples/s]
Generating train split: 93616686 examples [20:40, 99421.89 examples/s]
Generating train split: 93631686 examples [20:41, 63800.65 examples/s]
Generating train split: 93645686 examples [20:41, 62362.17 examples/s]
Generating train split: 93678686 examples [20:41, 95025.12 examples/s]
Generating train split: 93695686 examples [20:41, 102339.88 examples/s]
Generating train split: 93710686 examples [20:42, 107219.98 examples/s]
Generating train split: 93736686 examples [20:42, 115675.94 examples/s]
Generating train split: 93752686 examples [20:42, 61526.03 examples/s]
Generating train split: 93766686 examples [20:43, 62239.90 examples/s]
Generating train split: 93777686 examples [20:43, 61500.15 examples/s]
Generating train split: 93802686 examples [20:43, 88099.38 examples/s]
Generating train split: 93836686 examples [20:43, 98509.83 examples/s]
Generating train split: 93850686 examples [20:43, 103910.18 examples/s]
Generating train split: 93866686 examples [20:44, 72595.46 examples/s]
Generating train split: 93881686 examples [20:44, 82407.57 examples/s]
Generating train split: 93896686 examples [20:44, 85244.94 examples/s]
Generating train split: 93914686 examples [20:44, 94218.10 examples/s]
Generating train split: 93945686 examples [20:44, 132984.25 examples/s]
Generating train split: 93966686 examples [20:44, 132386.78 examples/s]
Generating train split: 93983686 examples [20:45, 76222.39 examples/s]
Generating train split: 93995686 examples [20:45, 76328.95 examples/s]
Generating train split: 94006686 examples [20:45, 71111.58 examples/s]
Generating train split: 94029686 examples [20:45, 96745.32 examples/s]
Generating train split: 94051686 examples [20:45, 114794.00 examples/s]
Generating train split: 94067686 examples [20:46, 108546.93 examples/s]
Generating train split: 94087686 examples [20:46, 123804.48 examples/s]
Generating train split: 94104686 examples [20:46, 124193.64 examples/s]
Generating train split: 94121686 examples [20:46, 72984.01 examples/s]
Generating train split: 94132686 examples [20:47, 51754.33 examples/s]
Generating train split: 94141686 examples [20:47, 55990.38 examples/s]
Generating train split: 94175686 examples [20:47, 97213.18 examples/s]
Generating train split: 94197686 examples [20:47, 113556.62 examples/s]
Generating train split: 94225686 examples [20:47, 142924.53 examples/s]
Generating train split: 94246686 examples [20:47, 123852.13 examples/s]
Generating train split: 94264686 examples [20:48, 66999.17 examples/s]
Generating train split: 94279686 examples [20:48, 68586.87 examples/s]
Generating train split: 94314686 examples [20:48, 104988.62 examples/s]
Generating train split: 94335686 examples [20:49, 101327.16 examples/s]
Generating train split: 94352686 examples [20:49, 104201.50 examples/s]
Generating train split: 94379686 examples [20:49, 81356.32 examples/s]
Generating train split: 94391686 examples [20:49, 76351.32 examples/s]
Generating train split: 94404686 examples [20:50, 79081.48 examples/s]
Generating train split: 94420686 examples [20:50, 78188.17 examples/s]
Generating train split: 94448686 examples [20:50, 108309.17 examples/s]
Generating train split: 94491686 examples [20:50, 163320.70 examples/s]
Generating train split: 94513686 examples [20:50, 105091.87 examples/s]
Generating train split: 94535686 examples [20:51, 96517.12 examples/s]
Generating train split: 94551686 examples [20:51, 84601.33 examples/s]
Generating train split: 94577686 examples [20:51, 108846.45 examples/s]
Generating train split: 94616686 examples [20:51, 104216.59 examples/s]
Generating train split: 94631686 examples [20:52, 106679.94 examples/s]
Generating train split: 94652686 examples [20:52, 75789.76 examples/s]
Generating train split: 94664686 examples [20:52, 78520.47 examples/s]
Generating train split: 94679686 examples [20:52, 76775.38 examples/s]
Generating train split: 94692686 examples [20:53, 76126.25 examples/s]
Generating train split: 94707686 examples [20:53, 83878.18 examples/s]
Generating train split: 94737686 examples [20:53, 119355.30 examples/s]
Generating train split: 94754686 examples [20:53, 116741.53 examples/s]
Generating train split: 94769686 examples [20:53, 115656.98 examples/s]
Generating train split: 94786686 examples [20:54, 66364.49 examples/s]
Generating train split: 94801686 examples [20:54, 67407.04 examples/s]
Generating train split: 94818686 examples [20:54, 79482.18 examples/s]
Generating train split: 94834686 examples [20:54, 80766.18 examples/s]
Generating train split: 94862686 examples [20:54, 112185.19 examples/s]
Generating train split: 94879686 examples [20:54, 121709.27 examples/s]
Generating train split: 94895686 examples [20:55, 117371.72 examples/s]
Generating train split: 94909686 examples [20:55, 95432.44 examples/s]
Generating train split: 94924686 examples [20:55, 54636.99 examples/s]
Generating train split: 94956686 examples [20:55, 87229.19 examples/s]
Generating train split: 94973686 examples [20:56, 90270.25 examples/s]
Generating train split: 94987686 examples [20:56, 75631.30 examples/s]
Generating train split: 95033686 examples [20:56, 118128.29 examples/s]
Generating train split: 95050686 examples [20:56, 93436.68 examples/s]
Generating train split: 95067686 examples [20:57, 75732.25 examples/s]
Generating train split: 95078686 examples [20:57, 78242.36 examples/s]
Generating train split: 95101686 examples [20:57, 98252.16 examples/s]
Generating train split: 95119686 examples [20:57, 90728.82 examples/s]
Generating train split: 95141686 examples [20:58, 82108.10 examples/s]
Generating train split: 95158686 examples [20:58, 94606.30 examples/s]
Generating train split: 95187686 examples [20:58, 96148.29 examples/s]
Generating train split: 95199686 examples [20:58, 91120.34 examples/s]
Generating train split: 95217686 examples [20:58, 78723.96 examples/s]
Generating train split: 95238686 examples [20:59, 97387.49 examples/s]
Generating train split: 95253686 examples [20:59, 100655.14 examples/s]
Generating train split: 95269686 examples [20:59, 82413.78 examples/s]
Generating train split: 95315686 examples [20:59, 140274.34 examples/s]
Generating train split: 95338686 examples [21:00, 101754.37 examples/s]
Generating train split: 95355686 examples [21:00, 79097.88 examples/s]
Generating train split: 95367686 examples [21:00, 69056.35 examples/s]
Generating train split: 95392686 examples [21:00, 91539.22 examples/s]
Generating train split: 95408686 examples [21:00, 96468.29 examples/s]
Generating train split: 95438686 examples [21:01, 126392.01 examples/s]
Generating train split: 95455686 examples [21:01, 119619.94 examples/s]
Generating train split: 95471686 examples [21:01, 127323.33 examples/s]
Generating train split: 95486686 examples [21:01, 74836.81 examples/s]
Generating train split: 95501686 examples [21:02, 57108.63 examples/s]
Generating train split: 95525686 examples [21:02, 78009.42 examples/s]
Generating train split: 95542686 examples [21:02, 88539.40 examples/s]
Generating train split: 95569686 examples [21:02, 118601.18 examples/s]
Generating train split: 95595686 examples [21:02, 121367.27 examples/s]
Generating train split: 95612686 examples [21:03, 65403.37 examples/s]
Generating train split: 95629686 examples [21:03, 57261.63 examples/s]
Generating train split: 95645686 examples [21:03, 66052.78 examples/s]
Generating train split: 95668686 examples [21:04, 85565.62 examples/s]
Generating train split: 95707686 examples [21:04, 130913.30 examples/s]
Generating train split: 95736686 examples [21:04, 151112.80 examples/s]
Generating train split: 95758686 examples [21:05, 74029.71 examples/s]
Generating train split: 95774686 examples [21:05, 60720.60 examples/s]
Generating train split: 95788686 examples [21:05, 68468.22 examples/s]
Generating train split: 95815686 examples [21:05, 93904.48 examples/s]
Generating train split: 95839686 examples [21:05, 113761.67 examples/s]
Generating train split: 95884686 examples [21:05, 170700.74 examples/s]
Generating train split: 95912686 examples [21:06, 66389.48 examples/s]
Generating train split: 95932686 examples [21:07, 75471.26 examples/s]
Generating train split: 95973686 examples [21:07, 108366.63 examples/s]
Generating train split: 95999686 examples [21:07, 117979.88 examples/s]
Generating train split: 96032686 examples [21:07, 146849.00 examples/s]
Generating train split: 96055686 examples [21:07, 99553.95 examples/s]
Generating train split: 96076686 examples [21:08, 87632.47 examples/s]
Generating train split: 96092686 examples [21:08, 86580.69 examples/s]
Generating train split: 96115686 examples [21:08, 105223.04 examples/s]
Generating train split: 96134686 examples [21:08, 116704.77 examples/s]
Generating train split: 96151686 examples [21:08, 120193.36 examples/s]
Generating train split: 96187686 examples [21:08, 163657.67 examples/s]
Generating train split: 96210686 examples [21:09, 59030.68 examples/s]
Generating train split: 96226686 examples [21:10, 68048.77 examples/s]
Generating train split: 96243686 examples [21:10, 72965.84 examples/s]
Generating train split: 96281686 examples [21:10, 111817.53 examples/s]
Generating train split: 96314686 examples [21:10, 145064.14 examples/s]
Generating train split: 96337686 examples [21:11, 58218.04 examples/s]
Generating train split: 96364686 examples [21:11, 75937.94 examples/s]
Generating train split: 96386686 examples [21:11, 91219.33 examples/s]
Generating train split: 96409686 examples [21:11, 108425.49 examples/s]
Generating train split: 96441686 examples [21:11, 140051.80 examples/s]
Generating train split: 96472686 examples [21:12, 166252.67 examples/s]
Generating train split: 96498686 examples [21:13, 66951.57 examples/s]
Generating train split: 96519686 examples [21:13, 77491.94 examples/s]
Generating train split: 96542686 examples [21:13, 94563.54 examples/s]
Generating train split: 96563686 examples [21:13, 105365.79 examples/s]
Generating train split: 96604686 examples [21:13, 149090.75 examples/s]
Generating train split: 96630686 examples [21:14, 70760.66 examples/s]
Generating train split: 96656686 examples [21:14, 76794.53 examples/s]
Generating train split: 96681686 examples [21:14, 94297.47 examples/s]
Generating train split: 96722686 examples [21:14, 134991.58 examples/s]
Generating train split: 96750686 examples [21:15, 151027.75 examples/s]
Generating train split: 96776686 examples [21:15, 80652.45 examples/s]
Generating train split: 96798686 examples [21:15, 89638.83 examples/s]
Generating train split: 96826686 examples [21:16, 92259.69 examples/s]
Generating train split: 96861686 examples [21:16, 123779.94 examples/s]
Generating train split: 96903686 examples [21:16, 163649.60 examples/s]
Generating train split: 96931686 examples [21:16, 157559.50 examples/s]
Generating train split: 96956686 examples [21:16, 141913.71 examples/s]
Generating train split: 96978686 examples [21:16, 146097.74 examples/s]
Generating train split: 96999686 examples [21:17, 150438.23 examples/s]
Generating train split: 97028686 examples [21:17, 175039.91 examples/s]
Generating train split: 97069686 examples [21:17, 219452.98 examples/s]
Generating train split: 97096686 examples [21:17, 210930.96 examples/s]
Generating train split: 97133686 examples [21:17, 236351.08 examples/s]
Generating train split: 97177686 examples [21:17, 282459.70 examples/s]
Generating train split: 97209686 examples [21:17, 229385.49 examples/s]
Generating train split: 97244686 examples [21:18, 210158.81 examples/s]
Generating train split: 97270686 examples [21:18, 145000.39 examples/s]
Generating train split: 97270686 examples [21:18, 76079.02 examples/s]
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:44,304 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:44,339 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 11:35:44,340 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:35:44,340 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:44,340 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:07,595 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:08,012 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 11:36:08,012 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:36:08,013 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:08,013 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:40:37,638 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 11:40:38,351 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 11:40:38,493 - root - INFO - Building dataloader...
+[titan] 2025-07-23 11:40:38,496 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 11:40:38,498 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 11:40:39,687 - fla.layers.mamba2 - WARNING - The fast path is not available because one of `(selective_state_update)` is None. Falling back to the naive implementation. To install follow https://github.com/state-spaces/mamba/#installation
+[titan] 2025-07-23 11:40:39,688 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 11:40:39,808 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 11:40:39,860 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 11:40:39,860 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 11:40:39,860 - root - WARNING - No norm found in model
+[titan] 2025-07-23 11:40:39,860 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 11:40:40,073 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 11:40:40,429 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 11:40:40,596 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 11:40:40,837 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 11:40:40,839 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 11:40:40,870 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 11:40:40,871 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 11:40:41,100 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 11:40:49,669 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 11:40:49,687 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 11:40:49,687 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 11:40:49,703 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 11:40:49,704 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 11:40:49,704 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 11:40:49,704 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 11:40:49,704 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 11:40:49,705 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 11:40:49,705 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank3]: Traceback (most recent call last):
+[rank3]: File "", line 198, in _run_module_as_main
+[rank3]: File "", line 88, in _run_code
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank3]: main(config)
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank3]: return f(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank3]: output = model(
+[rank3]: ^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank3]: return self._call_impl(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank3]: return inner()
+[rank3]: ^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank3]: result = forward_call(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank3]: return func(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank3]: outputs = self.backbone(
+[rank3]: ^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank3]: return self._call_impl(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank3]: return forward_call(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank3]: hidden_states = mixer_block(
+[rank3]: ^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank3]: return self._call_impl(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank3]: return inner()
+[rank3]: ^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank3]: result = forward_call(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank3]: return fn(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank3]: return self._call_impl(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank3]: return forward_call(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank3]: hidden_states = self.norm(hidden_states)
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank3]: hidden_states = self.mixer(
+[rank3]: ^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank3]: return self._call_impl(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank3]: return forward_call(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 601, in forward
+[rank3]: return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 528, in torch_forward
+[rank3]: G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]
+[rank3]: ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~
+[rank3]: torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 3 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003902 has 316.00 MiB memory in use. Process 696032 has 316.00 MiB memory in use. Process 1850007 has 21.15 GiB memory in use. Process 2711978 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/3/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/3/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/4/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/4/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..aecad2e8dadc343a1aed8f5bffb0349cb9274e5e
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/4/error.json
@@ -0,0 +1 @@
+{"message": {"message": "OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 4 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003903 has 316.00 MiB memory in use. Process 696034 has 316.00 MiB memory in use. Process 1850008 has 21.15 GiB memory in use. Process 2711979 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 601, in forward\n return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 528, in torch_forward\n G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]\n ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~\ntorch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 4 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003903 has 316.00 MiB memory in use. Process 696034 has 316.00 MiB memory in use. Process 1850008 has 21.15 GiB memory in use. Process 2711979 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)\n", "timestamp": "1753242220"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/4/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/4/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..f18564a14973f90c63d31bcb96a7fdd1259bd4ce
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/4/stderr.log
@@ -0,0 +1,387 @@
+[titan] 2025-07-23 11:13:04,821 - root - INFO - Starting job: default job
+[titan] 2025-07-23 11:13:04,821 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 11:13:04,822 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 11:13:06,869 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 11:13:06,872 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 11:13:06,915 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 11:13:06,915 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 11:13:06,916 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 11:13:07,194 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 11:13:07,703 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 11:13:07,704 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:07,704 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:08,426 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 11:13:08,426 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:13:08,426 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:08,427 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:38,105 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:38,138 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 11:35:38,138 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:35:38,139 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:38,139 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:07,589 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:08,013 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 11:36:08,013 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:36:08,014 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:08,014 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:40:36,111 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 11:40:36,821 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 11:40:36,944 - root - INFO - Building dataloader...
+[titan] 2025-07-23 11:40:36,947 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 11:40:36,948 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 11:40:39,687 - fla.layers.mamba2 - WARNING - The fast path is not available because one of `(selective_state_update)` is None. Falling back to the naive implementation. To install follow https://github.com/state-spaces/mamba/#installation
+[titan] 2025-07-23 11:40:39,688 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 11:40:39,805 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 11:40:39,858 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 11:40:39,858 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 11:40:39,858 - root - WARNING - No norm found in model
+[titan] 2025-07-23 11:40:39,859 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 11:40:40,074 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 11:40:40,428 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 11:40:40,596 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 11:40:40,838 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 11:40:40,840 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 11:40:40,866 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 11:40:40,866 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 11:40:41,118 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 11:40:48,318 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 11:40:48,320 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 11:40:48,321 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 11:40:48,321 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 11:40:48,322 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 11:40:48,322 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 11:40:48,322 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 11:40:48,322 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 11:40:48,322 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 11:40:48,322 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank4]: Traceback (most recent call last):
+[rank4]: File "", line 198, in _run_module_as_main
+[rank4]: File "", line 88, in _run_code
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank4]: main(config)
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank4]: return f(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank4]: output = model(
+[rank4]: ^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank4]: return self._call_impl(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank4]: return inner()
+[rank4]: ^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank4]: result = forward_call(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank4]: return func(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank4]: outputs = self.backbone(
+[rank4]: ^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank4]: return self._call_impl(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank4]: return forward_call(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank4]: hidden_states = mixer_block(
+[rank4]: ^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank4]: return self._call_impl(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank4]: return inner()
+[rank4]: ^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank4]: result = forward_call(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank4]: return fn(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank4]: return self._call_impl(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank4]: return forward_call(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank4]: hidden_states = self.norm(hidden_states)
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank4]: hidden_states = self.mixer(
+[rank4]: ^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank4]: return self._call_impl(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank4]: return forward_call(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 601, in forward
+[rank4]: return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 528, in torch_forward
+[rank4]: G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]
+[rank4]: ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~
+[rank4]: torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 4 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003903 has 316.00 MiB memory in use. Process 696034 has 316.00 MiB memory in use. Process 1850008 has 21.15 GiB memory in use. Process 2711979 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/4/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/4/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/5/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/5/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..ea0bc49bf0dc04be2af5b6a10bab6b363ff66973
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/5/error.json
@@ -0,0 +1 @@
+{"message": {"message": "OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 5 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003905 has 316.00 MiB memory in use. Process 696036 has 316.00 MiB memory in use. Process 1850009 has 21.15 GiB memory in use. Process 2711980 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 601, in forward\n return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 528, in torch_forward\n G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]\n ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~\ntorch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 5 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003905 has 316.00 MiB memory in use. Process 696036 has 316.00 MiB memory in use. Process 1850009 has 21.15 GiB memory in use. Process 2711980 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)\n", "timestamp": "1753242220"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/5/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/5/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..d1c58bf3952bc4702013d45cb3cd60ced2f5c6ab
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/5/stderr.log
@@ -0,0 +1,387 @@
+[titan] 2025-07-23 11:13:05,018 - root - INFO - Starting job: default job
+[titan] 2025-07-23 11:13:05,018 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 11:13:05,019 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 11:13:07,338 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 11:13:07,341 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 11:13:07,398 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 11:13:07,398 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 11:13:07,398 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 11:13:07,408 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 11:13:07,703 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 11:13:07,704 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:07,704 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:08,425 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 11:13:08,426 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:13:08,426 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:08,426 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:40,853 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:40,887 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 11:35:40,887 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:35:40,887 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:40,887 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:07,598 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:08,012 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 11:36:08,013 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:36:08,013 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:08,013 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:40:36,193 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 11:40:36,903 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 11:40:37,027 - root - INFO - Building dataloader...
+[titan] 2025-07-23 11:40:37,029 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 11:40:37,031 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 11:40:39,687 - fla.layers.mamba2 - WARNING - The fast path is not available because one of `(selective_state_update)` is None. Falling back to the naive implementation. To install follow https://github.com/state-spaces/mamba/#installation
+[titan] 2025-07-23 11:40:39,687 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 11:40:39,807 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 11:40:39,859 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 11:40:39,860 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 11:40:39,860 - root - WARNING - No norm found in model
+[titan] 2025-07-23 11:40:39,860 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 11:40:40,073 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 11:40:40,434 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 11:40:40,596 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 11:40:40,838 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 11:40:40,840 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 11:40:40,865 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 11:40:40,866 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 11:40:41,092 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 11:40:48,858 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 11:40:48,867 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 11:40:48,877 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 11:40:48,877 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 11:40:48,877 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 11:40:48,878 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 11:40:48,878 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 11:40:48,878 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 11:40:48,878 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 11:40:48,879 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank5]: Traceback (most recent call last):
+[rank5]: File "", line 198, in _run_module_as_main
+[rank5]: File "", line 88, in _run_code
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank5]: main(config)
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank5]: return f(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank5]: output = model(
+[rank5]: ^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank5]: return self._call_impl(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank5]: return inner()
+[rank5]: ^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank5]: result = forward_call(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank5]: return func(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank5]: outputs = self.backbone(
+[rank5]: ^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank5]: return self._call_impl(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank5]: return forward_call(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank5]: hidden_states = mixer_block(
+[rank5]: ^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank5]: return self._call_impl(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank5]: return inner()
+[rank5]: ^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank5]: result = forward_call(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank5]: return fn(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank5]: return self._call_impl(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank5]: return forward_call(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank5]: hidden_states = self.norm(hidden_states)
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank5]: hidden_states = self.mixer(
+[rank5]: ^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank5]: return self._call_impl(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank5]: return forward_call(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 601, in forward
+[rank5]: return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 528, in torch_forward
+[rank5]: G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]
+[rank5]: ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~
+[rank5]: torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 5 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003905 has 316.00 MiB memory in use. Process 696036 has 316.00 MiB memory in use. Process 1850009 has 21.15 GiB memory in use. Process 2711980 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/5/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/5/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/6/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/6/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..976ec72228dc4c2b7a8fa0af2109f374cd4ecad1
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/6/error.json
@@ -0,0 +1 @@
+{"message": {"message": "OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 6 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003904 has 316.00 MiB memory in use. Process 696035 has 316.00 MiB memory in use. Process 1850010 has 21.15 GiB memory in use. Process 2711981 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 601, in forward\n return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 528, in torch_forward\n G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]\n ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~\ntorch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 6 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003904 has 316.00 MiB memory in use. Process 696035 has 316.00 MiB memory in use. Process 1850010 has 21.15 GiB memory in use. Process 2711981 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)\n", "timestamp": "1753242216"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/6/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/6/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..17f3e8b6d063aaabf25d19aa2c9145b5a6ddc59a
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/6/stderr.log
@@ -0,0 +1,387 @@
+[titan] 2025-07-23 11:13:05,027 - root - INFO - Starting job: default job
+[titan] 2025-07-23 11:13:05,027 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 11:13:05,028 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 11:13:07,378 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 11:13:07,380 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 11:13:07,459 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 11:13:07,459 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 11:13:07,459 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 11:13:07,469 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 11:13:07,703 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 11:13:07,704 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:07,705 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:08,425 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 11:13:08,426 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:13:08,426 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:08,427 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:51,159 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:51,191 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 11:35:51,191 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:35:51,192 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:35:51,192 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:07,587 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:08,013 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 11:36:08,013 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:36:08,013 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:08,013 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:40:36,055 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 11:40:36,768 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 11:40:36,892 - root - INFO - Building dataloader...
+[titan] 2025-07-23 11:40:36,895 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 11:40:36,897 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 11:40:39,687 - fla.layers.mamba2 - WARNING - The fast path is not available because one of `(selective_state_update)` is None. Falling back to the naive implementation. To install follow https://github.com/state-spaces/mamba/#installation
+[titan] 2025-07-23 11:40:39,687 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 11:40:39,806 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 11:40:39,857 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 11:40:39,858 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 11:40:39,858 - root - WARNING - No norm found in model
+[titan] 2025-07-23 11:40:39,858 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 11:40:40,073 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 11:40:40,431 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 11:40:40,596 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 11:40:40,839 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 11:40:40,841 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 11:40:40,867 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 11:40:40,867 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 11:40:41,110 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 11:40:48,363 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 11:40:48,365 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 11:40:48,366 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 11:40:48,366 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 11:40:48,366 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 11:40:48,366 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 11:40:48,366 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 11:40:48,366 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 11:40:48,366 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 11:40:48,366 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank6]: Traceback (most recent call last):
+[rank6]: File "", line 198, in _run_module_as_main
+[rank6]: File "", line 88, in _run_code
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank6]: main(config)
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank6]: return f(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank6]: output = model(
+[rank6]: ^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank6]: return self._call_impl(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank6]: return inner()
+[rank6]: ^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank6]: result = forward_call(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank6]: return func(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank6]: outputs = self.backbone(
+[rank6]: ^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank6]: return self._call_impl(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank6]: return forward_call(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank6]: hidden_states = mixer_block(
+[rank6]: ^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank6]: return self._call_impl(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank6]: return inner()
+[rank6]: ^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank6]: result = forward_call(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank6]: return fn(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank6]: return self._call_impl(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank6]: return forward_call(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank6]: hidden_states = self.norm(hidden_states)
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank6]: hidden_states = self.mixer(
+[rank6]: ^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank6]: return self._call_impl(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank6]: return forward_call(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 601, in forward
+[rank6]: return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 528, in torch_forward
+[rank6]: G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]
+[rank6]: ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~
+[rank6]: torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 6 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003904 has 316.00 MiB memory in use. Process 696035 has 316.00 MiB memory in use. Process 1850010 has 21.15 GiB memory in use. Process 2711981 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/6/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/6/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/7/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/7/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..71f95deb8b75bd56758fa07dc9517d4490a49c32
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/7/error.json
@@ -0,0 +1 @@
+{"message": {"message": "OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 7 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003906 has 316.00 MiB memory in use. Process 696037 has 316.00 MiB memory in use. Process 1850011 has 21.15 GiB memory in use. Process 2711982 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 601, in forward\n return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 528, in torch_forward\n G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]\n ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~\ntorch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 7 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003906 has 316.00 MiB memory in use. Process 696037 has 316.00 MiB memory in use. Process 1850011 has 21.15 GiB memory in use. Process 2711982 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)\n", "timestamp": "1753242219"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/7/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/7/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..84e76aa051ebf02eb83b7be7a5e294777ff338c7
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/7/stderr.log
@@ -0,0 +1,387 @@
+[titan] 2025-07-23 11:13:04,946 - root - INFO - Starting job: default job
+[titan] 2025-07-23 11:13:04,947 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 11:13:04,947 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 11:13:07,188 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 11:13:07,191 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 11:13:07,236 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 11:13:07,236 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 11:13:07,236 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 11:13:07,305 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 11:13:07,703 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 11:13:07,704 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:07,704 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:08,425 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 11:13:08,426 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:13:08,426 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:13:08,426 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:07,808 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:07,840 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 11:36:07,840 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:36:07,840 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:07,840 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:08,113 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:08,197 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 11:36:08,198 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 11:36:08,198 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:36:08,198 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 11:40:36,585 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 11:40:37,348 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 11:40:37,522 - root - INFO - Building dataloader...
+[titan] 2025-07-23 11:40:37,524 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 11:40:37,526 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 11:40:39,687 - fla.layers.mamba2 - WARNING - The fast path is not available because one of `(selective_state_update)` is None. Falling back to the naive implementation. To install follow https://github.com/state-spaces/mamba/#installation
+[titan] 2025-07-23 11:40:39,687 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 11:40:39,804 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 11:40:39,856 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 11:40:39,856 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 11:40:39,857 - root - WARNING - No norm found in model
+[titan] 2025-07-23 11:40:39,857 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 11:40:40,109 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 11:40:40,429 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 11:40:40,596 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 11:40:40,838 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 11:40:40,840 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 11:40:40,865 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 11:40:40,866 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 11:40:41,123 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 11:40:50,425 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 11:40:50,428 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 11:40:50,429 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 11:40:50,431 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 11:40:50,431 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 11:40:50,431 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 11:40:50,431 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 11:40:50,432 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 11:40:50,432 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 11:40:50,432 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank7]: Traceback (most recent call last):
+[rank7]: File "", line 198, in _run_module_as_main
+[rank7]: File "", line 88, in _run_code
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank7]: main(config)
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank7]: return f(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank7]: output = model(
+[rank7]: ^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank7]: return self._call_impl(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank7]: return inner()
+[rank7]: ^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank7]: result = forward_call(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank7]: return func(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank7]: outputs = self.backbone(
+[rank7]: ^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank7]: return self._call_impl(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank7]: return forward_call(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank7]: hidden_states = mixer_block(
+[rank7]: ^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank7]: return self._call_impl(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank7]: return inner()
+[rank7]: ^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank7]: result = forward_call(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank7]: return fn(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank7]: return self._call_impl(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank7]: return forward_call(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank7]: hidden_states = self.norm(hidden_states)
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank7]: hidden_states = self.mixer(
+[rank7]: ^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank7]: return self._call_impl(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank7]: return forward_call(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 601, in forward
+[rank7]: return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 528, in torch_forward
+[rank7]: G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, :, :]
+[rank7]: ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~
+[rank7]: torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 GiB. GPU 7 has a total capacity of 95.00 GiB of which 63.94 GiB is free. Process 2003906 has 316.00 MiB memory in use. Process 696037 has 316.00 MiB memory in use. Process 1850011 has 21.15 GiB memory in use. Process 2711982 has 9.27 GiB memory in use. Of the allocated memory 7.99 GiB is allocated by PyTorch, and 73.66 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/7/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_77qh1j5t/attempt_0/7/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/0/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/0/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..2c7a9259aa9b367776fd2116bc21e4177467feb8
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/0/error.json
@@ -0,0 +1 @@
+{"message": {"message": "TypeError: 'NoneType' object is not callable", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 595, in forward\n return self.cuda_kernels_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 300, in cuda_kernels_forward\n out = mamba_split_conv1d_scan_combined(\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py\", line 947, in mamba_split_conv1d_scan_combined\n return MambaSplitConv1dScanCombinedFn.apply(zxbcdt, conv1d_weight, conv1d_bias, dt_bias, A, D, chunk_size, initial_states, seq_idx, dt_limit, return_final_states, activation, rmsnorm_weight, rmsnorm_eps, outproj_weight, outproj_bias, headdim, ngroups, norm_before_gate)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/autograd/function.py\", line 575, in apply\n return super().apply(*args, **kwargs) # type: ignore[misc]\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/amp/autocast_mode.py\", line 510, in decorate_fwd\n return fwd(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py\", line 792, in forward\n causal_conv1d_fwd_function(rearrange_and_update_stride(xBC, \"b s d -> b d s\"),\nTypeError: 'NoneType' object is not callable\n", "timestamp": "1753254643"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/0/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/0/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..fb8ea6f8a4c546263bcb48d30ec84685a63bb7a5
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/0/stderr.log
@@ -0,0 +1,484 @@
+[titan] 2025-07-23 15:07:39,311 - root - INFO - Starting job: default job
+[titan] 2025-07-23 15:07:39,311 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 15:07:39,313 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 15:07:39,313 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 15:07:39,321 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 15:07:39,408 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 15:07:39,409 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 15:07:39,409 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 15:07:39,658 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 15:07:39,763 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 15:07:39,763 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:39,763 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:39,882 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 15:07:39,883 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:07:39,883 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:39,883 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:34,205 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:34,238 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 15:08:34,238 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:08:34,238 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:34,238 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:34,503 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:34,591 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 15:08:34,591 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:08:34,591 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:34,591 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:41,198 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 15:08:41,974 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 15:08:42,171 - root - INFO - Building dataloader...
+[titan] 2025-07-23 15:08:42,174 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 15:08:42,175 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 15:08:42,388 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 15:08:42,487 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 15:08:42,539 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 15:08:42,540 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 15:08:42,540 - root - WARNING - No norm found in model
+[titan] 2025-07-23 15:08:42,540 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 15:08:42,624 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 15:08:42,697 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 15:08:42,698 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 15:08:42,747 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 15:08:42,749 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 15:08:42,773 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+wandb: Network error (InvalidURL), entering retry loop.
+wandb: W&B API key is configured. Use `wandb login --relogin` to force relogin
+wandb: Network error (InvalidURL), entering retry loop.
+[titan] 2025-07-23 15:10:24,327 - root - ERROR - Failed to create WandB logger: Run initialization has timed out after 90.0 sec. Please try increasing the timeout with the `init_timeout` setting: `wandb.init(settings=wandb.Settings(init_timeout=120))`.
+[titan] 2025-07-23 15:10:24,334 - root - INFO - TensorBoard logging enabled. Logs will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1508
+[titan] 2025-07-23 15:10:24,334 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 15:10:24,386 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 15:10:31,721 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 15:10:31,725 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 15:10:31,725 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 15:10:31,725 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 15:10:31,725 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 15:10:31,725 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 15:10:31,726 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 15:10:31,726 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 15:10:31,726 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 15:10:31,726 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+Traceback (most recent call last):
+ File "", line 198, in _run_module_as_main
+ File "", line 88, in _run_code
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+ main(config)
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+ return f(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+ output = model(
+ ^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+ return self._call_impl(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+ return inner()
+ ^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+ result = forward_call(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+ return func(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+ outputs = self.backbone(
+ ^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+ return self._call_impl(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+ return forward_call(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+ hidden_states = mixer_block(
+ ^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+ return self._call_impl(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+ return inner()
+ ^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+ result = forward_call(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+ return fn(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+ return self._call_impl(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+ return forward_call(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+ hidden_states = self.norm(hidden_states)
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+ hidden_states = self.mixer(
+ ^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+ return self._call_impl(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+ return forward_call(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 595, in forward
+ return self.cuda_kernels_forward(hidden_states, cache_params, cache_position, attention_mask)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 300, in cuda_kernels_forward
+ out = mamba_split_conv1d_scan_combined(
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py", line 947, in mamba_split_conv1d_scan_combined
+ return MambaSplitConv1dScanCombinedFn.apply(zxbcdt, conv1d_weight, conv1d_bias, dt_bias, A, D, chunk_size, initial_states, seq_idx, dt_limit, return_final_states, activation, rmsnorm_weight, rmsnorm_eps, outproj_weight, outproj_bias, headdim, ngroups, norm_before_gate)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/autograd/function.py", line 575, in apply
+ return super().apply(*args, **kwargs) # type: ignore[misc]
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/amp/autocast_mode.py", line 510, in decorate_fwd
+ return fwd(*args, **kwargs)
+ ^^^^^^^^^^^^^^^^^^^^
+ File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py", line 792, in forward
+ causal_conv1d_fwd_function(rearrange_and_update_stride(xBC, "b s d -> b d s"),
+TypeError: 'NoneType' object is not callable
+[rank0]: Traceback (most recent call last):
+[rank0]: File "", line 198, in _run_module_as_main
+[rank0]: File "", line 88, in _run_code
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank0]: main(config)
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank0]: return f(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank0]: output = model(
+[rank0]: ^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank0]: return self._call_impl(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank0]: return inner()
+[rank0]: ^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank0]: result = forward_call(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank0]: return func(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank0]: outputs = self.backbone(
+[rank0]: ^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank0]: return self._call_impl(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank0]: return forward_call(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank0]: hidden_states = mixer_block(
+[rank0]: ^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank0]: return self._call_impl(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank0]: return inner()
+[rank0]: ^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank0]: result = forward_call(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank0]: return fn(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank0]: return self._call_impl(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank0]: return forward_call(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank0]: hidden_states = self.norm(hidden_states)
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank0]: hidden_states = self.mixer(
+[rank0]: ^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank0]: return self._call_impl(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank0]: return forward_call(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 595, in forward
+[rank0]: return self.cuda_kernels_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 300, in cuda_kernels_forward
+[rank0]: out = mamba_split_conv1d_scan_combined(
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py", line 947, in mamba_split_conv1d_scan_combined
+[rank0]: return MambaSplitConv1dScanCombinedFn.apply(zxbcdt, conv1d_weight, conv1d_bias, dt_bias, A, D, chunk_size, initial_states, seq_idx, dt_limit, return_final_states, activation, rmsnorm_weight, rmsnorm_eps, outproj_weight, outproj_bias, headdim, ngroups, norm_before_gate)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/autograd/function.py", line 575, in apply
+[rank0]: return super().apply(*args, **kwargs) # type: ignore[misc]
+[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/amp/autocast_mode.py", line 510, in decorate_fwd
+[rank0]: return fwd(*args, **kwargs)
+[rank0]: ^^^^^^^^^^^^^^^^^^^^
+[rank0]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py", line 792, in forward
+[rank0]: causal_conv1d_fwd_function(rearrange_and_update_stride(xBC, "b s d -> b d s"),
+[rank0]: TypeError: 'NoneType' object is not callable
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/0/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/0/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/1/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/1/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..2c7a9259aa9b367776fd2116bc21e4177467feb8
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/1/error.json
@@ -0,0 +1 @@
+{"message": {"message": "TypeError: 'NoneType' object is not callable", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 595, in forward\n return self.cuda_kernels_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 300, in cuda_kernels_forward\n out = mamba_split_conv1d_scan_combined(\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py\", line 947, in mamba_split_conv1d_scan_combined\n return MambaSplitConv1dScanCombinedFn.apply(zxbcdt, conv1d_weight, conv1d_bias, dt_bias, A, D, chunk_size, initial_states, seq_idx, dt_limit, return_final_states, activation, rmsnorm_weight, rmsnorm_eps, outproj_weight, outproj_bias, headdim, ngroups, norm_before_gate)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/autograd/function.py\", line 575, in apply\n return super().apply(*args, **kwargs) # type: ignore[misc]\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/amp/autocast_mode.py\", line 510, in decorate_fwd\n return fwd(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py\", line 792, in forward\n causal_conv1d_fwd_function(rearrange_and_update_stride(xBC, \"b s d -> b d s\"),\nTypeError: 'NoneType' object is not callable\n", "timestamp": "1753254643"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/1/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/1/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..68792ec1ab6054ccbac7e67743d6ee4e943f9adf
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/1/stderr.log
@@ -0,0 +1,397 @@
+[titan] 2025-07-23 15:07:39,579 - root - INFO - Starting job: default job
+[titan] 2025-07-23 15:07:39,579 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 15:07:39,580 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 15:07:39,811 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 15:07:39,814 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 15:07:39,870 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 15:07:39,870 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 15:07:39,870 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 15:07:39,878 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 15:07:39,987 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 15:07:39,988 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:39,988 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:40,116 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 15:07:40,116 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:07:40,116 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:40,116 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:32,327 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:32,359 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 15:08:32,359 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:08:32,359 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:32,359 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:32,618 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:32,704 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 15:08:32,704 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:08:32,704 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:32,705 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:39,069 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 15:08:39,771 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 15:08:39,891 - root - INFO - Building dataloader...
+[titan] 2025-07-23 15:08:39,893 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 15:08:39,895 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 15:08:40,145 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 15:08:40,240 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 15:08:40,292 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 15:08:40,293 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 15:08:40,293 - root - WARNING - No norm found in model
+[titan] 2025-07-23 15:08:40,293 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 15:08:40,374 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 15:08:40,440 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 15:08:40,441 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 15:08:40,490 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 15:08:40,492 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 15:08:40,516 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 15:08:40,521 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 15:08:40,572 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 15:08:47,124 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 15:08:47,126 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 15:08:47,127 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 15:08:47,128 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 15:08:47,128 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 15:08:47,128 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 15:08:47,128 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 15:08:47,128 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 15:08:47,128 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 15:08:47,128 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank1]: Traceback (most recent call last):
+[rank1]: File "", line 198, in _run_module_as_main
+[rank1]: File "", line 88, in _run_code
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank1]: main(config)
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank1]: return f(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank1]: output = model(
+[rank1]: ^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank1]: return self._call_impl(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank1]: return inner()
+[rank1]: ^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank1]: result = forward_call(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank1]: return func(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank1]: outputs = self.backbone(
+[rank1]: ^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank1]: return self._call_impl(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank1]: return forward_call(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank1]: hidden_states = mixer_block(
+[rank1]: ^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank1]: return self._call_impl(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank1]: return inner()
+[rank1]: ^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank1]: result = forward_call(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank1]: return fn(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank1]: return self._call_impl(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank1]: return forward_call(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank1]: hidden_states = self.norm(hidden_states)
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank1]: hidden_states = self.mixer(
+[rank1]: ^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank1]: return self._call_impl(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank1]: return forward_call(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 595, in forward
+[rank1]: return self.cuda_kernels_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 300, in cuda_kernels_forward
+[rank1]: out = mamba_split_conv1d_scan_combined(
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py", line 947, in mamba_split_conv1d_scan_combined
+[rank1]: return MambaSplitConv1dScanCombinedFn.apply(zxbcdt, conv1d_weight, conv1d_bias, dt_bias, A, D, chunk_size, initial_states, seq_idx, dt_limit, return_final_states, activation, rmsnorm_weight, rmsnorm_eps, outproj_weight, outproj_bias, headdim, ngroups, norm_before_gate)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/autograd/function.py", line 575, in apply
+[rank1]: return super().apply(*args, **kwargs) # type: ignore[misc]
+[rank1]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/amp/autocast_mode.py", line 510, in decorate_fwd
+[rank1]: return fwd(*args, **kwargs)
+[rank1]: ^^^^^^^^^^^^^^^^^^^^
+[rank1]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py", line 792, in forward
+[rank1]: causal_conv1d_fwd_function(rearrange_and_update_stride(xBC, "b s d -> b d s"),
+[rank1]: TypeError: 'NoneType' object is not callable
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/1/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/1/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/2/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/2/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..2c7a9259aa9b367776fd2116bc21e4177467feb8
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/2/error.json
@@ -0,0 +1 @@
+{"message": {"message": "TypeError: 'NoneType' object is not callable", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 595, in forward\n return self.cuda_kernels_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 300, in cuda_kernels_forward\n out = mamba_split_conv1d_scan_combined(\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py\", line 947, in mamba_split_conv1d_scan_combined\n return MambaSplitConv1dScanCombinedFn.apply(zxbcdt, conv1d_weight, conv1d_bias, dt_bias, A, D, chunk_size, initial_states, seq_idx, dt_limit, return_final_states, activation, rmsnorm_weight, rmsnorm_eps, outproj_weight, outproj_bias, headdim, ngroups, norm_before_gate)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/autograd/function.py\", line 575, in apply\n return super().apply(*args, **kwargs) # type: ignore[misc]\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/amp/autocast_mode.py\", line 510, in decorate_fwd\n return fwd(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py\", line 792, in forward\n causal_conv1d_fwd_function(rearrange_and_update_stride(xBC, \"b s d -> b d s\"),\nTypeError: 'NoneType' object is not callable\n", "timestamp": "1753254643"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/2/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/2/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..ca5349b18817ef8459168af71ed8244eeb1a24bc
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/2/stderr.log
@@ -0,0 +1,397 @@
+[titan] 2025-07-23 15:07:43,124 - root - INFO - Starting job: default job
+[titan] 2025-07-23 15:07:43,124 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 15:07:43,125 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 15:07:43,401 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 15:07:43,403 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 15:07:43,459 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 15:07:43,460 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 15:07:43,460 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 15:07:43,533 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 15:07:43,640 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 15:07:43,640 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:43,640 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:43,753 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 15:07:43,753 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:07:43,753 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:43,754 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:34,285 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:34,317 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 15:08:34,317 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:08:34,317 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:34,317 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:34,582 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:34,672 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 15:08:34,672 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:08:34,672 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:34,673 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:41,985 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 15:08:42,774 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 15:08:42,989 - root - INFO - Building dataloader...
+[titan] 2025-07-23 15:08:42,991 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 15:08:42,993 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 15:08:43,215 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 15:08:43,312 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 15:08:43,364 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 15:08:43,364 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 15:08:43,364 - root - WARNING - No norm found in model
+[titan] 2025-07-23 15:08:43,365 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 15:08:43,446 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 15:08:43,511 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 15:08:43,512 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 15:08:43,562 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 15:08:43,564 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 15:08:43,588 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 15:08:43,593 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 15:08:43,646 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 15:08:53,279 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 15:08:53,283 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 15:08:53,283 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 15:08:53,283 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 15:08:53,284 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 15:08:53,284 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 15:08:53,284 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 15:08:53,284 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 15:08:53,284 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 15:08:53,284 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank2]: Traceback (most recent call last):
+[rank2]: File "", line 198, in _run_module_as_main
+[rank2]: File "", line 88, in _run_code
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank2]: main(config)
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank2]: return f(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank2]: output = model(
+[rank2]: ^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank2]: return self._call_impl(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank2]: return inner()
+[rank2]: ^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank2]: result = forward_call(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank2]: return func(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank2]: outputs = self.backbone(
+[rank2]: ^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank2]: return self._call_impl(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank2]: return forward_call(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank2]: hidden_states = mixer_block(
+[rank2]: ^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank2]: return self._call_impl(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank2]: return inner()
+[rank2]: ^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank2]: result = forward_call(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank2]: return fn(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank2]: return self._call_impl(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank2]: return forward_call(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank2]: hidden_states = self.norm(hidden_states)
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank2]: hidden_states = self.mixer(
+[rank2]: ^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank2]: return self._call_impl(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank2]: return forward_call(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 595, in forward
+[rank2]: return self.cuda_kernels_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 300, in cuda_kernels_forward
+[rank2]: out = mamba_split_conv1d_scan_combined(
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py", line 947, in mamba_split_conv1d_scan_combined
+[rank2]: return MambaSplitConv1dScanCombinedFn.apply(zxbcdt, conv1d_weight, conv1d_bias, dt_bias, A, D, chunk_size, initial_states, seq_idx, dt_limit, return_final_states, activation, rmsnorm_weight, rmsnorm_eps, outproj_weight, outproj_bias, headdim, ngroups, norm_before_gate)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/autograd/function.py", line 575, in apply
+[rank2]: return super().apply(*args, **kwargs) # type: ignore[misc]
+[rank2]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/amp/autocast_mode.py", line 510, in decorate_fwd
+[rank2]: return fwd(*args, **kwargs)
+[rank2]: ^^^^^^^^^^^^^^^^^^^^
+[rank2]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py", line 792, in forward
+[rank2]: causal_conv1d_fwd_function(rearrange_and_update_stride(xBC, "b s d -> b d s"),
+[rank2]: TypeError: 'NoneType' object is not callable
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/2/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/2/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/3/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/3/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..2c7a9259aa9b367776fd2116bc21e4177467feb8
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/3/error.json
@@ -0,0 +1 @@
+{"message": {"message": "TypeError: 'NoneType' object is not callable", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 595, in forward\n return self.cuda_kernels_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 300, in cuda_kernels_forward\n out = mamba_split_conv1d_scan_combined(\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py\", line 947, in mamba_split_conv1d_scan_combined\n return MambaSplitConv1dScanCombinedFn.apply(zxbcdt, conv1d_weight, conv1d_bias, dt_bias, A, D, chunk_size, initial_states, seq_idx, dt_limit, return_final_states, activation, rmsnorm_weight, rmsnorm_eps, outproj_weight, outproj_bias, headdim, ngroups, norm_before_gate)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/autograd/function.py\", line 575, in apply\n return super().apply(*args, **kwargs) # type: ignore[misc]\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/amp/autocast_mode.py\", line 510, in decorate_fwd\n return fwd(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py\", line 792, in forward\n causal_conv1d_fwd_function(rearrange_and_update_stride(xBC, \"b s d -> b d s\"),\nTypeError: 'NoneType' object is not callable\n", "timestamp": "1753254643"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/3/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/3/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..169280b727ecf283e7a2a17e7485cdda4b3575f1
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/3/stderr.log
@@ -0,0 +1,397 @@
+[titan] 2025-07-23 15:07:44,425 - root - INFO - Starting job: default job
+[titan] 2025-07-23 15:07:44,426 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 15:07:44,427 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 15:07:45,122 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 15:07:45,126 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 15:07:45,178 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 15:07:45,178 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 15:07:45,178 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 15:07:45,192 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 15:07:45,295 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 15:07:45,295 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:45,295 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:45,406 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 15:07:45,406 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:07:45,406 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:45,406 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:36,749 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:36,782 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 15:08:36,782 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:08:36,782 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:36,782 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:37,042 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:37,126 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 15:08:37,127 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:08:37,127 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:37,127 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:43,444 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 15:08:44,150 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 15:08:44,271 - root - INFO - Building dataloader...
+[titan] 2025-07-23 15:08:44,274 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 15:08:44,275 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 15:08:44,487 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 15:08:44,583 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 15:08:44,636 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 15:08:44,636 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 15:08:44,636 - root - WARNING - No norm found in model
+[titan] 2025-07-23 15:08:44,637 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 15:08:44,717 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 15:08:44,783 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 15:08:44,783 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 15:08:44,833 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 15:08:44,835 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 15:08:44,859 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 15:08:44,864 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 15:08:44,915 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 15:08:52,281 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 15:08:52,284 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 15:08:52,285 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 15:08:52,285 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 15:08:52,285 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 15:08:52,285 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 15:08:52,285 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 15:08:52,285 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 15:08:52,285 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 15:08:52,286 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank3]: Traceback (most recent call last):
+[rank3]: File "", line 198, in _run_module_as_main
+[rank3]: File "", line 88, in _run_code
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank3]: main(config)
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank3]: return f(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank3]: output = model(
+[rank3]: ^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank3]: return self._call_impl(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank3]: return inner()
+[rank3]: ^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank3]: result = forward_call(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank3]: return func(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank3]: outputs = self.backbone(
+[rank3]: ^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank3]: return self._call_impl(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank3]: return forward_call(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank3]: hidden_states = mixer_block(
+[rank3]: ^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank3]: return self._call_impl(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank3]: return inner()
+[rank3]: ^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank3]: result = forward_call(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank3]: return fn(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank3]: return self._call_impl(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank3]: return forward_call(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank3]: hidden_states = self.norm(hidden_states)
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank3]: hidden_states = self.mixer(
+[rank3]: ^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank3]: return self._call_impl(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank3]: return forward_call(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 595, in forward
+[rank3]: return self.cuda_kernels_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 300, in cuda_kernels_forward
+[rank3]: out = mamba_split_conv1d_scan_combined(
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py", line 947, in mamba_split_conv1d_scan_combined
+[rank3]: return MambaSplitConv1dScanCombinedFn.apply(zxbcdt, conv1d_weight, conv1d_bias, dt_bias, A, D, chunk_size, initial_states, seq_idx, dt_limit, return_final_states, activation, rmsnorm_weight, rmsnorm_eps, outproj_weight, outproj_bias, headdim, ngroups, norm_before_gate)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/autograd/function.py", line 575, in apply
+[rank3]: return super().apply(*args, **kwargs) # type: ignore[misc]
+[rank3]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/amp/autocast_mode.py", line 510, in decorate_fwd
+[rank3]: return fwd(*args, **kwargs)
+[rank3]: ^^^^^^^^^^^^^^^^^^^^
+[rank3]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py", line 792, in forward
+[rank3]: causal_conv1d_fwd_function(rearrange_and_update_stride(xBC, "b s d -> b d s"),
+[rank3]: TypeError: 'NoneType' object is not callable
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/3/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/3/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/4/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/4/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..2c7a9259aa9b367776fd2116bc21e4177467feb8
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/4/error.json
@@ -0,0 +1 @@
+{"message": {"message": "TypeError: 'NoneType' object is not callable", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 595, in forward\n return self.cuda_kernels_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 300, in cuda_kernels_forward\n out = mamba_split_conv1d_scan_combined(\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py\", line 947, in mamba_split_conv1d_scan_combined\n return MambaSplitConv1dScanCombinedFn.apply(zxbcdt, conv1d_weight, conv1d_bias, dt_bias, A, D, chunk_size, initial_states, seq_idx, dt_limit, return_final_states, activation, rmsnorm_weight, rmsnorm_eps, outproj_weight, outproj_bias, headdim, ngroups, norm_before_gate)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/autograd/function.py\", line 575, in apply\n return super().apply(*args, **kwargs) # type: ignore[misc]\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/amp/autocast_mode.py\", line 510, in decorate_fwd\n return fwd(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py\", line 792, in forward\n causal_conv1d_fwd_function(rearrange_and_update_stride(xBC, \"b s d -> b d s\"),\nTypeError: 'NoneType' object is not callable\n", "timestamp": "1753254643"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/4/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/4/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..46df845126b77d92c7f195a8d64fc0a496127cf2
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/4/stderr.log
@@ -0,0 +1,397 @@
+[titan] 2025-07-23 15:07:43,853 - root - INFO - Starting job: default job
+[titan] 2025-07-23 15:07:43,853 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 15:07:43,854 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 15:07:44,480 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 15:07:44,483 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 15:07:44,560 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 15:07:44,560 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 15:07:44,560 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 15:07:44,628 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 15:07:44,732 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 15:07:44,733 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:44,733 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:44,853 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 15:07:44,853 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:07:44,853 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:44,853 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:37,683 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:37,715 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 15:08:37,715 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:08:37,715 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:37,715 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:37,973 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:38,058 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 15:08:38,058 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:08:38,059 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:38,059 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:44,571 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 15:08:45,263 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 15:08:45,382 - root - INFO - Building dataloader...
+[titan] 2025-07-23 15:08:45,384 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 15:08:45,387 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 15:08:45,627 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 15:08:45,724 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 15:08:45,777 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 15:08:45,777 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 15:08:45,777 - root - WARNING - No norm found in model
+[titan] 2025-07-23 15:08:45,778 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 15:08:45,859 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 15:08:45,924 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 15:08:45,925 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 15:08:45,975 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 15:08:45,977 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 15:08:46,001 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 15:08:46,006 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 15:08:46,053 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 15:08:53,583 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 15:08:53,612 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 15:08:53,613 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 15:08:53,614 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 15:08:53,614 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 15:08:53,614 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 15:08:53,614 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 15:08:53,614 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 15:08:53,614 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 15:08:53,614 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank4]: Traceback (most recent call last):
+[rank4]: File "", line 198, in _run_module_as_main
+[rank4]: File "", line 88, in _run_code
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank4]: main(config)
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank4]: return f(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank4]: output = model(
+[rank4]: ^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank4]: return self._call_impl(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank4]: return inner()
+[rank4]: ^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank4]: result = forward_call(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank4]: return func(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank4]: outputs = self.backbone(
+[rank4]: ^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank4]: return self._call_impl(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank4]: return forward_call(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank4]: hidden_states = mixer_block(
+[rank4]: ^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank4]: return self._call_impl(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank4]: return inner()
+[rank4]: ^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank4]: result = forward_call(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank4]: return fn(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank4]: return self._call_impl(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank4]: return forward_call(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank4]: hidden_states = self.norm(hidden_states)
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank4]: hidden_states = self.mixer(
+[rank4]: ^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank4]: return self._call_impl(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank4]: return forward_call(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 595, in forward
+[rank4]: return self.cuda_kernels_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 300, in cuda_kernels_forward
+[rank4]: out = mamba_split_conv1d_scan_combined(
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py", line 947, in mamba_split_conv1d_scan_combined
+[rank4]: return MambaSplitConv1dScanCombinedFn.apply(zxbcdt, conv1d_weight, conv1d_bias, dt_bias, A, D, chunk_size, initial_states, seq_idx, dt_limit, return_final_states, activation, rmsnorm_weight, rmsnorm_eps, outproj_weight, outproj_bias, headdim, ngroups, norm_before_gate)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/autograd/function.py", line 575, in apply
+[rank4]: return super().apply(*args, **kwargs) # type: ignore[misc]
+[rank4]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/amp/autocast_mode.py", line 510, in decorate_fwd
+[rank4]: return fwd(*args, **kwargs)
+[rank4]: ^^^^^^^^^^^^^^^^^^^^
+[rank4]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py", line 792, in forward
+[rank4]: causal_conv1d_fwd_function(rearrange_and_update_stride(xBC, "b s d -> b d s"),
+[rank4]: TypeError: 'NoneType' object is not callable
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/4/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/4/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/5/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/5/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..2c7a9259aa9b367776fd2116bc21e4177467feb8
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/5/error.json
@@ -0,0 +1 @@
+{"message": {"message": "TypeError: 'NoneType' object is not callable", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 595, in forward\n return self.cuda_kernels_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 300, in cuda_kernels_forward\n out = mamba_split_conv1d_scan_combined(\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py\", line 947, in mamba_split_conv1d_scan_combined\n return MambaSplitConv1dScanCombinedFn.apply(zxbcdt, conv1d_weight, conv1d_bias, dt_bias, A, D, chunk_size, initial_states, seq_idx, dt_limit, return_final_states, activation, rmsnorm_weight, rmsnorm_eps, outproj_weight, outproj_bias, headdim, ngroups, norm_before_gate)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/autograd/function.py\", line 575, in apply\n return super().apply(*args, **kwargs) # type: ignore[misc]\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/amp/autocast_mode.py\", line 510, in decorate_fwd\n return fwd(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py\", line 792, in forward\n causal_conv1d_fwd_function(rearrange_and_update_stride(xBC, \"b s d -> b d s\"),\nTypeError: 'NoneType' object is not callable\n", "timestamp": "1753254643"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/5/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/5/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..235cb77f8fdf71bba49fb4ea5fbba076dc155050
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/5/stderr.log
@@ -0,0 +1,397 @@
+[titan] 2025-07-23 15:07:44,429 - root - INFO - Starting job: default job
+[titan] 2025-07-23 15:07:44,429 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 15:07:44,431 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 15:07:45,164 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 15:07:45,167 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 15:07:45,218 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 15:07:45,218 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 15:07:45,218 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 15:07:45,229 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 15:07:45,330 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 15:07:45,330 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:45,330 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:45,439 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 15:07:45,439 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:07:45,439 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:45,439 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:43,202 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:43,235 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 15:08:43,235 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:08:43,235 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:43,235 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:43,497 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:43,588 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 15:08:43,588 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:08:43,588 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:43,588 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:50,284 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 15:08:51,233 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 15:08:51,405 - root - INFO - Building dataloader...
+[titan] 2025-07-23 15:08:51,407 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 15:08:51,410 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 15:08:51,667 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 15:08:51,778 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 15:08:51,830 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 15:08:51,831 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 15:08:51,831 - root - WARNING - No norm found in model
+[titan] 2025-07-23 15:08:51,831 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 15:08:51,914 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 15:08:51,982 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 15:08:51,983 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 15:08:52,037 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 15:08:52,038 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 15:08:52,063 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 15:08:52,068 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 15:08:52,147 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 15:09:00,474 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 15:09:00,476 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 15:09:00,477 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 15:09:00,477 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 15:09:00,477 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 15:09:00,477 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 15:09:00,477 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 15:09:00,477 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 15:09:00,477 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 15:09:00,478 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank5]: Traceback (most recent call last):
+[rank5]: File "", line 198, in _run_module_as_main
+[rank5]: File "", line 88, in _run_code
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank5]: main(config)
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank5]: return f(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank5]: output = model(
+[rank5]: ^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank5]: return self._call_impl(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank5]: return inner()
+[rank5]: ^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank5]: result = forward_call(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank5]: return func(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank5]: outputs = self.backbone(
+[rank5]: ^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank5]: return self._call_impl(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank5]: return forward_call(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank5]: hidden_states = mixer_block(
+[rank5]: ^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank5]: return self._call_impl(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank5]: return inner()
+[rank5]: ^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank5]: result = forward_call(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank5]: return fn(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank5]: return self._call_impl(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank5]: return forward_call(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank5]: hidden_states = self.norm(hidden_states)
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank5]: hidden_states = self.mixer(
+[rank5]: ^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank5]: return self._call_impl(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank5]: return forward_call(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 595, in forward
+[rank5]: return self.cuda_kernels_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 300, in cuda_kernels_forward
+[rank5]: out = mamba_split_conv1d_scan_combined(
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py", line 947, in mamba_split_conv1d_scan_combined
+[rank5]: return MambaSplitConv1dScanCombinedFn.apply(zxbcdt, conv1d_weight, conv1d_bias, dt_bias, A, D, chunk_size, initial_states, seq_idx, dt_limit, return_final_states, activation, rmsnorm_weight, rmsnorm_eps, outproj_weight, outproj_bias, headdim, ngroups, norm_before_gate)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/autograd/function.py", line 575, in apply
+[rank5]: return super().apply(*args, **kwargs) # type: ignore[misc]
+[rank5]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/amp/autocast_mode.py", line 510, in decorate_fwd
+[rank5]: return fwd(*args, **kwargs)
+[rank5]: ^^^^^^^^^^^^^^^^^^^^
+[rank5]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py", line 792, in forward
+[rank5]: causal_conv1d_fwd_function(rearrange_and_update_stride(xBC, "b s d -> b d s"),
+[rank5]: TypeError: 'NoneType' object is not callable
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/5/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/5/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/6/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/6/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..2c7a9259aa9b367776fd2116bc21e4177467feb8
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/6/error.json
@@ -0,0 +1 @@
+{"message": {"message": "TypeError: 'NoneType' object is not callable", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 595, in forward\n return self.cuda_kernels_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 300, in cuda_kernels_forward\n out = mamba_split_conv1d_scan_combined(\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py\", line 947, in mamba_split_conv1d_scan_combined\n return MambaSplitConv1dScanCombinedFn.apply(zxbcdt, conv1d_weight, conv1d_bias, dt_bias, A, D, chunk_size, initial_states, seq_idx, dt_limit, return_final_states, activation, rmsnorm_weight, rmsnorm_eps, outproj_weight, outproj_bias, headdim, ngroups, norm_before_gate)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/autograd/function.py\", line 575, in apply\n return super().apply(*args, **kwargs) # type: ignore[misc]\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/amp/autocast_mode.py\", line 510, in decorate_fwd\n return fwd(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py\", line 792, in forward\n causal_conv1d_fwd_function(rearrange_and_update_stride(xBC, \"b s d -> b d s\"),\nTypeError: 'NoneType' object is not callable\n", "timestamp": "1753254643"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/6/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/6/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..1f15f323b124a34636b92ba559d934bb26603fb6
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/6/stderr.log
@@ -0,0 +1,397 @@
+[titan] 2025-07-23 15:07:43,875 - root - INFO - Starting job: default job
+[titan] 2025-07-23 15:07:43,875 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 15:07:43,877 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 15:07:44,487 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 15:07:44,489 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 15:07:44,558 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 15:07:44,558 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 15:07:44,558 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 15:07:44,609 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 15:07:44,712 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 15:07:44,712 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:44,712 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:44,825 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 15:07:44,826 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:07:44,826 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:44,826 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:39,360 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:39,392 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 15:08:39,392 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:08:39,392 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:39,392 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:39,651 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:39,735 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 15:08:39,736 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:08:39,736 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:39,736 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:46,050 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 15:08:46,748 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 15:08:46,868 - root - INFO - Building dataloader...
+[titan] 2025-07-23 15:08:46,870 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 15:08:46,872 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 15:08:47,083 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 15:08:47,179 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 15:08:47,231 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 15:08:47,232 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 15:08:47,232 - root - WARNING - No norm found in model
+[titan] 2025-07-23 15:08:47,232 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 15:08:47,319 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 15:08:47,402 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 15:08:47,404 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 15:08:47,468 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 15:08:47,471 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 15:08:47,503 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 15:08:47,508 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 15:08:47,570 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 15:08:55,252 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 15:08:55,254 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 15:08:55,255 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 15:08:55,255 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 15:08:55,256 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 15:08:55,256 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 15:08:55,256 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 15:08:55,256 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 15:08:55,256 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 15:08:55,256 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank6]: Traceback (most recent call last):
+[rank6]: File "", line 198, in _run_module_as_main
+[rank6]: File "", line 88, in _run_code
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank6]: main(config)
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank6]: return f(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank6]: output = model(
+[rank6]: ^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank6]: return self._call_impl(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank6]: return inner()
+[rank6]: ^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank6]: result = forward_call(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank6]: return func(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank6]: outputs = self.backbone(
+[rank6]: ^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank6]: return self._call_impl(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank6]: return forward_call(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank6]: hidden_states = mixer_block(
+[rank6]: ^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank6]: return self._call_impl(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank6]: return inner()
+[rank6]: ^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank6]: result = forward_call(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank6]: return fn(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank6]: return self._call_impl(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank6]: return forward_call(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank6]: hidden_states = self.norm(hidden_states)
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank6]: hidden_states = self.mixer(
+[rank6]: ^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank6]: return self._call_impl(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank6]: return forward_call(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 595, in forward
+[rank6]: return self.cuda_kernels_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 300, in cuda_kernels_forward
+[rank6]: out = mamba_split_conv1d_scan_combined(
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py", line 947, in mamba_split_conv1d_scan_combined
+[rank6]: return MambaSplitConv1dScanCombinedFn.apply(zxbcdt, conv1d_weight, conv1d_bias, dt_bias, A, D, chunk_size, initial_states, seq_idx, dt_limit, return_final_states, activation, rmsnorm_weight, rmsnorm_eps, outproj_weight, outproj_bias, headdim, ngroups, norm_before_gate)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/autograd/function.py", line 575, in apply
+[rank6]: return super().apply(*args, **kwargs) # type: ignore[misc]
+[rank6]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/amp/autocast_mode.py", line 510, in decorate_fwd
+[rank6]: return fwd(*args, **kwargs)
+[rank6]: ^^^^^^^^^^^^^^^^^^^^
+[rank6]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py", line 792, in forward
+[rank6]: causal_conv1d_fwd_function(rearrange_and_update_stride(xBC, "b s d -> b d s"),
+[rank6]: TypeError: 'NoneType' object is not callable
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/6/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/6/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/7/error.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/7/error.json
new file mode 100644
index 0000000000000000000000000000000000000000..2c7a9259aa9b367776fd2116bc21e4177467feb8
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/7/error.json
@@ -0,0 +1 @@
+{"message": {"message": "TypeError: 'NoneType' object is not callable", "extraInfo": {"py_callstack": "Traceback (most recent call last):\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py\", line 355, in wrapper\n return f(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py\", line 487, in main\n output = model(\n ^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py\", line 172, in wrapped_func\n return func(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 526, in forward\n outputs = self.backbone(\n ^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 405, in forward\n hidden_states = mixer_block(\n ^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1857, in _call_impl\n return inner()\n ^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1805, in inner\n result = forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py\", line 655, in _fn\n return fn(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 161, in forward\n hidden_states = self.norm(hidden_states)\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py\", line 165, in torch_dynamo_resume_in_forward_at_161\n hidden_states = self.mixer(\n ^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1751, in _wrapped_call_impl\n return self._call_impl(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py\", line 1762, in _call_impl\n return forward_call(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 595, in forward\n return self.cuda_kernels_forward(hidden_states, cache_params, cache_position, attention_mask)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py\", line 300, in cuda_kernels_forward\n out = mamba_split_conv1d_scan_combined(\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py\", line 947, in mamba_split_conv1d_scan_combined\n return MambaSplitConv1dScanCombinedFn.apply(zxbcdt, conv1d_weight, conv1d_bias, dt_bias, A, D, chunk_size, initial_states, seq_idx, dt_limit, return_final_states, activation, rmsnorm_weight, rmsnorm_eps, outproj_weight, outproj_bias, headdim, ngroups, norm_before_gate)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/autograd/function.py\", line 575, in apply\n return super().apply(*args, **kwargs) # type: ignore[misc]\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/amp/autocast_mode.py\", line 510, in decorate_fwd\n return fwd(*args, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^\n File \"/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py\", line 792, in forward\n causal_conv1d_fwd_function(rearrange_and_update_stride(xBC, \"b s d -> b d s\"),\nTypeError: 'NoneType' object is not callable\n", "timestamp": "1753254643"}}}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/7/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/7/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..03cbc29a6a775a539a6c58a7cff034268a4a55ed
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/7/stderr.log
@@ -0,0 +1,397 @@
+[titan] 2025-07-23 15:07:42,604 - root - INFO - Starting job: default job
+[titan] 2025-07-23 15:07:42,604 - root - INFO - [32m{
+ "activation_checkpoint": {
+ "mode": "none",
+ "selective_ac_option": "2"
+ },
+ "activation_offload": {
+ "mode": "none"
+ },
+ "checkpoint": {
+ "async_mode": "disabled",
+ "create_seed_checkpoint": false,
+ "enable_checkpoint": true,
+ "exclude_from_loading": [],
+ "export_dtype": "float32",
+ "folder": "checkpoint",
+ "interval": 8192,
+ "interval_type": "steps",
+ "keep_latest_k": 100,
+ "load_step": -1,
+ "model_weights_only": false
+ },
+ "comm": {
+ "init_timeout_seconds": 300,
+ "trace_buf_size": 20000,
+ "train_timeout_seconds": 100
+ },
+ "experimental": {
+ "context_parallel_degree": 1,
+ "context_parallel_rotate_method": "allgather",
+ "custom_model_path": "",
+ "enable_async_tensor_parallel": false,
+ "enable_compiled_autograd": false,
+ "pipeline_parallel_degree": 1,
+ "pipeline_parallel_microbatches": null,
+ "pipeline_parallel_schedule": "1F1B",
+ "pipeline_parallel_schedule_csv": "",
+ "pipeline_parallel_split_points": []
+ },
+ "fault_tolerance": {
+ "enable": false,
+ "group_size": 0,
+ "min_replica_size": 1,
+ "replica_id": 0
+ },
+ "float8": {
+ "enable_fsdp_float8_all_gather": false,
+ "force_recompute_fp8_weight_in_bwd": false,
+ "precompute_float8_dynamic_scale_for_fsdp": false,
+ "recipe_name": null
+ },
+ "job": {
+ "config_file": "flame/models/fla.toml",
+ "description": "default job",
+ "dump_folder": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2",
+ "print_args": true,
+ "use_for_integration_test": false
+ },
+ "lr_scheduler": {
+ "decay_ratio": 1.0,
+ "decay_type": "linear",
+ "lr_min": 0.01,
+ "warmup_steps": 100
+ },
+ "memory_estimation": {
+ "disable_fake_mode": false,
+ "enabled": false
+ },
+ "metrics": {
+ "disable_color_printing": false,
+ "enable_tensorboard": true,
+ "enable_wandb": true,
+ "log_freq": 1,
+ "save_for_all_ranks": false,
+ "save_tb_folder": "tb"
+ },
+ "model": {
+ "config": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json",
+ "converters": [],
+ "name": "fla",
+ "print_after_conversion": false,
+ "tokenizer_path": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer"
+ },
+ "optimizer": {
+ "early_step_in_backward": false,
+ "eps": 1e-08,
+ "implementation": "fused",
+ "lr": 0.0003,
+ "name": "AdamW"
+ },
+ "profiling": {
+ "enable_memory_snapshot": false,
+ "enable_profiling": true,
+ "profile_freq": 512,
+ "save_memory_snapshot_folder": "memory_snapshot",
+ "save_traces_folder": "profile_trace"
+ },
+ "training": {
+ "batch_size": 8,
+ "compile": true,
+ "context_len": 8192,
+ "data_dir": null,
+ "data_files": null,
+ "data_parallel_replicate_degree": 1,
+ "data_parallel_shard_degree": -1,
+ "data_probs": "0.55,0.3,0.15",
+ "dataset": "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro",
+ "dataset_name": "default,default,default",
+ "dataset_split": "train,train,train",
+ "deterministic": false,
+ "disable_loss_parallel": false,
+ "enable_cpu_offload": false,
+ "fsdp_reshard_after_forward": "default",
+ "gc_freq": 50,
+ "gradient_accumulation_steps": 2,
+ "max_norm": 1.0,
+ "mixed_precision_param": "bfloat16",
+ "mixed_precision_reduce": "float32",
+ "num_workers": 32,
+ "persistent_workers": false,
+ "pin_memory": false,
+ "prefetch_factor": 2,
+ "seed": 42,
+ "seq_len": 8192,
+ "skip_nan_inf": true,
+ "steps": 95366,
+ "streaming": true,
+ "tensor_parallel_degree": 1,
+ "varlen": false
+ }
+}[39m
+[titan] 2025-07-23 15:07:42,606 - root - INFO - [GC] Initial GC collection. 0.00 seconds.
+[titan] 2025-07-23 15:07:42,833 - root - WARNING - ENV[TORCH_NCCL_ASYNC_ERROR_HANDLING] = 1 will be overridden to 3 based on job config
+[titan] 2025-07-23 15:07:42,835 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 15:07:42,889 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 15:07:42,890 - root - INFO - Peak FLOPS used for computing MFU: 3.120e+14
+[titan] 2025-07-23 15:07:42,890 - root - INFO - Building 1-D device mesh with ['dp_shard'], [8]
+[titan] 2025-07-23 15:07:42,900 - root - INFO - Loading tokenizer...
+[titan] 2025-07-23 15:07:43,010 - root - INFO - LlamaTokenizerFast(name_or_path='/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/tokenizer', vocab_size=32000, model_max_length=10000000000, is_fast=True, padding_side='left', truncation_side='right', special_tokens={'bos_token': '', 'eos_token': '', 'unk_token': ''}, clean_up_tokenization_spaces=False, added_tokens_decoder={
+ 0: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 1: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+ 2: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True),
+}
+)
+[titan] 2025-07-23 15:07:43,010 - root - INFO - Loading dataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged,/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default,default,default
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:43,011 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:43,140 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample:default (p = 0.550)[39m:
+IterableDataset({
+ features: ['text', 'id', 'dump', 'url', 'file_path', 'language', 'language_score', 'token_count', 'score', 'int_score'],
+ num_shards: 140
+})
+[titan] 2025-07-23 15:07:43,140 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:07:43,140 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample has insufficient shards (140). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:07:43,140 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/fineweb-edu-sample' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:34,899 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:34,931 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged:default (p = 0.300)[39m:
+IterableDataset({
+ features: ['repo', 'content'],
+ num_shards: 1
+})
+[titan] 2025-07-23 15:08:34,931 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:08:34,931 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged has insufficient shards (1). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:34,931 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/small_repos_20B_sample_merged' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:35,195 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:35,281 - root - INFO - Subset [36m/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro:default (p = 0.150)[39m:
+IterableDataset({
+ features: ['text', 'cc-path', 'domain', 'lang', 'lang_score', 'timestamp', 'url', 'math_score'],
+ num_shards: 100
+})
+[titan] 2025-07-23 15:08:35,281 - root - INFO - Shuffling the dataset with seed 42
+[titan] 2025-07-23 15:08:35,281 - root - WARNING - [31mDataset /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro has insufficient shards (100). Need 256 shards minimum for desired data parallel workers × 32 dataloader workers. Resharding dataset to 256 shards and disabling streaming mode.[39m
+`trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:35,281 - datasets.load - ERROR - `trust_remote_code` is not supported anymore.
+Please check that the Hugging Face dataset '/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/dataset/megamath-web-pro' isn't based on a loading script and remove `trust_remote_code`.
+If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
+[titan] 2025-07-23 15:08:41,805 - root - INFO - Interleaving 3 datasets with probabilities [0.55, 0.3, 0.15]
+[titan] 2025-07-23 15:08:42,518 - root - INFO - IterableDataset({
+ features: ['text', 'content'],
+ num_shards: 256
+})
+[titan] 2025-07-23 15:08:42,644 - root - INFO - Building dataloader...
+[titan] 2025-07-23 15:08:42,646 - root - INFO - Loading model config from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/configs/mamba2_6_1_340M.json
+[titan] 2025-07-23 15:08:42,648 - root - INFO - Building model from the config
+[32mMamba2Config {
+ "architectures": [
+ "Mamba2ForCausalLM"
+ ],
+ "attn": {
+ "layers": [
+ 5,
+ 11,
+ 17,
+ 23
+ ],
+ "num_heads": 16,
+ "num_kv_heads": 8,
+ "qkv_bias": false,
+ "rope_theta": 160000.0,
+ "window_size": null
+ },
+ "attn_mode": "chunk",
+ "bos_token_id": 1,
+ "chunk_size": 256,
+ "conv_kernel": 4,
+ "eos_token_id": 2,
+ "expand": 2,
+ "fuse_cross_entropy": true,
+ "fuse_norm": true,
+ "fuse_swiglu": true,
+ "head_dim": 64,
+ "hidden_act": "silu",
+ "hidden_size": 1024,
+ "initializer_range": 0.02,
+ "model_type": "mamba2",
+ "n_groups": 1,
+ "norm_eps": 1e-05,
+ "num_heads": 32,
+ "num_hidden_layers": 48,
+ "pad_token_id": 0,
+ "rescale_prenorm_residual": true,
+ "residual_in_fp32": true,
+ "rms_norm": true,
+ "state_size": 128,
+ "tie_word_embeddings": false,
+ "time_step_floor": 0.0001,
+ "time_step_limit": [
+ 0.0,
+ Infinity
+ ],
+ "time_step_max": 0.1,
+ "time_step_min": 0.001,
+ "time_step_rank": 128,
+ "transformers_version": "4.53.3",
+ "use_bias": false,
+ "use_cache": true,
+ "use_conv_bias": true,
+ "use_l2warp": false,
+ "vocab_size": 32000
+}
+[39m
+[titan] 2025-07-23 15:08:42,859 - fla.layers.mamba2 - WARNING - The CUDA backend is not available because `causal_conv1d` is None. Falling back to the Triton backend. To install follow https://github.com/Dao-AILab/causal-conv1d
+[titan] 2025-07-23 15:08:42,957 - root - INFO - [34m
+Mamba2ForCausalLM(
+ (backbone): Mamba2Model(
+ (embeddings): Embedding(32000, 1024)
+ (layers): ModuleList(
+ (0-47): 48 x Mamba2Block(
+ (norm): RMSNorm(1024, eps=1e-05)
+ (mixer): Mamba2(
+ (conv1d): Conv1d(2304, 2304, kernel_size=(4,), stride=(1,), padding=(3,), groups=2304)
+ (in_proj): Linear(in_features=1024, out_features=4384, bias=False)
+ (norm): RMSNormGated()
+ (out_proj): Linear(in_features=2048, out_features=1024, bias=False)
+ )
+ )
+ )
+ (norm_f): RMSNorm(1024, eps=1e-05)
+ )
+ (lm_head): Linear(in_features=1024, out_features=32000, bias=False)
+ (criterion): FusedLinearCrossEntropyLoss()
+)[39m
+
+[titan] 2025-07-23 15:08:43,010 - root - INFO - Compiling each block with torch.compile
+[titan] 2025-07-23 15:08:43,010 - root - INFO - Compiling the embedding, norm, and lm_head layers with torch.compile
+[titan] 2025-07-23 15:08:43,010 - root - WARNING - No norm found in model
+[titan] 2025-07-23 15:08:43,011 - root - INFO - Compiling the entire model with torch.compile
+[titan] 2025-07-23 15:08:43,091 - root - INFO - Applied FSDP to the model
+[titan] 2025-07-23 15:08:43,152 - fla.models.mamba2.modeling_mamba2 - WARNING - `A_log` is a DTensor, skipping initialization
+[titan] 2025-07-23 15:08:43,153 - fla.models.mamba2.modeling_mamba2 - WARNING - `dt_bias` is a DTensor, skipping initialization
+[titan] 2025-07-23 15:08:43,201 - root - INFO - CUDA memory usage for model: 0.19GiB(0.20%)
+[titan] 2025-07-23 15:08:43,203 - root - WARNING - Warmup (100) + decay (95366) steps exceed total training steps (95366). Adjusting decay steps to 95266.
+[titan] 2025-07-23 15:08:43,226 - root - INFO - Checkpointing active. Checkpoints will be loaded from and saved to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/checkpoint
+[titan] 2025-07-23 15:08:43,231 - root - INFO - CUDA capacity: NVIDIA H20 with 95.00GiB memory
+[titan] 2025-07-23 15:08:43,280 - root - WARNING - Peak flops undefined for: NVIDIA H20, fallback to A100
+[titan] 2025-07-23 15:08:50,502 - root - INFO - [31m***** Running training *****[39m
+[titan] 2025-07-23 15:08:50,504 - root - INFO - [32m Training starts at step 1
+[titan] 2025-07-23 15:08:50,506 - root - INFO - [32m Number of tokens per sequence = 8,192
+[titan] 2025-07-23 15:08:50,506 - root - INFO - [32m Gradient Accumulation steps = 2
+[titan] 2025-07-23 15:08:50,506 - root - INFO - [32m Instantaneous batch size (per device) = 8
+[titan] 2025-07-23 15:08:50,506 - root - INFO - [32m Global batch size (w. parallel, distributed & accumulation) = 128 (1,048,576 tokens)
+[titan] 2025-07-23 15:08:50,506 - root - INFO - [32m Total optimization steps = 95,366 (99,998,498,816 tokens)
+[titan] 2025-07-23 15:08:50,507 - root - INFO - [32m Warmup steps = 100 (104,857,600 tokens)
+[titan] 2025-07-23 15:08:50,507 - root - INFO - [32m Number of parameters = 382,387,712 [39m
+[titan] 2025-07-23 15:08:50,507 - root - INFO - Profiling active. Traces will be saved at /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/profile_trace
+/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/variables/functions.py:1263: UserWarning: Dynamo does not know how to trace the builtin `cuda_utils.get_device_properties.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind).
+If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround.
+If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`.
+ torch._dynamo.utils.warn_once(explanation + "\n" + "\n".join(hints))
+[rank7]: Traceback (most recent call last):
+[rank7]: File "", line 198, in _run_module_as_main
+[rank7]: File "", line 88, in _run_code
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 615, in
+[rank7]: main(config)
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
+[rank7]: return f(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/flame/train.py", line 487, in main
+[rank7]: output = model(
+[rank7]: ^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank7]: return self._call_impl(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank7]: return inner()
+[rank7]: ^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank7]: result = forward_call(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
+[rank7]: return func(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 526, in forward
+[rank7]: outputs = self.backbone(
+[rank7]: ^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank7]: return self._call_impl(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank7]: return forward_call(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 405, in forward
+[rank7]: hidden_states = mixer_block(
+[rank7]: ^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank7]: return self._call_impl(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1857, in _call_impl
+[rank7]: return inner()
+[rank7]: ^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1805, in inner
+[rank7]: result = forward_call(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 655, in _fn
+[rank7]: return fn(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank7]: return self._call_impl(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank7]: return forward_call(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 161, in forward
+[rank7]: hidden_states = self.norm(hidden_states)
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/models/mamba2/modeling_mamba2.py", line 165, in torch_dynamo_resume_in_forward_at_161
+[rank7]: hidden_states = self.mixer(
+[rank7]: ^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1751, in _wrapped_call_impl
+[rank7]: return self._call_impl(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1762, in _call_impl
+[rank7]: return forward_call(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 595, in forward
+[rank7]: return self.cuda_kernels_forward(hidden_states, cache_params, cache_position, attention_mask)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flash-linear-attention/fla/layers/mamba2.py", line 300, in cuda_kernels_forward
+[rank7]: out = mamba_split_conv1d_scan_combined(
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py", line 947, in mamba_split_conv1d_scan_combined
+[rank7]: return MambaSplitConv1dScanCombinedFn.apply(zxbcdt, conv1d_weight, conv1d_bias, dt_bias, A, D, chunk_size, initial_states, seq_idx, dt_limit, return_final_states, activation, rmsnorm_weight, rmsnorm_eps, outproj_weight, outproj_bias, headdim, ngroups, norm_before_gate)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/autograd/function.py", line 575, in apply
+[rank7]: return super().apply(*args, **kwargs) # type: ignore[misc]
+[rank7]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/torch/amp/autocast_mode.py", line 510, in decorate_fwd
+[rank7]: return fwd(*args, **kwargs)
+[rank7]: ^^^^^^^^^^^^^^^^^^^^
+[rank7]: File "/mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/.venv/lib/python3.11/site-packages/mamba_ssm/ops/triton/ssd_combined.py", line 792, in forward
+[rank7]: causal_conv1d_fwd_function(rearrange_and_update_stride(xBC, "b s d -> b d s"),
+[rank7]: TypeError: 'NoneType' object is not callable
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/7/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_r3tps1xj/attempt_0/7/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/0/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/0/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..9d92f69d8262075eafce22ea726f5abdf98e4521
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/0/stderr.log
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:117c2d1396d33cc7e76fdb9c97cb05862ce308b64fe1656469b0224ff9a5061b
+size 28955181
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/0/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/0/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e5dc269d2c8237051b7e4355aac6e33d77b20294
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/0/stdout.log
@@ -0,0 +1,3 @@
+[1;34mwandb[0m:
+[1;34mwandb[0m: 🚀 View run [33mmamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2[0m at: [34m[0m
+[1;34mwandb[0m: Find logs at: [1;35mexp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/run-20250723_155414-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552/logs[0m
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/1/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/1/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..90a3083eafdecd3e77ce1117e00248a5b476c3b6
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/1/stderr.log
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a08bc6b625bbcec2f40e26580d3bd5f9814ce8228c59f8e8cc47dafa103a6a58
+size 28954684
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/1/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/1/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/2/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/2/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..7a42a6f0eeb2ec3b5c4095cbdb345d9c79306696
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/2/stderr.log
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:63a78af2803cfec87bb5efd840493146442fed6cef87a9e112deec31074f22fd
+size 28954686
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/2/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/2/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/3/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/3/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..80d468b7ce9ff22e24441e0315c9eb2fd7a00843
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/3/stderr.log
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:df3d44e4428ffc3aa07f695d6f0b66bf2ca04930b12c28dd4b4ba0a0c0c62efa
+size 28954685
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/3/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/3/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/4/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/4/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..2f14628a242f87de74a47bd73974b4d579bac0ab
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/4/stderr.log
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ad1e252441ef71c0580b5095222eb3c71d1bef5e7e4578bf19e648244d993a6d
+size 28954685
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/4/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/4/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/5/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/5/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..2bc16d5ed84e8b12b5b97dc51c22fddbba5f4ded
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/5/stderr.log
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8311c4ea59ca424a6d186bebe66fc59d22f7500f838c23028d378311eb0975fe
+size 28954685
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/5/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/5/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/6/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/6/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..857bfa0f5bdbb05d0bce62e9298f78019dc9b8ba
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/6/stderr.log
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8a40c17a824eab9916f9c2392268e1b98ee282a543d4f6b524f100dedcae7753
+size 28954685
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/6/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/6/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/7/stderr.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/7/stderr.log
new file mode 100644
index 0000000000000000000000000000000000000000..9bec193a9704236f3f8c1273dd0d88cc09818c59
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/7/stderr.log
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1f87ea77e4505b9e5f71ad1c3e6b3b6e577ed2542b22b97a19a9356433b09789
+size 28954685
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/7/stdout.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/logs/none_v3h3fbcf/attempt_0/7/stdout.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/model.safetensors b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..80210beb2175f658af10061761e225a874d548ee
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9ccf331bbca9e17b512e6d199e3002d6fe69c042a44b9b867783b3ed6a625060
+size 1529597152
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/model_size=382m b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/model_size=382m
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/special_tokens_map.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..451134b2ddc2e78555d1e857518c54b4bdc2e87d
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/special_tokens_map.json
@@ -0,0 +1,23 @@
+{
+ "bos_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "eos_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "unk_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ }
+}
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1140/events.out.tfevents.1753242153.TENCENT64.site.306046.0 b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1140/events.out.tfevents.1753242153.TENCENT64.site.306046.0
new file mode 100644
index 0000000000000000000000000000000000000000..c7e676bda7ec52f595a1a9b21bc4b94f7768f43d
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1140/events.out.tfevents.1753242153.TENCENT64.site.306046.0
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4146871b3c9be36510d5c910ae5196f44c46f102f3bac6be21159a4350f1f380
+size 88
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1140/wandb/debug-internal.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1140/wandb/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..655f9b820ab71f5654fda379006dcf44c0da55f6
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1140/wandb/debug-internal.log
@@ -0,0 +1,8 @@
+{"time":"2025-07-23T11:41:03.340954687+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.0"}
+{"time":"2025-07-23T11:41:04.585656278+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T11:41:06.736194203+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T11:41:10.967597934+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T11:41:19.985299359+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T11:41:39.065583721+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T11:42:12.387420422+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T11:43:12.407046332+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1140/wandb/debug.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1140/wandb/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..f7d2aa724ae03a0534ab0ed60c1e92b869b9b59a
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1140/wandb/debug.log
@@ -0,0 +1,15 @@
+2025-07-23 11:41:00,777 INFO MainThread:306046 [wandb_setup.py:_flush():80] Current SDK version is 0.21.0
+2025-07-23 11:41:00,777 INFO MainThread:306046 [wandb_setup.py:_flush():80] Configure stats pid to 306046
+2025-07-23 11:41:00,777 INFO MainThread:306046 [wandb_setup.py:_flush():80] Loading settings from /root/.config/wandb/settings
+2025-07-23 11:41:00,777 INFO MainThread:306046 [wandb_setup.py:_flush():80] Loading settings from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/wandb/settings
+2025-07-23 11:41:00,777 INFO MainThread:306046 [wandb_setup.py:_flush():80] Loading settings from environment variables
+2025-07-23 11:41:00,778 INFO MainThread:306046 [wandb_init.py:setup_run_log_directory():703] Logging user logs to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1140/wandb/run-20250723_114100-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231112/logs/debug.log
+2025-07-23 11:41:00,778 INFO MainThread:306046 [wandb_init.py:setup_run_log_directory():704] Logging internal logs to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1140/wandb/run-20250723_114100-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231112/logs/debug-internal.log
+2025-07-23 11:41:00,778 INFO MainThread:306046 [wandb_init.py:init():830] calling init triggers
+2025-07-23 11:41:00,778 INFO MainThread:306046 [wandb_init.py:init():835] wandb.init called with sweep_config: {}
+config: {'_wandb': {}}
+2025-07-23 11:41:00,778 INFO MainThread:306046 [wandb_init.py:init():871] starting backend
+2025-07-23 11:41:03,210 INFO MainThread:306046 [wandb_init.py:init():874] sending inform_init request
+2025-07-23 11:41:03,320 INFO MainThread:306046 [wandb_init.py:init():882] backend started and connected
+2025-07-23 11:41:03,321 INFO MainThread:306046 [wandb_init.py:init():953] updated telemetry
+2025-07-23 11:41:03,608 INFO MainThread:306046 [wandb_init.py:init():977] communicating run to backend with 90.0 second timeout
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1140/wandb/run-20250723_114100-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231112/logs/debug-internal.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1140/wandb/run-20250723_114100-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231112/logs/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..655f9b820ab71f5654fda379006dcf44c0da55f6
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1140/wandb/run-20250723_114100-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231112/logs/debug-internal.log
@@ -0,0 +1,8 @@
+{"time":"2025-07-23T11:41:03.340954687+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.0"}
+{"time":"2025-07-23T11:41:04.585656278+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T11:41:06.736194203+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T11:41:10.967597934+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T11:41:19.985299359+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T11:41:39.065583721+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T11:42:12.387420422+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T11:43:12.407046332+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1140/wandb/run-20250723_114100-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231112/logs/debug.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1140/wandb/run-20250723_114100-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231112/logs/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..f7d2aa724ae03a0534ab0ed60c1e92b869b9b59a
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1140/wandb/run-20250723_114100-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231112/logs/debug.log
@@ -0,0 +1,15 @@
+2025-07-23 11:41:00,777 INFO MainThread:306046 [wandb_setup.py:_flush():80] Current SDK version is 0.21.0
+2025-07-23 11:41:00,777 INFO MainThread:306046 [wandb_setup.py:_flush():80] Configure stats pid to 306046
+2025-07-23 11:41:00,777 INFO MainThread:306046 [wandb_setup.py:_flush():80] Loading settings from /root/.config/wandb/settings
+2025-07-23 11:41:00,777 INFO MainThread:306046 [wandb_setup.py:_flush():80] Loading settings from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/wandb/settings
+2025-07-23 11:41:00,777 INFO MainThread:306046 [wandb_setup.py:_flush():80] Loading settings from environment variables
+2025-07-23 11:41:00,778 INFO MainThread:306046 [wandb_init.py:setup_run_log_directory():703] Logging user logs to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1140/wandb/run-20250723_114100-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231112/logs/debug.log
+2025-07-23 11:41:00,778 INFO MainThread:306046 [wandb_init.py:setup_run_log_directory():704] Logging internal logs to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1140/wandb/run-20250723_114100-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231112/logs/debug-internal.log
+2025-07-23 11:41:00,778 INFO MainThread:306046 [wandb_init.py:init():830] calling init triggers
+2025-07-23 11:41:00,778 INFO MainThread:306046 [wandb_init.py:init():835] wandb.init called with sweep_config: {}
+config: {'_wandb': {}}
+2025-07-23 11:41:00,778 INFO MainThread:306046 [wandb_init.py:init():871] starting backend
+2025-07-23 11:41:03,210 INFO MainThread:306046 [wandb_init.py:init():874] sending inform_init request
+2025-07-23 11:41:03,320 INFO MainThread:306046 [wandb_init.py:init():882] backend started and connected
+2025-07-23 11:41:03,321 INFO MainThread:306046 [wandb_init.py:init():953] updated telemetry
+2025-07-23 11:41:03,608 INFO MainThread:306046 [wandb_init.py:init():977] communicating run to backend with 90.0 second timeout
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1428/events.out.tfevents.1753252244.TENCENT64.site.430813.0 b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1428/events.out.tfevents.1753252244.TENCENT64.site.430813.0
new file mode 100644
index 0000000000000000000000000000000000000000..08b8fddfd2986ca77f054d67cbb0632d406059b2
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1428/events.out.tfevents.1753252244.TENCENT64.site.430813.0
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6cabdd678210d0b88ff4aa1a5021496b9a58417a4bfa741536aafcdee80700b6
+size 88
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1428/wandb/debug-internal.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1428/wandb/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..9b0fde86f0a514c2d96d38a6b3969d90c520ffd9
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1428/wandb/debug-internal.log
@@ -0,0 +1,8 @@
+{"time":"2025-07-23T14:29:14.323738565+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.0"}
+{"time":"2025-07-23T14:29:14.84297462+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T14:29:17.181658577+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T14:29:21.79775038+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T14:29:31.71812748+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T14:29:50.349160073+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T14:30:28.919888486+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T14:31:28.941753958+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1428/wandb/debug.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1428/wandb/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..ce9961dc1505314a2fb0d89efe29fcb859c7eb1e
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1428/wandb/debug.log
@@ -0,0 +1,15 @@
+2025-07-23 14:29:12,054 INFO MainThread:430813 [wandb_setup.py:_flush():80] Current SDK version is 0.21.0
+2025-07-23 14:29:12,054 INFO MainThread:430813 [wandb_setup.py:_flush():80] Configure stats pid to 430813
+2025-07-23 14:29:12,054 INFO MainThread:430813 [wandb_setup.py:_flush():80] Loading settings from /root/.config/wandb/settings
+2025-07-23 14:29:12,054 INFO MainThread:430813 [wandb_setup.py:_flush():80] Loading settings from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/wandb/settings
+2025-07-23 14:29:12,054 INFO MainThread:430813 [wandb_setup.py:_flush():80] Loading settings from environment variables
+2025-07-23 14:29:12,055 INFO MainThread:430813 [wandb_init.py:setup_run_log_directory():703] Logging user logs to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1428/wandb/run-20250723_142911-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231426/logs/debug.log
+2025-07-23 14:29:12,055 INFO MainThread:430813 [wandb_init.py:setup_run_log_directory():704] Logging internal logs to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1428/wandb/run-20250723_142911-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231426/logs/debug-internal.log
+2025-07-23 14:29:12,055 INFO MainThread:430813 [wandb_init.py:init():830] calling init triggers
+2025-07-23 14:29:12,055 INFO MainThread:430813 [wandb_init.py:init():835] wandb.init called with sweep_config: {}
+config: {'_wandb': {}}
+2025-07-23 14:29:12,055 INFO MainThread:430813 [wandb_init.py:init():871] starting backend
+2025-07-23 14:29:14,221 INFO MainThread:430813 [wandb_init.py:init():874] sending inform_init request
+2025-07-23 14:29:14,272 INFO MainThread:430813 [wandb_init.py:init():882] backend started and connected
+2025-07-23 14:29:14,274 INFO MainThread:430813 [wandb_init.py:init():953] updated telemetry
+2025-07-23 14:29:14,433 INFO MainThread:430813 [wandb_init.py:init():977] communicating run to backend with 90.0 second timeout
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1428/wandb/run-20250723_142911-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231426/logs/debug-internal.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1428/wandb/run-20250723_142911-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231426/logs/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..9b0fde86f0a514c2d96d38a6b3969d90c520ffd9
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1428/wandb/run-20250723_142911-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231426/logs/debug-internal.log
@@ -0,0 +1,8 @@
+{"time":"2025-07-23T14:29:14.323738565+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.0"}
+{"time":"2025-07-23T14:29:14.84297462+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T14:29:17.181658577+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T14:29:21.79775038+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T14:29:31.71812748+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T14:29:50.349160073+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T14:30:28.919888486+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T14:31:28.941753958+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1428/wandb/run-20250723_142911-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231426/logs/debug.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1428/wandb/run-20250723_142911-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231426/logs/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..ce9961dc1505314a2fb0d89efe29fcb859c7eb1e
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1428/wandb/run-20250723_142911-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231426/logs/debug.log
@@ -0,0 +1,15 @@
+2025-07-23 14:29:12,054 INFO MainThread:430813 [wandb_setup.py:_flush():80] Current SDK version is 0.21.0
+2025-07-23 14:29:12,054 INFO MainThread:430813 [wandb_setup.py:_flush():80] Configure stats pid to 430813
+2025-07-23 14:29:12,054 INFO MainThread:430813 [wandb_setup.py:_flush():80] Loading settings from /root/.config/wandb/settings
+2025-07-23 14:29:12,054 INFO MainThread:430813 [wandb_setup.py:_flush():80] Loading settings from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/wandb/settings
+2025-07-23 14:29:12,054 INFO MainThread:430813 [wandb_setup.py:_flush():80] Loading settings from environment variables
+2025-07-23 14:29:12,055 INFO MainThread:430813 [wandb_init.py:setup_run_log_directory():703] Logging user logs to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1428/wandb/run-20250723_142911-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231426/logs/debug.log
+2025-07-23 14:29:12,055 INFO MainThread:430813 [wandb_init.py:setup_run_log_directory():704] Logging internal logs to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1428/wandb/run-20250723_142911-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231426/logs/debug-internal.log
+2025-07-23 14:29:12,055 INFO MainThread:430813 [wandb_init.py:init():830] calling init triggers
+2025-07-23 14:29:12,055 INFO MainThread:430813 [wandb_init.py:init():835] wandb.init called with sweep_config: {}
+config: {'_wandb': {}}
+2025-07-23 14:29:12,055 INFO MainThread:430813 [wandb_init.py:init():871] starting backend
+2025-07-23 14:29:14,221 INFO MainThread:430813 [wandb_init.py:init():874] sending inform_init request
+2025-07-23 14:29:14,272 INFO MainThread:430813 [wandb_init.py:init():882] backend started and connected
+2025-07-23 14:29:14,274 INFO MainThread:430813 [wandb_init.py:init():953] updated telemetry
+2025-07-23 14:29:14,433 INFO MainThread:430813 [wandb_init.py:init():977] communicating run to backend with 90.0 second timeout
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1508/events.out.tfevents.1753254624.TENCENT64.site.541431.0 b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1508/events.out.tfevents.1753254624.TENCENT64.site.541431.0
new file mode 100644
index 0000000000000000000000000000000000000000..bd6e17587e69bc490dd59363f5cd86efd4434b71
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1508/events.out.tfevents.1753254624.TENCENT64.site.541431.0
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8d7f10852840a1ee1a430c368cf177955559fbe295bae9c4f7ec0109b3bef750
+size 88
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1508/wandb/debug-internal.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1508/wandb/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..2fb8171b6e21688661186e5576ab3d4157f5f2b9
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1508/wandb/debug-internal.log
@@ -0,0 +1,7 @@
+{"time":"2025-07-23T15:08:54.275790393+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.0"}
+{"time":"2025-07-23T15:08:54.538485652+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:08:56.6694466+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:09:01.59828124+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:09:11.28028992+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:09:29.996431148+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:10:06.03981597+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1508/wandb/debug.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1508/wandb/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..58eb7bb99c9198afa48f5ff99c4ad6885a7e5133
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1508/wandb/debug.log
@@ -0,0 +1,15 @@
+2025-07-23 15:08:54,033 INFO MainThread:541431 [wandb_setup.py:_flush():80] Current SDK version is 0.21.0
+2025-07-23 15:08:54,033 INFO MainThread:541431 [wandb_setup.py:_flush():80] Configure stats pid to 541431
+2025-07-23 15:08:54,033 INFO MainThread:541431 [wandb_setup.py:_flush():80] Loading settings from /root/.config/wandb/settings
+2025-07-23 15:08:54,034 INFO MainThread:541431 [wandb_setup.py:_flush():80] Loading settings from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/wandb/settings
+2025-07-23 15:08:54,034 INFO MainThread:541431 [wandb_setup.py:_flush():80] Loading settings from environment variables
+2025-07-23 15:08:54,034 INFO MainThread:541431 [wandb_init.py:setup_run_log_directory():703] Logging user logs to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1508/wandb/run-20250723_150853-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231507/logs/debug.log
+2025-07-23 15:08:54,034 INFO MainThread:541431 [wandb_init.py:setup_run_log_directory():704] Logging internal logs to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1508/wandb/run-20250723_150853-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231507/logs/debug-internal.log
+2025-07-23 15:08:54,034 INFO MainThread:541431 [wandb_init.py:init():830] calling init triggers
+2025-07-23 15:08:54,035 INFO MainThread:541431 [wandb_init.py:init():835] wandb.init called with sweep_config: {}
+config: {'_wandb': {}}
+2025-07-23 15:08:54,035 INFO MainThread:541431 [wandb_init.py:init():871] starting backend
+2025-07-23 15:08:54,244 INFO MainThread:541431 [wandb_init.py:init():874] sending inform_init request
+2025-07-23 15:08:54,264 INFO MainThread:541431 [wandb_init.py:init():882] backend started and connected
+2025-07-23 15:08:54,266 INFO MainThread:541431 [wandb_init.py:init():953] updated telemetry
+2025-07-23 15:08:54,324 INFO MainThread:541431 [wandb_init.py:init():977] communicating run to backend with 90.0 second timeout
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1508/wandb/run-20250723_150853-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231507/logs/debug-internal.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1508/wandb/run-20250723_150853-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231507/logs/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..2fb8171b6e21688661186e5576ab3d4157f5f2b9
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1508/wandb/run-20250723_150853-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231507/logs/debug-internal.log
@@ -0,0 +1,7 @@
+{"time":"2025-07-23T15:08:54.275790393+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.0"}
+{"time":"2025-07-23T15:08:54.538485652+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:08:56.6694466+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:09:01.59828124+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:09:11.28028992+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:09:29.996431148+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:10:06.03981597+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1508/wandb/run-20250723_150853-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231507/logs/debug.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1508/wandb/run-20250723_150853-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231507/logs/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..58eb7bb99c9198afa48f5ff99c4ad6885a7e5133
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1508/wandb/run-20250723_150853-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231507/logs/debug.log
@@ -0,0 +1,15 @@
+2025-07-23 15:08:54,033 INFO MainThread:541431 [wandb_setup.py:_flush():80] Current SDK version is 0.21.0
+2025-07-23 15:08:54,033 INFO MainThread:541431 [wandb_setup.py:_flush():80] Configure stats pid to 541431
+2025-07-23 15:08:54,033 INFO MainThread:541431 [wandb_setup.py:_flush():80] Loading settings from /root/.config/wandb/settings
+2025-07-23 15:08:54,034 INFO MainThread:541431 [wandb_setup.py:_flush():80] Loading settings from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/wandb/settings
+2025-07-23 15:08:54,034 INFO MainThread:541431 [wandb_setup.py:_flush():80] Loading settings from environment variables
+2025-07-23 15:08:54,034 INFO MainThread:541431 [wandb_init.py:setup_run_log_directory():703] Logging user logs to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1508/wandb/run-20250723_150853-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231507/logs/debug.log
+2025-07-23 15:08:54,034 INFO MainThread:541431 [wandb_init.py:setup_run_log_directory():704] Logging internal logs to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1508/wandb/run-20250723_150853-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231507/logs/debug-internal.log
+2025-07-23 15:08:54,034 INFO MainThread:541431 [wandb_init.py:init():830] calling init triggers
+2025-07-23 15:08:54,035 INFO MainThread:541431 [wandb_init.py:init():835] wandb.init called with sweep_config: {}
+config: {'_wandb': {}}
+2025-07-23 15:08:54,035 INFO MainThread:541431 [wandb_init.py:init():871] starting backend
+2025-07-23 15:08:54,244 INFO MainThread:541431 [wandb_init.py:init():874] sending inform_init request
+2025-07-23 15:08:54,264 INFO MainThread:541431 [wandb_init.py:init():882] backend started and connected
+2025-07-23 15:08:54,266 INFO MainThread:541431 [wandb_init.py:init():953] updated telemetry
+2025-07-23 15:08:54,324 INFO MainThread:541431 [wandb_init.py:init():977] communicating run to backend with 90.0 second timeout
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/events.out.tfevents.1753257347.TENCENT64.site.647798.0 b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/events.out.tfevents.1753257347.TENCENT64.site.647798.0
new file mode 100644
index 0000000000000000000000000000000000000000..7b648a1d0172cf2860a383079bd247ddf913e9de
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/events.out.tfevents.1753257347.TENCENT64.site.647798.0
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:cb015aa87fc61af85f5a06bcb99ba99ea60d4a6db1d03685b9b9927fa575df4f
+size 97469568
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/debug-internal.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..6400306d0d776139e44e80fc1d699856eb70cc51
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/debug-internal.log
@@ -0,0 +1,82 @@
+{"time":"2025-07-23T15:54:17.202033271+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.0"}
+{"time":"2025-07-23T15:54:17.726221602+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:54:19.870675818+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:54:23.95323127+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:54:32.343392135+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:54:50.177516176+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:55:22.32776313+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:56:22.349459294+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:57:22.373014277+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:58:22.393570956+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:59:22.415931625+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:00:22.435251318+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:01:22.453760112+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:02:22.47481149+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:03:22.494641285+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:04:22.514637397+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:05:22.53908081+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:06:22.554718667+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:07:22.575950592+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:08:22.59516599+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:09:22.618204936+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:10:22.640285491+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:10:22.640561155+08:00","level":"ERROR","msg":"Failed to load features, feature will default to disabled","error":"api: failed sending: POST https://api.wandb.ai/graphql giving up after 21 attempt(s): Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:10:22.655245653+08:00","level":"INFO","msg":"stream: created new stream","id":"mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552"}
+{"time":"2025-07-23T16:10:22.65537423+08:00","level":"INFO","msg":"stream: started","id":"mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552"}
+{"time":"2025-07-23T16:10:22.655404089+08:00","level":"INFO","msg":"writer: Do: started","stream_id":"mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552"}
+{"time":"2025-07-23T16:10:22.655443978+08:00","level":"INFO","msg":"sender: started","stream_id":"mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552"}
+{"time":"2025-07-23T16:10:22.655438278+08:00","level":"INFO","msg":"handler: started","stream_id":"mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552"}
+{"time":"2025-07-23T16:10:22.674547549+08:00","level":"WARN","msg":"runupserter: server does not expand metric globs but the x_server_side_expand_glob_metrics setting is set; ignoring"}
+{"time":"2025-07-23T16:10:22.677445015+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:10:24.855217555+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:10:29.31752334+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:10:37.829357598+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:10:54.257058229+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:11:33.270872917+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:12:33.290922758+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:13:33.309386134+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:14:33.328175182+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:15:33.345597329+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:16:33.368160115+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:17:33.384016261+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:18:33.403106053+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:19:33.424551514+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:20:22.660752778+08:00","level":"WARN","msg":"sender: taking a long time","seconds":600.000167592,"work":"RunUpdateWork; Control(mailbox_slot:\"ld5863kv3ux3\" connection_id:\"1(@)\")"}
+{"time":"2025-07-23T16:20:33.446092222+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:21:33.4685356+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:22:33.488932883+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:23:33.508398277+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:24:33.527479449+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:25:33.544379368+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:26:33.564775821+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:26:33.565144643+08:00","level":"ERROR","msg":"runupserter: failed to init run","error":"api: failed sending: POST https://api.wandb.ai/graphql giving up after 21 attempt(s): Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:26:33.565396718+08:00","level":"INFO","msg":"sender: succeeded after taking longer than expected","seconds":970.90488677,"work":"RunUpdateWork; Control(mailbox_slot:\"ld5863kv3ux3\" connection_id:\"1(@)\")"}
+{"time":"2025-07-29T06:43:53.318006912+08:00","level":"INFO","msg":"stream: closing","id":"mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552"}
+{"time":"2025-07-29T06:43:53.338897634+08:00","level":"ERROR","msg":"sender: uploadConfigFile: stream: no run"}
+{"time":"2025-07-29T06:43:53.394096172+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:43:55.875989199+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:44:00.555698012+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:44:09.89131533+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:44:27.958200974+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:45:05.127170144+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:46:05.147917701+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:47:05.170266993+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:48:05.191985969+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:49:05.214191483+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:50:05.231554686+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:51:05.248024048+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:52:05.267259389+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:53:05.288484475+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:54:05.304448647+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:55:05.327261179+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:56:05.351203715+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:57:05.372446132+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:58:05.395235895+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:59:05.417328624+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T07:00:05.43906279+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T07:00:05.439375173+08:00","level":"ERROR","msg":"runfiles: CreateRunFiles returned error: api: failed sending: POST https://api.wandb.ai/graphql giving up after 21 attempt(s): Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T07:00:05.485814133+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
+{"time":"2025-07-29T07:00:05.486040248+08:00","level":"INFO","msg":"handler: closed","stream_id":"mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552"}
+{"time":"2025-07-29T07:00:05.486115986+08:00","level":"INFO","msg":"sender: closed","stream_id":"mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552"}
+{"time":"2025-07-29T07:00:05.486103247+08:00","level":"INFO","msg":"writer: Close: closed","stream_id":"mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552"}
+{"time":"2025-07-29T07:00:05.486725823+08:00","level":"INFO","msg":"stream: closed","id":"mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552"}
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/debug.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..a3f349e5f98c0379797ba2d97e67c38438cfcd7b
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/debug.log
@@ -0,0 +1,15 @@
+2025-07-23 15:54:14,964 INFO MainThread:647798 [wandb_setup.py:_flush():80] Current SDK version is 0.21.0
+2025-07-23 15:54:14,964 INFO MainThread:647798 [wandb_setup.py:_flush():80] Configure stats pid to 647798
+2025-07-23 15:54:14,965 INFO MainThread:647798 [wandb_setup.py:_flush():80] Loading settings from /root/.config/wandb/settings
+2025-07-23 15:54:14,965 INFO MainThread:647798 [wandb_setup.py:_flush():80] Loading settings from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/wandb/settings
+2025-07-23 15:54:14,965 INFO MainThread:647798 [wandb_setup.py:_flush():80] Loading settings from environment variables
+2025-07-23 15:54:14,965 INFO MainThread:647798 [wandb_init.py:setup_run_log_directory():703] Logging user logs to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/run-20250723_155414-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552/logs/debug.log
+2025-07-23 15:54:14,965 INFO MainThread:647798 [wandb_init.py:setup_run_log_directory():704] Logging internal logs to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/run-20250723_155414-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552/logs/debug-internal.log
+2025-07-23 15:54:14,965 INFO MainThread:647798 [wandb_init.py:init():830] calling init triggers
+2025-07-23 15:54:14,965 INFO MainThread:647798 [wandb_init.py:init():835] wandb.init called with sweep_config: {}
+config: {'_wandb': {}}
+2025-07-23 15:54:14,965 INFO MainThread:647798 [wandb_init.py:init():871] starting backend
+2025-07-23 15:54:17,123 INFO MainThread:647798 [wandb_init.py:init():874] sending inform_init request
+2025-07-23 15:54:17,174 INFO MainThread:647798 [wandb_init.py:init():882] backend started and connected
+2025-07-23 15:54:17,175 INFO MainThread:647798 [wandb_init.py:init():953] updated telemetry
+2025-07-23 15:54:17,331 INFO MainThread:647798 [wandb_init.py:init():977] communicating run to backend with 90.0 second timeout
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/run-20250723_155414-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552/files/output.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/run-20250723_155414-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552/files/output.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/run-20250723_155414-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552/files/wandb-summary.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/run-20250723_155414-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552/files/wandb-summary.json
new file mode 100644
index 0000000000000000000000000000000000000000..b0a620d0c1047a4dd8a400939b6da246ed8063a7
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/run-20250723_155414-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552/files/wandb-summary.json
@@ -0,0 +1 @@
+{"_wandb":{"runtime":0},"_runtime":0}
\ No newline at end of file
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/run-20250723_155414-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552/logs/debug-internal.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/run-20250723_155414-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552/logs/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..6400306d0d776139e44e80fc1d699856eb70cc51
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/run-20250723_155414-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552/logs/debug-internal.log
@@ -0,0 +1,82 @@
+{"time":"2025-07-23T15:54:17.202033271+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.0"}
+{"time":"2025-07-23T15:54:17.726221602+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:54:19.870675818+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:54:23.95323127+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:54:32.343392135+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:54:50.177516176+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:55:22.32776313+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:56:22.349459294+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:57:22.373014277+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:58:22.393570956+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T15:59:22.415931625+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:00:22.435251318+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:01:22.453760112+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:02:22.47481149+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:03:22.494641285+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:04:22.514637397+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:05:22.53908081+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:06:22.554718667+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:07:22.575950592+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:08:22.59516599+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:09:22.618204936+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:10:22.640285491+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:10:22.640561155+08:00","level":"ERROR","msg":"Failed to load features, feature will default to disabled","error":"api: failed sending: POST https://api.wandb.ai/graphql giving up after 21 attempt(s): Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:10:22.655245653+08:00","level":"INFO","msg":"stream: created new stream","id":"mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552"}
+{"time":"2025-07-23T16:10:22.65537423+08:00","level":"INFO","msg":"stream: started","id":"mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552"}
+{"time":"2025-07-23T16:10:22.655404089+08:00","level":"INFO","msg":"writer: Do: started","stream_id":"mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552"}
+{"time":"2025-07-23T16:10:22.655443978+08:00","level":"INFO","msg":"sender: started","stream_id":"mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552"}
+{"time":"2025-07-23T16:10:22.655438278+08:00","level":"INFO","msg":"handler: started","stream_id":"mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552"}
+{"time":"2025-07-23T16:10:22.674547549+08:00","level":"WARN","msg":"runupserter: server does not expand metric globs but the x_server_side_expand_glob_metrics setting is set; ignoring"}
+{"time":"2025-07-23T16:10:22.677445015+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:10:24.855217555+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:10:29.31752334+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:10:37.829357598+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:10:54.257058229+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:11:33.270872917+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:12:33.290922758+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:13:33.309386134+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:14:33.328175182+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:15:33.345597329+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:16:33.368160115+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:17:33.384016261+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:18:33.403106053+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:19:33.424551514+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:20:22.660752778+08:00","level":"WARN","msg":"sender: taking a long time","seconds":600.000167592,"work":"RunUpdateWork; Control(mailbox_slot:\"ld5863kv3ux3\" connection_id:\"1(@)\")"}
+{"time":"2025-07-23T16:20:33.446092222+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:21:33.4685356+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:22:33.488932883+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:23:33.508398277+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:24:33.527479449+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:25:33.544379368+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:26:33.564775821+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:26:33.565144643+08:00","level":"ERROR","msg":"runupserter: failed to init run","error":"api: failed sending: POST https://api.wandb.ai/graphql giving up after 21 attempt(s): Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-23T16:26:33.565396718+08:00","level":"INFO","msg":"sender: succeeded after taking longer than expected","seconds":970.90488677,"work":"RunUpdateWork; Control(mailbox_slot:\"ld5863kv3ux3\" connection_id:\"1(@)\")"}
+{"time":"2025-07-29T06:43:53.318006912+08:00","level":"INFO","msg":"stream: closing","id":"mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552"}
+{"time":"2025-07-29T06:43:53.338897634+08:00","level":"ERROR","msg":"sender: uploadConfigFile: stream: no run"}
+{"time":"2025-07-29T06:43:53.394096172+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:43:55.875989199+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:44:00.555698012+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:44:09.89131533+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:44:27.958200974+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:45:05.127170144+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:46:05.147917701+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:47:05.170266993+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:48:05.191985969+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:49:05.214191483+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:50:05.231554686+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:51:05.248024048+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:52:05.267259389+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:53:05.288484475+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:54:05.304448647+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:55:05.327261179+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:56:05.351203715+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:57:05.372446132+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:58:05.395235895+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T06:59:05.417328624+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T07:00:05.43906279+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T07:00:05.439375173+08:00","level":"ERROR","msg":"runfiles: CreateRunFiles returned error: api: failed sending: POST https://api.wandb.ai/graphql giving up after 21 attempt(s): Post \"https://api.wandb.ai/graphql\": proxyconnect tcp: dial tcp: lookup http on 21.19.172.23:53: no such host"}
+{"time":"2025-07-29T07:00:05.485814133+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
+{"time":"2025-07-29T07:00:05.486040248+08:00","level":"INFO","msg":"handler: closed","stream_id":"mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552"}
+{"time":"2025-07-29T07:00:05.486115986+08:00","level":"INFO","msg":"sender: closed","stream_id":"mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552"}
+{"time":"2025-07-29T07:00:05.486103247+08:00","level":"INFO","msg":"writer: Close: closed","stream_id":"mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552"}
+{"time":"2025-07-29T07:00:05.486725823+08:00","level":"INFO","msg":"stream: closed","id":"mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552"}
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/run-20250723_155414-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552/logs/debug.log b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/run-20250723_155414-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552/logs/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..a3f349e5f98c0379797ba2d97e67c38438cfcd7b
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/run-20250723_155414-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552/logs/debug.log
@@ -0,0 +1,15 @@
+2025-07-23 15:54:14,964 INFO MainThread:647798 [wandb_setup.py:_flush():80] Current SDK version is 0.21.0
+2025-07-23 15:54:14,964 INFO MainThread:647798 [wandb_setup.py:_flush():80] Configure stats pid to 647798
+2025-07-23 15:54:14,965 INFO MainThread:647798 [wandb_setup.py:_flush():80] Loading settings from /root/.config/wandb/settings
+2025-07-23 15:54:14,965 INFO MainThread:647798 [wandb_setup.py:_flush():80] Loading settings from /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/dataset/Selection/pretrain-linear-moe/flame/wandb/settings
+2025-07-23 15:54:14,965 INFO MainThread:647798 [wandb_setup.py:_flush():80] Loading settings from environment variables
+2025-07-23 15:54:14,965 INFO MainThread:647798 [wandb_init.py:setup_run_log_directory():703] Logging user logs to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/run-20250723_155414-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552/logs/debug.log
+2025-07-23 15:54:14,965 INFO MainThread:647798 [wandb_init.py:setup_run_log_directory():704] Logging internal logs to /mnt/nanjingcephfs/project_wx-rec-alg-bdc-exp/bwzheng/yulan/hyw/pretrain-linear-moe/flame/exp/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/run-20250723_155414-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552/logs/debug-internal.log
+2025-07-23 15:54:14,965 INFO MainThread:647798 [wandb_init.py:init():830] calling init triggers
+2025-07-23 15:54:14,965 INFO MainThread:647798 [wandb_init.py:init():835] wandb.init called with sweep_config: {}
+config: {'_wandb': {}}
+2025-07-23 15:54:14,965 INFO MainThread:647798 [wandb_init.py:init():871] starting backend
+2025-07-23 15:54:17,123 INFO MainThread:647798 [wandb_init.py:init():874] sending inform_init request
+2025-07-23 15:54:17,174 INFO MainThread:647798 [wandb_init.py:init():882] backend started and connected
+2025-07-23 15:54:17,175 INFO MainThread:647798 [wandb_init.py:init():953] updated telemetry
+2025-07-23 15:54:17,331 INFO MainThread:647798 [wandb_init.py:init():977] communicating run to backend with 90.0 second timeout
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/run-20250723_155414-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552/run-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552.wandb b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/run-20250723_155414-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552/run-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552.wandb
new file mode 100644
index 0000000000000000000000000000000000000000..61a7e352c0cb0ceb3ba4bfbd7b9880d7b1b88c47
Binary files /dev/null and b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tb/20250723-1553/wandb/run-20250723_155414-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552/run-mamba2-mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2-202507231552.wandb differ
diff --git a/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tokenizer.json b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b667161bc937dfaed6f27e9e7849a664c3e869b3
--- /dev/null
+++ b/mamba2_6_1_340M.json-ctx8192-steps95366-lr3e-4-decay_typelinear-decay_ratio1-bs8-nn1-gas2/tokenizer.json
@@ -0,0 +1,268063 @@
+{
+ "version": "1.0",
+ "truncation": null,
+ "padding": null,
+ "added_tokens": [
+ {
+ "id": 0,
+ "content": "",
+ "single_word": false,
+ "lstrip": false,
+ "rstrip": false,
+ "normalized": false,
+ "special": true
+ },
+ {
+ "id": 1,
+ "content": "",
+ "single_word": false,
+ "lstrip": false,
+ "rstrip": false,
+ "normalized": false,
+ "special": true
+ },
+ {
+ "id": 2,
+ "content": "",
+ "single_word": false,
+ "lstrip": false,
+ "rstrip": false,
+ "normalized": false,
+ "special": true
+ }
+ ],
+ "normalizer": {
+ "type": "Sequence",
+ "normalizers": [
+ {
+ "type": "Prepend",
+ "prepend": "▁"
+ },
+ {
+ "type": "Replace",
+ "pattern": {
+ "String": " "
+ },
+ "content": "▁"
+ }
+ ]
+ },
+ "pre_tokenizer": null,
+ "post_processor": {
+ "type": "TemplateProcessing",
+ "single": [
+ {
+ "SpecialToken": {
+ "id": "",
+ "type_id": 0
+ }
+ },
+ {
+ "Sequence": {
+ "id": "A",
+ "type_id": 0
+ }
+ }
+ ],
+ "pair": [
+ {
+ "SpecialToken": {
+ "id": "",
+ "type_id": 0
+ }
+ },
+ {
+ "Sequence": {
+ "id": "A",
+ "type_id": 0
+ }
+ },
+ {
+ "SpecialToken": {
+ "id": "",
+ "type_id": 1
+ }
+ },
+ {
+ "Sequence": {
+ "id": "B",
+ "type_id": 1
+ }
+ }
+ ],
+ "special_tokens": {
+ "": {
+ "id": "",
+ "ids": [
+ 1
+ ],
+ "tokens": [
+ ""
+ ]
+ }
+ }
+ },
+ "decoder": {
+ "type": "Sequence",
+ "decoders": [
+ {
+ "type": "Replace",
+ "pattern": {
+ "String": "▁"
+ },
+ "content": " "
+ },
+ {
+ "type": "ByteFallback"
+ },
+ {
+ "type": "Fuse"
+ },
+ {
+ "type": "Strip",
+ "content": " ",
+ "start": 1,
+ "stop": 0
+ }
+ ]
+ },
+ "model": {
+ "type": "BPE",
+ "dropout": null,
+ "unk_token": "",
+ "continuing_subword_prefix": null,
+ "end_of_word_suffix": null,
+ "fuse_unk": true,
+ "byte_fallback": true,
+ "ignore_merges": false,
+ "vocab": {
+ "": 0,
+ "": 1,
+ "": 2,
+ "<0x00>": 3,
+ "<0x01>": 4,
+ "<0x02>": 5,
+ "<0x03>": 6,
+ "<0x04>": 7,
+ "<0x05>": 8,
+ "<0x06>": 9,
+ "<0x07>": 10,
+ "<0x08>": 11,
+ "<0x09>": 12,
+ "<0x0A>": 13,
+ "<0x0B>": 14,
+ "<0x0C>": 15,
+ "<0x0D>": 16,
+ "<0x0E>": 17,
+ "<0x0F>": 18,
+ "<0x10>": 19,
+ "<0x11>": 20,
+ "<0x12>": 21,
+ "<0x13>": 22,
+ "<0x14>": 23,
+ "<0x15>": 24,
+ "<0x16>": 25,
+ "<0x17>": 26,
+ "<0x18>": 27,
+ "<0x19>": 28,
+ "<0x1A>": 29,
+ "<0x1B>": 30,
+ "<0x1C>": 31,
+ "<0x1D>": 32,
+ "<0x1E>": 33,
+ "<0x1F>": 34,
+ "<0x20>": 35,
+ "<0x21>": 36,
+ "<0x22>": 37,
+ "<0x23>": 38,
+ "<0x24>": 39,
+ "<0x25>": 40,
+ "<0x26>": 41,
+ "<0x27>": 42,
+ "<0x28>": 43,
+ "<0x29>": 44,
+ "<0x2A>": 45,
+ "<0x2B>": 46,
+ "<0x2C>": 47,
+ "<0x2D>": 48,
+ "<0x2E>": 49,
+ "<0x2F>": 50,
+ "<0x30>": 51,
+ "<0x31>": 52,
+ "<0x32>": 53,
+ "<0x33>": 54,
+ "<0x34>": 55,
+ "<0x35>": 56,
+ "<0x36>": 57,
+ "<0x37>": 58,
+ "<0x38>": 59,
+ "<0x39>": 60,
+ "<0x3A>": 61,
+ "<0x3B>": 62,
+ "<0x3C>": 63,
+ "<0x3D>": 64,
+ "<0x3E>": 65,
+ "<0x3F>": 66,
+ "<0x40>": 67,
+ "<0x41>": 68,
+ "<0x42>": 69,
+ "<0x43>": 70,
+ "<0x44>": 71,
+ "<0x45>": 72,
+ "<0x46>": 73,
+ "<0x47>": 74,
+ "<0x48>": 75,
+ "<0x49>": 76,
+ "<0x4A>": 77,
+ "<0x4B>": 78,
+ "<0x4C>": 79,
+ "<0x4D>": 80,
+ "<0x4E>": 81,
+ "<0x4F>": 82,
+ "<0x50>": 83,
+ "<0x51>": 84,
+ "<0x52>": 85,
+ "<0x53>": 86,
+ "<0x54>": 87,
+ "<0x55>": 88,
+ "<0x56>": 89,
+ "<0x57>": 90,
+ "<0x58>": 91,
+ "<0x59>": 92,
+ "<0x5A>": 93,
+ "<0x5B>": 94,
+ "<0x5C>": 95,
+ "<0x5D>": 96,
+ "<0x5E>": 97,
+ "<0x5F>": 98,
+ "<0x60>": 99,
+ "<0x61>": 100,
+ "<0x62>": 101,
+ "<0x63>": 102,
+ "<0x64>": 103,
+ "<0x65>": 104,
+ "<0x66>": 105,
+ "<0x67>": 106,
+ "<0x68>": 107,
+ "<0x69>": 108,
+ "<0x6A>": 109,
+ "<0x6B>": 110,
+ "<0x6C>": 111,
+ "<0x6D>": 112,
+ "<0x6E>": 113,
+ "<0x6F>": 114,
+ "<0x70>": 115,
+ "<0x71>": 116,
+ "<0x72>": 117,
+ "<0x73>": 118,
+ "<0x74>": 119,
+ "<0x75>": 120,
+ "<0x76>": 121,
+ "<0x77>": 122,
+ "<0x78>": 123,
+ "<0x79>": 124,
+ "<0x7A>": 125,
+ "<0x7B>": 126,
+ "<0x7C>": 127,
+ "<0x7D>": 128,
+ "<0x7E>": 129,
+ "<0x7F>": 130,
+ "<0x80>": 131,
+ "<0x81>": 132,
+ "<0x82>": 133,
+ "<0x83>": 134,
+ "<0x84>": 135,
+ "<0x85>": 136,
+ "<0x86>": 137,
+ "<0x87>": 138,
+ "<0x88>": 139,
+ "<0x89>": 140,
+ "<0x8A>": 141,
+ "<0x8B>": 142,
+ "<0x8C>": 143,
+ "<0x8D>": 144,
+ "<0x8E>": 145,
+ "<0x8F>": 146,
+ "<0x90>": 147,
+ "<0x91>": 148,
+ "<0x92>": 149,
+ "<0x93>": 150,
+ "<0x94>": 151,
+ "<0x95>": 152,
+ "<0x96>": 153,
+ "<0x97>": 154,
+ "<0x98>": 155,
+ "<0x99>": 156,
+ "<0x9A>": 157,
+ "<0x9B>": 158,
+ "<0x9C>": 159,
+ "<0x9D>": 160,
+ "<0x9E>": 161,
+ "<0x9F>": 162,
+ "<0xA0>": 163,
+ "<0xA1>": 164,
+ "<0xA2>": 165,
+ "<0xA3>": 166,
+ "<0xA4>": 167,
+ "<0xA5>": 168,
+ "<0xA6>": 169,
+ "<0xA7>": 170,
+ "<0xA8>": 171,
+ "<0xA9>": 172,
+ "<0xAA>": 173,
+ "<0xAB>": 174,
+ "<0xAC>": 175,
+ "<0xAD>": 176,
+ "<0xAE>": 177,
+ "<0xAF>": 178,
+ "<0xB0>": 179,
+ "<0xB1>": 180,
+ "<0xB2>": 181,
+ "<0xB3>": 182,
+ "<0xB4>": 183,
+ "<0xB5>": 184,
+ "<0xB6>": 185,
+ "<0xB7>": 186,
+ "<0xB8>": 187,
+ "<0xB9>": 188,
+ "<0xBA>": 189,
+ "<0xBB>": 190,
+ "<0xBC>": 191,
+ "<0xBD>": 192,
+ "<0xBE>": 193,
+ "<0xBF>": 194,
+ "<0xC0>": 195,
+ "<0xC1>": 196,
+ "<0xC2>": 197,
+ "<0xC3>": 198,
+ "<0xC4>": 199,
+ "<0xC5>": 200,
+ "<0xC6>": 201,
+ "<0xC7>": 202,
+ "<0xC8>": 203,
+ "<0xC9>": 204,
+ "<0xCA>": 205,
+ "<0xCB>": 206,
+ "<0xCC>": 207,
+ "<0xCD>": 208,
+ "<0xCE>": 209,
+ "<0xCF>": 210,
+ "<0xD0>": 211,
+ "<0xD1>": 212,
+ "<0xD2>": 213,
+ "<0xD3>": 214,
+ "<0xD4>": 215,
+ "<0xD5>": 216,
+ "<0xD6>": 217,
+ "<0xD7>": 218,
+ "<0xD8>": 219,
+ "<0xD9>": 220,
+ "<0xDA>": 221,
+ "<0xDB>": 222,
+ "<0xDC>": 223,
+ "<0xDD>": 224,
+ "<0xDE>": 225,
+ "<0xDF>": 226,
+ "<0xE0>": 227,
+ "<0xE1>": 228,
+ "<0xE2>": 229,
+ "<0xE3>": 230,
+ "<0xE4>": 231,
+ "<0xE5>": 232,
+ "<0xE6>": 233,
+ "<0xE7>": 234,
+ "<0xE8>": 235,
+ "<0xE9>": 236,
+ "<0xEA>": 237,
+ "<0xEB>": 238,
+ "<0xEC>": 239,
+ "<0xED>": 240,
+ "<0xEE>": 241,
+ "<0xEF>": 242,
+ "<0xF0>": 243,
+ "<0xF1>": 244,
+ "<0xF2>": 245,
+ "<0xF3>": 246,
+ "<0xF4>": 247,
+ "<0xF5>": 248,
+ "<0xF6>": 249,
+ "<0xF7>": 250,
+ "<0xF8>": 251,
+ "<0xF9>": 252,
+ "<0xFA>": 253,
+ "<0xFB>": 254,
+ "<0xFC>": 255,
+ "<0xFD>": 256,
+ "<0xFE>": 257,
+ "<0xFF>": 258,
+ "▁▁": 259,
+ "▁▁▁▁": 260,
+ "▁t": 261,
+ "in": 262,
+ "er": 263,
+ "▁a": 264,
+ "he": 265,
+ "on": 266,
+ "re": 267,
+ "▁s": 268,
+ "en": 269,
+ "at": 270,
+ "or": 271,
+ "▁the": 272,
+ "▁▁▁▁▁▁▁▁": 273,
+ "es": 274,
+ "▁w": 275,
+ "an": 276,
+ "▁c": 277,
+ "is": 278,
+ "it": 279,
+ "ou": 280,
+ "▁d": 281,
+ "al": 282,
+ "ar": 283,
+ "▁p": 284,
+ "▁f": 285,
+ "ed": 286,
+ "▁b": 287,
+ "ing": 288,
+ "▁o": 289,
+ "▁m": 290,
+ "le": 291,
+ "nd": 292,
+ "as": 293,
+ "ic": 294,
+ "▁h": 295,
+ "ion": 296,
+ "▁in": 297,
+ "▁to": 298,
+ "et": 299,
+ "om": 300,
+ "el": 301,
+ "▁of": 302,
+ "st": 303,
+ "▁and": 304,
+ "▁l": 305,
+ "▁th": 306,
+ "▁n": 307,
+ "ent": 308,
+ "il": 309,
+ "ct": 310,
+ "ro": 311,
+ "▁re": 312,
+ "id": 313,
+ "am": 314,
+ "▁I": 315,
+ "ad": 316,
+ "▁e": 317,
+ "▁S": 318,
+ "▁g": 319,
+ "▁T": 320,
+ "im": 321,
+ "ot": 322,
+ "ac": 323,
+ "ur": 324,
+ "▁(": 325,
+ "ig": 326,
+ "▁=": 327,
+ "ol": 328,
+ "ut": 329,
+ "▁A": 330,
+ "se": 331,
+ "▁u": 332,
+ "ve": 333,
+ "▁C": 334,
+ "if": 335,
+ "ow": 336,
+ "▁y": 337,
+ "ch": 338,
+ "ay": 339,
+ "▁de": 340,
+ "▁st": 341,
+ "▁|": 342,
+ "ver": 343,
+ ");": 344,
+ "▁\"": 345,
+ "ly": 346,
+ "▁be": 347,
+ "**": 348,
+ "▁is": 349,
+ "od": 350,
+ "▁M": 351,
+ "ation": 352,
+ "ul": 353,
+ "▁for": 354,
+ "▁▁▁▁▁": 355,
+ "▁on": 356,
+ "ag": 357,
+ "ce": 358,
+ "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁": 359,
+ "ter": 360,
+ "ir": 361,
+ "th": 362,
+ "▁v": 363,
+ "qu": 364,
+ "▁B": 365,
+ "em": 366,
+ "▁P": 367,
+ "▁you": 368,
+ "▁that": 369,
+ "un": 370,
+ "▁{": 371,
+ "ith": 372,
+ "ri": 373,
+ "est": 374,
+ "ab": 375,
+ "--": 376,
+ "ap": 377,
+ "▁it": 378,
+ "▁con": 379,
+ "ate": 380,
+ "us": 381,
+ "▁H": 382,
+ "um": 383,
+ "▁D": 384,
+ "os": 385,
+ "pe": 386,
+ "▁-": 387,
+ "▁wh": 388,
+ "▁al": 389,
+ "▁as": 390,
+ "and": 391,
+ "ist": 392,
+ "▁L": 393,
+ "▁W": 394,
+ "▁with": 395,
+ "▁an": 396,
+ "ere": 397,
+ "▁*": 398,
+ "▁R": 399,
+ "▁he": 400,
+ "▁F": 401,
+ "oc": 402,
+ "▁was": 403,
+ "ers": 404,
+ "ke": 405,
+ "out": 406,
+ "ht": 407,
+ "▁r": 408,
+ "ess": 409,
+ "op": 410,
+ "res": 411,
+ "ie": 412,
+ "▁E": 413,
+ "▁\\": 414,
+ "▁The": 415,
+ "end": 416,
+ "ld": 417,
+ "▁N": 418,
+ "ort": 419,
+ "▁G": 420,
+ "//": 421,
+ "▁#": 422,
+ "our": 423,
+ "te": 424,
+ "ill": 425,
+ "ain": 426,
+ "▁se": 427,
+ "▁▁▁▁▁▁": 428,
+ "▁$": 429,
+ "▁pro": 430,
+ "ore": 431,
+ "▁com": 432,
+ "ame": 433,
+ "tr": 434,
+ "▁ne": 435,
+ "rom": 436,
+ "ub": 437,
+ "▁at": 438,
+ "▁ex": 439,
+ "ant": 440,
+ "ue": 441,
+ "▁or": 442,
+ "▁}": 443,
+ "art": 444,
+ "ction": 445,
+ "▁k": 446,
+ "pt": 447,
+ "nt": 448,
+ "iv": 449,
+ "de": 450,
+ "▁O": 451,
+ "pl": 452,
+ "urn": 453,
+ "ight": 454,
+ "all": 455,
+ "▁this": 456,
+ "ser": 457,
+ "ave": 458,
+ "▁not": 459,
+ "▁are": 460,
+ "▁j": 461,
+ "▁le": 462,
+ "iz": 463,
+ "▁'": 464,
+ "age": 465,
+ "ment": 466,
+ "▁tr": 467,
+ "ack": 468,
+ "ust": 469,
+ "()": 470,
+ "->": 471,
+ "ity": 472,
+ "ine": 473,
+ "ould": 474,
+ "▁J": 475,
+ "og": 476,
+ "▁from": 477,
+ "▁we": 478,
+ "ell": 479,
+ "▁sh": 480,
+ "▁en": 481,
+ "ure": 482,
+ "port": 483,
+ "▁ch": 484,
+ "ne": 485,
+ "▁by": 486,
+ "per": 487,
+ "ard": 488,
+ "ass": 489,
+ "ge": 490,
+ "ak": 491,
+ "are": 492,
+ "ok": 493,
+ "av": 494,
+ "ive": 495,
+ "ff": 496,
+ "ies": 497,
+ "ath": 498,
+ "turn": 499,
+ "▁U": 500,
+ "int": 501,
+ "----": 502,
+ "▁im": 503,
+ "ost": 504,
+ "ial": 505,
+ "▁have": 506,
+ "ind": 507,
+ "ip": 508,
+ "ans": 509,
+ "xt": 510,
+ "▁do": 511,
+ "cl": 512,
+ "▁if": 513,
+ "con": 514,
+ "ia": 515,
+ "▁his": 516,
+ "ult": 517,
+ "rou": 518,
+ "▁su": 519,
+ "ra": 520,
+ "▁un": 521,
+ "able": 522,
+ "▁<": 523,
+ "▁K": 524,
+ "ome": 525,
+ "▁qu": 526,
+ "get": 527,
+ "▁me": 528,
+ "ast": 529,
+ "ect": 530,
+ "▁##": 531,
+ "to": 532,
+ "▁cl": 533,
+ "▁ab": 534,
+ "ice": 535,
+ "ire": 536,
+ "ber": 537,
+ "one": 538,
+ "ich": 539,
+ "hen": 540,
+ "▁can": 541,
+ "▁Th": 542,
+ "▁la": 543,
+ "▁all": 544,
+ "ime": 545,
+ "ile": 546,
+ "ide": 547,
+ "\",": 548,
+ "▁pl": 549,
+ "▁V": 550,
+ "ru": 551,
+ "orm": 552,
+ "▁had": 553,
+ "ud": 554,
+ "ase": 555,
+ "ord": 556,
+ "),": 557,
+ "▁▁▁▁▁▁▁▁▁▁▁▁": 558,
+ "▁her": 559,
+ "▁In": 560,
+ "ace": 561,
+ "▁but": 562,
+ "ata": 563,
+ "::": 564,
+ "****": 565,
+ "ong": 566,
+ "▁&": 567,
+ "..": 568,
+ "▁▁▁▁▁▁▁▁▁▁▁▁▁": 569,
+ "ite": 570,
+ "ype": 571,
+ "act": 572,
+ "ode": 573,
+ "▁your": 574,
+ "▁out": 575,
+ "▁go": 576,
+ "lic": 577,
+ "ally": 578,
+ "▁so": 579,
+ "ork": 580,
+ "au": 581,
+ "▁up": 582,
+ "▁_": 583,
+ "ll": 584,
+ "==": 585,
+ "▁my": 586,
+ "pp": 587,
+ "cc": 588,
+ "▁//": 589,
+ "▁they": 590,
+ "gh": 591,
+ "▁us": 592,
+ "ib": 593,
+ "ions": 594,
+ "ach": 595,
+ "ens": 596,
+ "▁ar": 597,
+ "ob": 598,
+ "elf": 599,
+ "ook": 600,
+ "ated": 601,
+ "ang": 602,
+ "ign": 603,
+ "▁return": 604,
+ "▁res": 605,
+ "ck": 606,
+ "ous": 607,
+ "ст": 608,
+ ").": 609,
+ "▁п": 610,
+ ".\"": 611,
+ "на": 612,
+ "▁i": 613,
+ "ail": 614,
+ "ep": 615,
+ "▁ad": 616,
+ "ance": 617,
+ "(\"": 618,
+ "▁**": 619,
+ "ther": 620,
+ "ake": 621,
+ "▁will": 622,
+ "▁comp": 623,
+ "▁one": 624,
+ "▁get": 625,
+ "ov": 626,
+ "▁Y": 627,
+ "ary": 628,
+ "ock": 629,
+ "▁she": 630,
+ "che": 631,
+ "ft": 632,
+ "▁new": 633,
+ "▁des": 634,
+ "▁li": 635,
+ "ence": 636,
+ "▁sa": 637,
+ "ress": 638,
+ "▁el": 639,
+ "▁und": 640,
+ "eg": 641,
+ "fer": 642,
+ "ry": 643,
+ "ear": 644,
+ "ose": 645,
+ "very": 646,
+ "',": 647,
+ "▁+": 648,
+ "▁в": 649,
+ "▁He": 650,
+ "ublic": 651,
+ "▁their": 652,
+ "ize": 653,
+ "▁were": 654,
+ "ink": 655,
+ "own": 656,
+ "In": 657,
+ "{\\": 658,
+ "▁has": 659,
+ "▁per": 660,
+ "▁It": 661,
+ "▁St": 662,
+ "her": 663,
+ "ject": 664,
+ "ра": 665,
+ "ild": 666,
+ "so": 667,
+ "▁sp": 668,
+ "ни": 669,
+ "du": 670,
+ "row": 671,
+ "alue": 672,
+ "set": 673,
+ "form": 674,
+ "com": 675,
+ "▁man": 676,
+ "ont": 677,
+ "ull": 678,
+ "▁cont": 679,
+ "▁more": 680,
+ "ick": 681,
+ "▁would": 682,
+ "▁ev": 683,
+ "▁about": 684,
+ "ition": 685,
+ "▁z": 686,
+ "ound": 687,
+ "ree": 688,
+ "▁Ch": 689,
+ "▁which": 690,
+ "io": 691,
+ "();": 692,
+ "▁who": 693,
+ "err": 694,
+ "ory": 695,
+ "ount": 696,
+ "ations": 697,
+ "▁с": 698,
+ "ring": 699,
+ "": 700,
+ "▁fe": 701,
+ "ко": 702,
+ "но": 703,
+ "▁dis": 704,
+ "ma": 705,
+ "▁them": 706,
+ "▁any": 707,
+ "▁no": 708,
+ "--------": 709,
+ "▁pre": 710,
+ "▁te": 711,
+ "▁ro": 712,
+ "▁him": 713,
+ "▁:": 714,
+ "up": 715,
+ "▁int": 716,
+ "▁ag": 717,
+ "St": 718,
+ "ark": 719,
+ "ex": 720,
+ "ph": 721,
+ "ient": 722,
+ "ely": 723,
+ "▁pr": 724,
+ "ER": 725,
+ "▁import": 726,
+ "▁time": 727,
+ "ро": 728,
+ "pro": 729,
+ "User": 730,
+ "lo": 731,
+ "▁/": 732,
+ "▁[": 733,
+ "ors": 734,
+ "=\"": 735,
+ "▁there": 736,
+ "▁like": 737,
+ "old": 738,
+ "▁when": 739,
+ "vers": 740,
+ "▁some": 741,
+ "ings": 742,
+ "))": 743,
+ "▁part": 744,
+ "ical": 745,
+ "▁fun": 746,
+ "▁kn": 747,
+ "ays": 748,
+ "ier": 749,
+ "▁been": 750,
+ "ove": 751,
+ "▁sc": 752,
+ "ian": 753,
+ "▁over": 754,
+ "iel": 755,
+ "▁▁▁▁▁▁▁▁▁▁": 756,
+ "▁pe": 757,
+ "rib": 758,
+ "put": 759,
+ "ec": 760,
+ "eth": 761,
+ "aram": 762,
+ "app": 763,
+ "▁–": 764,
+ "▁stat": 765,
+ "pon": 766,
+ "▁what": 767,
+ "ption": 768,
+ "we": 769,
+ "ade": 770,
+ "▁work": 771,
+ "text": 772,
+ "▁said": 773,
+ "▁###": 774,
+ "IN": 775,
+ "▁just": 776,
+ "irst": 777,
+ "▁into": 778,
+ "▁const": 779,
+ "ource": 780,
+ "tt": 781,
+ "ps": 782,
+ "pr": 783,
+ "erv": 784,
+ "itt": 785,
+ "ug": 786,
+ "_{": 787,
+ "ents": 788,
+ "ish": 789,
+ "ener": 790,
+ "▁inter": 791,
+ "ple": 792,
+ "oll": 793,
+ "mer": 794,
+ "ater": 795,
+ "ool": 796,
+ "ef": 797,
+ "▁public": 798,
+ "▁other": 799,
+ "ре": 800,
+ "▁def": 801,
+ "▁@": 802,
+ "го": 803,
+ "oint": 804,
+ "▁off": 805,
+ "oid": 806,
+ "return": 807,
+ "▁set": 808,
+ "wo": 809,
+ "fter": 810,
+ "sh": 811,
+ "********": 812,
+ "▁our": 813,
+ "riv": 814,
+ "iss": 815,
+ "▁We": 816,
+ "ng": 817,
+ "▁ob": 818,
+ "ss": 819,
+ "gr": 820,
+ "▁than": 821,
+ "pect": 822,
+ "ied": 823,
+ "sc": 824,
+ "iew": 825,
+ "der": 826,
+ "yst": 827,
+ "ev": 828,
+ "▁could": 829,
+ "ann": 830,
+ "enc": 831,
+ "ON": 832,
+ "ix": 833,
+ "anc": 834,
+ "▁also": 835,
+ "reat": 836,
+ "▁am": 837,
+ "▁bec": 838,
+ "▁и": 839,
+ "ual": 840,
+ "pec": 841,
+ "▁.": 842,
+ "▁bl": 843,
+ "lect": 844,
+ "ople": 845,
+ "ys": 846,
+ "▁gr": 847,
+ "ict": 848,
+ "ik": 849,
+ "tring": 850,
+ "▁This": 851,
+ "▁back": 852,
+ "▁о": 853,
+ "▁fin": 854,
+ "atch": 855,
+ "Con": 856,
+ "('": 857,
+ "erm": 858,
+ "▁==": 859,
+ "__": 860,
+ "name": 861,
+ ",\"": 862,
+ "▁did": 863,
+ "ise": 864,
+ "▁only": 865,
+ "ruct": 866,
+ "les": 867,
+ "▁then": 868,
+ "ause": 869,
+ "ва": 870,
+ "▁its": 871,
+ "rit": 872,
+ "▁know": 873,
+ "ield": 874,
+ "▁class": 875,
+ "▁>": 876,
+ "▁em": 877,
+ "▁$\\": 878,
+ "▁year": 879,
+ "wn": 880,
+ "},": 881,
+ "▁del": 882,
+ "ale": 883,
+ "ty": 884,
+ "fig": 885,
+ "sp": 886,
+ "hed": 887,
+ "round": 888,
+ "ew": 889,
+ "▁di": 890,
+ "▁der": 891,
+ "ри": 892,
+ "red": 893,
+ "this": 894,
+ "let": 895,
+ "RE": 896,
+ "ax": 897,
+ "fr": 898,
+ "essage": 899,
+ "ough": 900,
+ "▁comm": 901,
+ "fo": 902,
+ "uch": 903,
+ "oy": 904,
+ "▁people": 905,
+ "ystem": 906,
+ "▁first": 907,
+ "▁function": 908,
+ "ange": 909,
+ "▁how": 910,
+ "▁et": 911,
+ "ah": 912,
+ "▁look": 913,
+ "то": 914,
+ "und": 915,
+ "▁under": 916,
+ "ка": 917,
+ "▁!": 918,
+ "ray": 919,
+ "ST": 920,
+ "ific": 921,
+ "ли": 922,
+ "read": 923,
+ "▁bet": 924,
+ "ious": 925,
+ "arg": 926,
+ "▁need": 927,
+ "math": 928,
+ "▁на": 929,
+ "ert": 930,
+ "▁op": 931,
+ "▁acc": 932,
+ "Pro": 933,
+ "▁est": 934,
+ "▁Un": 935,
+ "▁ent": 936,
+ "▁rec": 937,
+ "▁use": 938,
+ "ен": 939,
+ "▁par": 940,
+ "az": 941,
+ "▁д": 942,
+ "▁Wh": 943,
+ "self": 944,
+ "▁ke": 945,
+ "та": 946,
+ "▁want": 947,
+ "▁end": 948,
+ "▁don": 949,
+ "ek": 950,
+ "ren": 951,
+ "Name": 952,
+ "▁=>": 953,
+ "▁app": 954,
+ "▁que": 955,
+ "igh": 956,
+ "▁bu": 957,
+ "equ": 958,
+ "vel": 959,
+ "▁act": 960,
+ "cre": 961,
+ "AT": 962,
+ "▁var": 963,
+ "cess": 964,
+ "====": 965,
+ "Ex": 966,
+ "▁add": 967,
+ "▁mod": 968,
+ "ung": 969,
+ "▁where": 970,
+ "ning": 971,
+ "▁fl": 972,
+ "als": 973,
+ "tern": 974,
+ "}}": 975,
+ "▁Al": 976,
+ "▁pos": 977,
+ "ank": 978,
+ "▁ap": 979,
+ "eng": 980,
+ "▁“": 981,
+ "ble": 982,
+ "▁reg": 983,
+ "^{": 984,
+ "▁She": 985,
+ "▁*/": 986,
+ "ude": 987,
+ "add": 988,
+ "▁two": 989,
+ "▁col": 990,
+ "▁sm": 991,
+ "air": 992,
+ "▁may": 993,
+ "fore": 994,
+ "▁You": 995,
+ "rough": 996,
+ "▁che": 997,
+ "▁att": 998,
+ "oth": 999,
+ "ла": 1000,
+ "▁co": 1001,
+ "ates": 1002,
+ "▁rem": 1003,
+ "ood": 1004,
+ "Type": 1005,
+ "led": 1006,
+ "ful": 1007,
+ "▁self": 1008,
+ "of": 1009,
+ "▁Ar": 1010,
+ "que": 1011,
+ "▁every": 1012,
+ "ref": 1013,
+ "The": 1014,
+ "▁And": 1015,
+ "▁rel": 1016,
+ "OR": 1017,
+ "Id": 1018,
+ "▁even": 1019,
+ "EN": 1020,
+ "▁hand": 1021,
+ "ait": 1022,
+ "▁should": 1023,
+ "▁after": 1024,
+ "▁dif": 1025,
+ "ght": 1026,
+ "ife": 1027,
+ "ator": 1028,
+ "ash": 1029,
+ "ribut": 1030,
+ "umber": 1031,
+ "▁see": 1032,
+ "ms": 1033,
+ "▁call": 1034,
+ "yn": 1035,
+ "dd": 1036,
+ "▁es": 1037,
+ "▁make": 1038,
+ "other": 1039,
+ "▁—": 1040,
+ "\");": 1041,
+ "str": 1042,
+ "▁long": 1043,
+ "lement": 1044,
+ "▁wor": 1045,
+ "its": 1046,
+ "▁If": 1047,
+ "alse": 1048,
+ "ль": 1049,
+ "ward": 1050,
+ "▁по": 1051,
+ "val": 1052,
+ "ons": 1053,
+ "▁Z": 1054,
+ "▁now": 1055,
+ "data": 1056,
+ "amp": 1057,
+ "ense": 1058,
+ "▁through": 1059,
+ "▁down": 1060,
+ "att": 1061,
+ "▁static": 1062,
+ "ics": 1063,
+ "##": 1064,
+ "pos": 1065,
+ "▁void": 1066,
+ "aw": 1067,
+ "oun": 1068,
+ "▁way": 1069,
+ "ible": 1070,
+ "vent": 1071,
+ "ower": 1072,
+ "▁think": 1073,
+ "ts": 1074,
+ "*/": 1075,
+ "▁again": 1076,
+ "ating": 1077,
+ "те": 1078,
+ "ner": 1079,
+ "▁most": 1080,
+ "line": 1081,
+ "ym": 1082,
+ "▁sub": 1083,
+ "erson": 1084,
+ "▁requ": 1085,
+ "AL": 1086,
+ "AR": 1087,
+ "abel": 1088,
+ "ond": 1089,
+ "));": 1090,
+ "▁Se": 1091,
+ "▁But": 1092,
+ "alk": 1093,
+ "▁An": 1094,
+ "new": 1095,
+ "▁because": 1096,
+ "ger": 1097,
+ "ular": 1098,
+ "roup": 1099,
+ "ta": 1100,
+ "...": 1101,
+ "▁cons": 1102,
+ "▁right": 1103,
+ "▁fr": 1104,
+ "be": 1105,
+ "ily": 1106,
+ "ки": 1107,
+ "▁ph": 1108,
+ "ead": 1109,
+ "?\"": 1110,
+ "▁gu": 1111,
+ "▁else": 1112,
+ "▁som": 1113,
+ "rent": 1114,
+ "co": 1115,
+ "ement": 1116,
+ "▁str": 1117,
+ "ault": 1118,
+ "▁з": 1119,
+ "ло": 1120,
+ "sert": 1121,
+ "var": 1122,
+ "type": 1123,
+ "▁Com": 1124,
+ "ле": 1125,
+ "ins": 1126,
+ "me": 1127,
+ "way": 1128,
+ "ident": 1129,
+ "▁prov": 1130,
+ "▁м": 1131,
+ "▁true": 1132,
+ "▁Pro": 1133,
+ "fl": 1134,
+ "▁sl": 1135,
+ "▁As": 1136,
+ "}\\": 1137,
+ "ID": 1138,
+ "ues": 1139,
+ "▁inst": 1140,
+ "▁name": 1141,
+ "ox": 1142,
+ "▁)": 1143,
+ "li": 1144,
+ "ames": 1145,
+ "Res": 1146,
+ "▁sur": 1147,
+ "param": 1148,
+ "▁start": 1149,
+ "aj": 1150,
+ "SE": 1151,
+ "ask": 1152,
+ "IT": 1153,
+ "String": 1154,
+ "▁ass": 1155,
+ "▁play": 1156,
+ "ting": 1157,
+ "ton": 1158,
+ "▁before": 1159,
+ "▁pol": 1160,
+ "arch": 1161,
+ "▁well": 1162,
+ "Com": 1163,
+ "any": 1164,
+ "olog": 1165,
+ "▁err": 1166,
+ "▁these": 1167,
+ "ars": 1168,
+ "eb": 1169,
+ "▁br": 1170,
+ "▁incl": 1171,
+ "▁hel": 1172,
+ "ern": 1173,
+ "ody": 1174,
+ "во": 1175,
+ "▁ind": 1176,
+ "----------------": 1177,
+ "▁data": 1178,
+ "▁good": 1179,
+ "LE": 1180,
+ "],": 1181,
+ "▁av": 1182,
+ "▁ac": 1183,
+ "ider": 1184,
+ "не": 1185,
+ "▁Q": 1186,
+ "▁min": 1187,
+ "▁much": 1188,
+ "ci": 1189,
+ "els": 1190,
+ "▁cur": 1191,
+ "▁value": 1192,
+ "ery": 1193,
+ "uf": 1194,
+ "▁loc": 1195,
+ "reak": 1196,
+ "ative": 1197,
+ "imes": 1198,
+ "Cl": 1199,
+ "▁,": 1200,
+ "▁ser": 1201,
+ "▁die": 1202,
+ "▁trans": 1203,
+ "▁result": 1204,
+ "ext": 1205,
+ "▁aut": 1206,
+ "land": 1207,
+ "▁&&": 1208,
+ "Ch": 1209,
+ "ten": 1210,
+ "}$": 1211,
+ "▁type": 1212,
+ "cond": 1213,
+ "ices": 1214,
+ "▁very": 1215,
+ "▁own": 1216,
+ "▁fil": 1217,
+ "ities": 1218,
+ "▁produ": 1219,
+ "▁read": 1220,
+ "▁form": 1221,
+ "▁case": 1222,
+ "ather": 1223,
+ "ти": 1224,
+ "да": 1225,
+ "ер": 1226,
+ "Th": 1227,
+ "aut": 1228,
+ "▁spec": 1229,
+ "ij": 1230,
+ "bl": 1231,
+ "ility": 1232,
+ "▁é": 1233,
+ "▁er": 1234,
+ "▁does": 1235,
+ "▁here": 1236,
+ "the": 1237,
+ "ures": 1238,
+ "▁%": 1239,
+ "min": 1240,
+ "▁null": 1241,
+ "rap": 1242,
+ "\")": 1243,
+ "rr": 1244,
+ "List": 1245,
+ "right": 1246,
+ "▁User": 1247,
+ "UL": 1248,
+ "ational": 1249,
+ "▁being": 1250,
+ "AN": 1251,
+ "sk": 1252,
+ "▁car": 1253,
+ "ole": 1254,
+ "▁dist": 1255,
+ "plic": 1256,
+ "ollow": 1257,
+ "▁pres": 1258,
+ "▁such": 1259,
+ "ream": 1260,
+ "ince": 1261,
+ "gan": 1262,
+ "▁For": 1263,
+ "\":": 1264,
+ "son": 1265,
+ "rivate": 1266,
+ "▁years": 1267,
+ "▁serv": 1268,
+ "▁made": 1269,
+ "def": 1270,
+ ";\r": 1271,
+ "▁gl": 1272,
+ "▁bel": 1273,
+ "▁list": 1274,
+ "▁cor": 1275,
+ "▁det": 1276,
+ "ception": 1277,
+ "egin": 1278,
+ "▁б": 1279,
+ "▁char": 1280,
+ "trans": 1281,
+ "▁fam": 1282,
+ "▁!=": 1283,
+ "ouse": 1284,
+ "▁dec": 1285,
+ "ica": 1286,
+ "▁many": 1287,
+ "aking": 1288,
+ "▁à": 1289,
+ "▁sim": 1290,
+ "ages": 1291,
+ "uff": 1292,
+ "ased": 1293,
+ "man": 1294,
+ "▁Sh": 1295,
+ "iet": 1296,
+ "irect": 1297,
+ "▁Re": 1298,
+ "▁differ": 1299,
+ "▁find": 1300,
+ "ethod": 1301,
+ "▁\r": 1302,
+ "ines": 1303,
+ "▁inv": 1304,
+ "▁point": 1305,
+ "▁They": 1306,
+ "▁used": 1307,
+ "ctions": 1308,
+ "▁still": 1309,
+ "ió": 1310,
+ "ined": 1311,
+ "▁while": 1312,
+ "It": 1313,
+ "ember": 1314,
+ "▁say": 1315,
+ "▁help": 1316,
+ "▁cre": 1317,
+ "▁x": 1318,
+ "▁Tr": 1319,
+ "ument": 1320,
+ "▁sk": 1321,
+ "ought": 1322,
+ "ually": 1323,
+ "message": 1324,
+ "▁Con": 1325,
+ "▁mon": 1326,
+ "ared": 1327,
+ "work": 1328,
+ "):": 1329,
+ "ister": 1330,
+ "arn": 1331,
+ "ized": 1332,
+ "Data": 1333,
+ "orn": 1334,
+ "▁head": 1335,
+ "DE": 1336,
+ "▁Le": 1337,
+ "▁person": 1338,
+ "ments": 1339,
+ "ength": 1340,
+ "▁false": 1341,
+ "▁med": 1342,
+ "▁De": 1343,
+ "ache": 1344,
+ "ited": 1345,
+ "▁let": 1346,
+ "▁show": 1347,
+ "▁same": 1348,
+ "uss": 1349,
+ "▁gener": 1350,
+ "▁у": 1351,
+ "cur": 1352,
+ "▁real": 1353,
+ "ced": 1354,
+ "\">": 1355,
+ "struct": 1356,
+ "begin": 1357,
+ "cept": 1358,
+ "▁bo": 1359,
+ "ired": 1360,
+ "▁Fr": 1361,
+ "▁stud": 1362,
+ "dev": 1363,
+ "Ar": 1364,
+ "(\\": 1365,
+ "▁Cl": 1366,
+ "ween": 1367,
+ "▁too": 1368,
+ "▁test": 1369,
+ "▁day": 1370,
+ "oh": 1371,
+ "▁follow": 1372,
+ "ature": 1373,
+ "ze": 1374,
+ "ien": 1375,
+ "reg": 1376,
+ "ces": 1377,
+ "uring": 1378,
+ "amb": 1379,
+ "ina": 1380,
+ "cri": 1381,
+ "▁ed": 1382,
+ "SS": 1383,
+ "uck": 1384,
+ "▁/*": 1385,
+ "CT": 1386,
+ "▁There": 1387,
+ "▁take": 1388,
+ "par": 1389,
+ "ule": 1390,
+ "cal": 1391,
+ "for": 1392,
+ "****************": 1393,
+ "source": 1394,
+ "▁those": 1395,
+ "col": 1396,
+ "▁eff": 1397,
+ "mod": 1398,
+ "cont": 1399,
+ "}{": 1400,
+ "▁around": 1401,
+ "press": 1402,
+ "by": 1403,
+ "▁going": 1404,
+ "ponse": 1405,
+ "▁С": 1406,
+ "▁line": 1407,
+ "date": 1408,
+ "code": 1409,
+ "['": 1410,
+ "▁life": 1411,
+ "ason": 1412,
+ "▁using": 1413,
+ "▁val": 1414,
+ "▁du": 1415,
+ "yp": 1416,
+ "▁▁▁▁▁▁▁▁▁▁▁▁▁▁": 1417,
+ "▁On": 1418,
+ "▁found": 1419,
+ "olut": 1420,
+ "']": 1421,
+ "arent": 1422,
+ "▁string": 1423,
+ "▁met": 1424,
+ "▁wr": 1425,
+ "ush": 1426,
+ "string": 1427,
+ "size": 1428,
+ "▁ver": 1429,
+ "▁each": 1430,
+ "value": 1431,
+ "▁last": 1432,
+ "▁got": 1433,
+ "ven": 1434,
+ "back": 1435,
+ "Set": 1436,
+ "ey": 1437,
+ "rol": 1438,
+ "▁cr": 1439,
+ "thing": 1440,
+ "ret": 1441,
+ "és": 1442,
+ "ism": 1443,
+ "▁between": 1444,
+ "Ob": 1445,
+ "ething": 1446,
+ "mp": 1447,
+ "▁lo": 1448,
+ "ats": 1449,
+ "▁New": 1450,
+ "ви": 1451,
+ "ado": 1452,
+ "dex": 1453,
+ "ди": 1454,
+ "▁pass": 1455,
+ "wh": 1456,
+ "▁den": 1457,
+ "Get": 1458,
+ "apt": 1459,
+ "▁ask": 1460,
+ "▁sup": 1461,
+ "Value": 1462,
+ "ны": 1463,
+ "▁try": 1464,
+ "lation": 1465,
+ "day": 1466,
+ "ness": 1467,
+ "ets": 1468,
+ "▁exper": 1469,
+ "Tr": 1470,
+ "▁Mar": 1471,
+ "serv": 1472,
+ "br": 1473,
+ "▁number": 1474,
+ "inal": 1475,
+ "cent": 1476,
+ "/*": 1477,
+ "not": 1478,
+ "ional": 1479,
+ "▁final": 1480,
+ "')": 1481,
+ "▁run": 1482,
+ "over": 1483,
+ "▁never": 1484,
+ "uc": 1485,
+ "▁high": 1486,
+ "yle": 1487,
+ "▁ins": 1488,
+ "▁best": 1489,
+ "ittle": 1490,
+ "ric": 1491,
+ "▁sign": 1492,
+ "▁dem": 1493,
+ "iness": 1494,
+ "gy": 1495,
+ "▁war": 1496,
+ "ished": 1497,
+ "▁giv": 1498,
+ "key": 1499,
+ "▁X": 1500,
+ "($": 1501,
+ "▁child": 1502,
+ "less": 1503,
+ "ways": 1504,
+ "incl": 1505,
+ "rop": 1506,
+ "raw": 1507,
+ "://": 1508,
+ "▁«": 1509,
+ "no": 1510,
+ "indow": 1511,
+ "fe": 1512,
+ "riend": 1513,
+ "▁les": 1514,
+ "▁los": 1515,
+ "file": 1516,
+ "formation": 1517,
+ "ccess": 1518,
+ "▁В": 1519,
+ "na": 1520,
+ "▁il": 1521,
+ "ision": 1522,
+ "ler": 1523,
+ "▁art": 1524,
+ "Cont": 1525,
+ "▁world": 1526,
+ "▁turn": 1527,
+ "▁really": 1528,
+ "▁Ex": 1529,
+ "ма": 1530,
+ "▁П": 1531,
+ "ters": 1532,
+ "arget": 1533,
+ "Err": 1534,
+ "▁happ": 1535,
+ "time": 1536,
+ "▁So": 1537,
+ "div": 1538,
+ "▁didn": 1539,
+ "ada": 1540,
+ "oot": 1541,
+ "})": 1542,
+ "▁sch": 1543,
+ "▁cle": 1544,
+ "▁something": 1545,
+ "().": 1546,
+ "▁cour": 1547,
+ "ever": 1548,
+ "ants": 1549,
+ "▁?": 1550,
+ "To": 1551,
+ "▁`": 1552,
+ "try": 1553,
+ "ux": 1554,
+ "ais": 1555,
+ "ross": 1556,
+ "hip": 1557,
+ "▁rep": 1558,
+ "label": 1559,
+ "▁both": 1560,
+ "*,": 1561,
+ "ott": 1562,
+ "ми": 1563,
+ "ane": 1564,
+ "▁open": 1565,
+ "ww": 1566,
+ "▁come": 1567,
+ "▁ext": 1568,
+ "rem": 1569,
+ "_{\\": 1570,
+ "▁old": 1571,
+ "ched": 1572,
+ "._": 1573,
+ "ME": 1574,
+ "ify": 1575,
+ "gg": 1576,
+ "Col": 1577,
+ "view": 1578,
+ "▁bus": 1579,
+ "▁must": 1580,
+ "▁different": 1581,
+ "log": 1582,
+ "ists": 1583,
+ "roll": 1584,
+ "ai": 1585,
+ "▁за": 1586,
+ "▁system": 1587,
+ "ivers": 1588,
+ "atus": 1589,
+ "ote": 1590,
+ "med": 1591,
+ "].": 1592,
+ "akes": 1593,
+ "RO": 1594,
+ "▁cent": 1595,
+ "gram": 1596,
+ "▁private": 1597,
+ "▁great": 1598,
+ "\";": 1599,
+ "opy": 1600,
+ "▁feel": 1601,
+ "▁How": 1602,
+ "////": 1603,
+ "IC": 1604,
+ "▁dr": 1605,
+ "ains": 1606,
+ "lock": 1607,
+ "En": 1608,
+ "▁Sch": 1609,
+ "▁mat": 1610,
+ "▁home": 1611,
+ "perty": 1612,
+ "test": 1613,
+ "loc": 1614,
+ "▁wom": 1615,
+ "sw": 1616,
+ "arly": 1617,
+ "▁En": 1618,
+ "▁ко": 1619,
+ "den": 1620,
+ "ста": 1621,
+ "▁а": 1622,
+ "eter": 1623,
+ "▁includ": 1624,
+ "ULL": 1625,
+ "▁mem": 1626,
+ "▁po": 1627,
+ "▁little": 1628,
+ "▁arg": 1629,
+ "▁},": 1630,
+ "include": 1631,
+ "eta": 1632,
+ "▁place": 1633,
+ "idth": 1634,
+ "ustom": 1635,
+ "▁||": 1636,
+ "▁tem": 1637,
+ "ried": 1638,
+ "▁fact": 1639,
+ "ience": 1640,
+ "▁Pl": 1641,
+ "opt": 1642,
+ "ele": 1643,
+ "go": 1644,
+ "AC": 1645,
+ "inter": 1646,
+ "========": 1647,
+ "(),": 1648,
+ "ots": 1649,
+ "ral": 1650,
+ "ique": 1651,
+ "aving": 1652,
+ "ml": 1653,
+ "▁thought": 1654,
+ "frac": 1655,
+ "▁care": 1656,
+ "());": 1657,
+ "▁put": 1658,
+ "▁might": 1659,
+ "▁Amer": 1660,
+ "▁(!": 1661,
+ "ample": 1662,
+ "alth": 1663,
+ "▁few": 1664,
+ "▁state": 1665,
+ "sub": 1666,
+ "▁Or": 1667,
+ "];": 1668,
+ "▁size": 1669,
+ "▁Sp": 1670,
+ "▁without": 1671,
+ "▁poss": 1672,
+ "eq": 1673,
+ "play": 1674,
+ "▁expect": 1675,
+ "▁second": 1676,
+ "▁String": 1677,
+ "uild": 1678,
+ "▁next": 1679,
+ "++": 1680,
+ "requ": 1681,
+ "▁All": 1682,
+ "▁men": 1683,
+ "▁When": 1684,
+ "iter": 1685,
+ "ament": 1686,
+ "net": 1687,
+ "▁К": 1688,
+ "ron": 1689,
+ "aint": 1690,
+ "▁Is": 1691,
+ "ве": 1692,
+ "pend": 1693,
+ "translation": 1694,
+ "▁го": 1695,
+ "че": 1696,
+ "▁van": 1697,
+ "▁another": 1698,
+ "▁ret": 1699,
+ "▁La": 1700,
+ "Mod": 1701,
+ "ION": 1702,
+ "list": 1703,
+ "▁post": 1704,
+ "da": 1705,
+ "ware": 1706,
+ "▁word": 1707,
+ "Error": 1708,
+ "▁seem": 1709,
+ "▁contin": 1710,
+ "atic": 1711,
+ "▁three": 1712,
+ "Object": 1713,
+ "▁partic": 1714,
+ "$.": 1715,
+ "▁mark": 1716,
+ "▁vis": 1717,
+ "rc": 1718,
+ "▁sw": 1719,
+ "ptions": 1720,
+ "▁break": 1721,
+ "▁things": 1722,
+ "ute": 1723,
+ "ui": 1724,
+ "▁That": 1725,
+ "urs": 1726,
+ "gl": 1727,
+ "ру": 1728,
+ "▁file": 1729,
+ "use": 1730,
+ "igned": 1731,
+ "part": 1732,
+ "Un": 1733,
+ "▁equ": 1734,
+ "(&": 1735,
+ "▁lead": 1736,
+ "rm": 1737,
+ "ained": 1738,
+ "▁Be": 1739,
+ "path": 1740,
+ "▁small": 1741,
+ "ager": 1742,
+ "▁always": 1743,
+ "▁El": 1744,
+ "▁order": 1745,
+ "▁ey": 1746,
+ "▁won": 1747,
+ "ape": 1748,
+ "▁left": 1749,
+ "ava": 1750,
+ "item": 1751,
+ "hor": 1752,
+ "▁away": 1753,
+ "bb": 1754,
+ "fun": 1755,
+ "▁Ind": 1756,
+ "mb": 1757,
+ "▁struct": 1758,
+ "▁process": 1759,
+ "▁support": 1760,
+ ");\r": 1761,
+ "ión": 1762,
+ "LO": 1763,
+ "▁oper": 1764,
+ "UT": 1765,
+ "▁·": 1766,
+ "PE": 1767,
+ "load": 1768,
+ "off": 1769,
+ "▁No": 1770,
+ "ives": 1771,
+ "ican": 1772,
+ "▁ve": 1773,
+ "action": 1774,
+ "';": 1775,
+ "▁vo": 1776,
+ "$,": 1777,
+ "▁Gr": 1778,
+ "pre": 1779,
+ "ny": 1780,
+ "aining": 1781,
+ "ior": 1782,
+ "init": 1783,
+ "lection": 1784,
+ "arm": 1785,
+ "umn": 1786,
+ "ags": 1787,
+ "ци": 1788,
+ "ско": 1789,
+ "version": 1790,
+ "▁To": 1791,
+ "▁ref": 1792,
+ "stand": 1793,
+ "▁At": 1794,
+ "ift": 1795,
+ "▁ein": 1796,
+ "face": 1797,
+ "bo": 1798,
+ "ified": 1799,
+ "ved": 1800,
+ "sum": 1801,
+ "une": 1802,
+ "ital": 1803,
+ "ump": 1804,
+ "comm": 1805,
+ "▁mov": 1806,
+ "elt": 1807,
+ "▁von": 1808,
+ "velop": 1809,
+ "ctor": 1810,
+ "head": 1811,
+ "cle": 1812,
+ "▁build": 1813,
+ "inc": 1814,
+ ".'": 1815,
+ "bs": 1816,
+ "info": 1817,
+ "chn": 1818,
+ "▁week": 1819,
+ "▁book": 1820,
+ "HE": 1821,
+ "bar": 1822,
+ "icense": 1823,
+ "▁What": 1824,
+ "▁quest": 1825,
+ "urch": 1826,
+ "ato": 1827,
+ "left": 1828,
+ "▁mar": 1829,
+ "▁top": 1830,
+ "FF": 1831,
+ "▁friend": 1832,
+ "▁beh": 1833,
+ "▁field": 1834,
+ "▁against": 1835,
+ "ract": 1836,
+ "ization": 1837,
+ "user": 1838,
+ "chen": 1839,
+ "▁keep": 1840,
+ "AD": 1841,
+ "itor": 1842,
+ "▁non": 1843,
+ "ird": 1844,
+ "ope": 1845,
+ "▁rest": 1846,
+ "▁dev": 1847,
+ "▁__": 1848,
+ "▁una": 1849,
+ "▁term": 1850,
+ "IS": 1851,
+ "▁pop": 1852,
+ "rist": 1853,
+ "▁since": 1854,
+ "ves": 1855,
+ "▁hard": 1856,
+ "pi": 1857,
+ "util": 1858,
+ "▁soc": 1859,
+ "ene": 1860,
+ "Exception": 1861,
+ "▁local": 1862,
+ "▁direct": 1863,
+ "▁sure": 1864,
+ "▁bro": 1865,
+ "▁da": 1866,
+ "▁": 1867,
+ "▁current": 1868,
+ "':": 1869,
+ "Wh": 1870,
+ "▁information": 1871,
+ "▁ide": 1872,
+ "▁better": 1873,
+ "Text": 1874,
+ "raph": 1875,
+ "▁stand": 1876,
+ "▁check": 1877,
+ "▁к": 1878,
+ "▁na": 1879,
+ "((": 1880,
+ "outh": 1881,
+ "aps": 1882,
+ "▁unt": 1883,
+ "bf": 1884,
+ "▁conf": 1885,
+ "▁spe": 1886,
+ "itle": 1887,
+ "▁Col": 1888,
+ "class": 1889,
+ "ural": 1890,
+ "bers": 1891,
+ "MA": 1892,
+ "ession": 1893,
+ "▁М": 1894,
+ "Info": 1895,
+ "▁Br": 1896,
+ "▁eas": 1897,
+ "ervice": 1898,
+ "aus": 1899,
+ "ari": 1900,
+ "по": 1901,
+ "▁coun": 1902,
+ "де": 1903,
+ "())": 1904,
+ "ling": 1905,
+ "ED": 1906,
+ "ably": 1907,
+ "▁pat": 1908,
+ "org": 1909,
+ "▁id": 1910,
+ "▁г": 1911,
+ "▁tell": 1912,
+ "lex": 1913,
+ "▁allow": 1914,
+ "reen": 1915,
+ "my": 1916,
+ "▁consider": 1917,
+ "▁team": 1918,
+ "lease": 1919,
+ "htt": 1920,
+ "▁Pr": 1921,
+ "/**": 1922,
+ "▁sing": 1923,
+ "Requ": 1924,
+ "Re": 1925,
+ "ides": 1926,
+ "ches": 1927,
+ "▁object": 1928,
+ "ially": 1929,
+ "By": 1930,
+ "ся": 1931,
+ "ided": 1932,
+ "▁free": 1933,
+ "▁proble": 1934,
+ "cite": 1935,
+ "▁);": 1936,
+ "ission": 1937,
+ "▁during": 1938,
+ "▁--": 1939,
+ "ither": 1940,
+ "ля": 1941,
+ "▁leg": 1942,
+ "▁sit": 1943,
+ "ically": 1944,
+ "▁key": 1945,
+ "leg": 1946,
+ "tra": 1947,
+ "▁mom": 1948,
+ "▁expl": 1949,
+ "▁develop": 1950,
+ "▁event": 1951,
+ "▁NULL": 1952,
+ "ohn": 1953,
+ "▁///": 1954,
+ "▁business": 1955,
+ "ча": 1956,
+ "▁prof": 1957,
+ "error": 1958,
+ "▁por": 1959,
+ "▁commun": 1960,
+ "Ind": 1961,
+ "ium": 1962,
+ "Test": 1963,
+ "▁Ad": 1964,
+ "ouble": 1965,
+ "▁son": 1966,
+ "rite": 1967,
+ "ready": 1968,
+ "▁{\r": 1969,
+ "▁thing": 1970,
+ "ня": 1971,
+ "▁Ph": 1972,
+ "ped": 1973,
+ "сь": 1974,
+ "ived": 1975,
+ "You": 1976,
+ "arl": 1977,
+ "const": 1978,
+ "../": 1979,
+ "Se": 1980,
+ "Sh": 1981,
+ "▁power": 1982,
+ "ribute": 1983,
+ "▁My": 1984,
+ "▁talk": 1985,
+ "itch": 1986,
+ "▁called": 1987,
+ "▁came": 1988,
+ "▁belie": 1989,
+ "UR": 1990,
+ "Add": 1991,
+ "▁Res": 1992,
+ "aster": 1993,
+ "ella": 1994,
+ "obal": 1995,
+ "▁until": 1996,
+ "▁hum": 1997,
+ "CO": 1998,
+ "ately": 1999,
+ "####": 2000,
+ "public": 2001,
+ "[]": 2002,
+ "▁room": 2003,
+ "len": 2004,
+ "▁family": 2005,
+ "por": 2006,
+ "▁program": 2007,
+ "▁hist": 2008,
+ "▁mus": 2009,
+ "arge": 2010,
+ "oney": 2011,
+ "Im": 2012,
+ "else": 2013,
+ "ails": 2014,
+ "af": 2015,
+ "▁love": 2016,
+ "är": 2017,
+ "ases": 2018,
+ "pha": 2019,
+ "ours": 2020,
+ "dis": 2021,
+ "map": 2022,
+ "iver": 2023,
+ "ör": 2024,
+ "▁Bl": 2025,
+ "ateg": 2026,
+ "state": 2027,
+ "State": 2028,
+ "ertain": 2029,
+ "▁effect": 2030,
+ "print": 2031,
+ "▁big": 2032,
+ "index": 2033,
+ "▁pub": 2034,
+ "vert": 2035,
+ "ero": 2036,
+ "md": 2037,
+ "▁method": 2038,
+ "▁game": 2039,
+ "ries": 2040,
+ "lete": 2041,
+ "Item": 2042,
+ "ING": 2043,
+ "resent": 2044,
+ "ality": 2045,
+ "pty": 2046,
+ "ley": 2047,
+ "ocument": 2048,
+ "▁beg": 2049,
+ "TR": 2050,
+ "}.": 2051,
+ "▁school": 2052,
+ "hes": 2053,
+ "до": 2054,
+ "▁lot": 2055,
+ "▁took": 2056,
+ "▁adv": 2057,
+ "▁cap": 2058,
+ "MP": 2059,
+ "unk": 2060,
+ "▁light": 2061,
+ "▁later": 2062,
+ ".,": 2063,
+ "Key": 2064,
+ "itions": 2065,
+ "▁enough": 2066,
+ "▁/**": 2067,
+ "▁went": 2068,
+ "ão": 2069,
+ "▁though": 2070,
+ "▁group": 2071,
+ "▁mean": 2072,
+ "ски": 2073,
+ "AP": 2074,
+ "▁num": 2075,
+ "▁cond": 2076,
+ "ні": 2077,
+ "▁given": 2078,
+ "▁why": 2079,
+ "▁rece": 2080,
+ "▁side": 2081,
+ "▁far": 2082,
+ "Context": 2083,
+ "ме": 2084,
+ "▁log": 2085,
+ "View": 2086,
+ "▁<<": 2087,
+ "fil": 2088,
+ "aces": 2089,
+ "ency": 2090,
+ "oad": 2091,
+ "ered": 2092,
+ "▁product": 2093,
+ "ET": 2094,
+ "▁param": 2095,
+ "▁prote": 2096,
+ "tes": 2097,
+ "Time": 2098,
+ "je": 2099,
+ "olution": 2100,
+ "▁ра": 2101,
+ "▁month": 2102,
+ "ference": 2103,
+ "▁appe": 2104,
+ "▁face": 2105,
+ "ened": 2106,
+ "tract": 2107,
+ "▁less": 2108,
+ "AS": 2109,
+ "ée": 2110,
+ "▁give": 2111,
+ "▁kind": 2112,
+ "▁count": 2113,
+ "count": 2114,
+ "▁stop": 2115,
+ "▁gover": 2116,
+ "ka": 2117,
+ "▁error": 2118,
+ "ences": 2119,
+ "▁mil": 2120,
+ "alf": 2121,
+ "ync": 2122,
+ "vious": 2123,
+ "ho": 2124,
+ "▁night": 2125,
+ "era": 2126,
+ "▁про": 2127,
+ "▁sol": 2128,
+ "men": 2129,
+ "▁water": 2130,
+ "ering": 2131,
+ "▁lim": 2132,
+ "Param": 2133,
+ "▁house": 2134,
+ "▁System": 2135,
+ "▁pay": 2136,
+ "▁:=": 2137,
+ "uro": 2138,
+ "oci": 2139,
+ "zy": 2140,
+ "▁already": 2141,
+ ",\\": 2142,
+ "length": 2143,
+ "▁si": 2144,
+ "▁interest": 2145,
+ "aff": 2146,
+ "cted": 2147,
+ "ention": 2148,
+ "▁до": 2149,
+ "ume": 2150,
+ "▁appro": 2151,
+ "bre": 2152,
+ "IG": 2153,
+ "▁throw": 2154,
+ "mathcal": 2155,
+ "irl": 2156,
+ "▁prom": 2157,
+ "oss": 2158,
+ "▁request": 2159,
+ "equation": 2160,
+ "ology": 2161,
+ "mit": 2162,
+ "▁pack": 2163,
+ "ino": 2164,
+ "array": 2165,
+ "za": 2166,
+ "til": 2167,
+ "UN": 2168,
+ "▁present": 2169,
+ "▁organ": 2170,
+ "File": 2171,
+ "▁orig": 2172,
+ "▁full": 2173,
+ "istr": 2174,
+ "▁flo": 2175,
+ "hr": 2176,
+ "▁assert": 2177,
+ "ards": 2178,
+ "url": 2179,
+ "enn": 2180,
+ "sl": 2181,
+ "▁А": 2182,
+ "▁cho": 2183,
+ "▁level": 2184,
+ "OT": 2185,
+ "word": 2186,
+ "▁body": 2187,
+ "▁user": 2188,
+ "ía": 2189,
+ "Qu": 2190,
+ "▁main": 2191,
+ "AB": 2192,
+ "ploy": 2193,
+ "Event": 2194,
+ "▁super": 2195,
+ "oken": 2196,
+ "▁Н": 2197,
+ "As": 2198,
+ "thers": 2199,
+ "мо": 2200,
+ "ку": 2201,
+ "▁days": 2202,
+ "▁done": 2203,
+ "▁view": 2204,
+ "side": 2205,
+ "си": 2206,
+ "');": 2207,
+ "▁vol": 2208,
+ "▁tot": 2209,
+ "case": 2210,
+ "▁aff": 2211,
+ "Request": 2212,
+ "▁Man": 2213,
+ "\\\\": 2214,
+ "▁John": 2215,
+ "▁Б": 2216,
+ "orth": 2217,
+ "▁je": 2218,
+ "▁une": 2219,
+ "la": 2220,
+ "[\"": 2221,
+ "field": 2222,
+ "▁US": 2223,
+ "ico": 2224,
+ "▁perform": 2225,
+ "ailable": 2226,
+ "Config": 2227,
+ "Or": 2228,
+ "▁model": 2229,
+ "ales": 2230,
+ "▁create": 2231,
+ "▁ann": 2232,
+ "ances": 2233,
+ "IL": 2234,
+ "ination": 2235,
+ "▁Im": 2236,
+ "ante": 2237,
+ "ana": 2238,
+ "ан": 2239,
+ "▁told": 2240,
+ "config": 2241,
+ "\"]": 2242,
+ "met": 2243,
+ "lt": 2244,
+ "▁text": 2245,
+ "▁May": 2246,
+ "▁org": 2247,
+ "▁port": 2248,
+ "Pl": 2249,
+ "ently": 2250,
+ "▁door": 2251,
+ "US": 2252,
+ "▁(*": 2253,
+ "kt": 2254,
+ "ES": 2255,
+ "ential": 2256,
+ "▁iss": 2257,
+ "▁inc": 2258,
+ "Node": 2259,
+ "ively": 2260,
+ "▁asked": 2261,
+ "irt": 2262,
+ "▁Te": 2263,
+ "▁report": 2264,
+ "▁chang": 2265,
+ "сти": 2266,
+ "▁along": 2267,
+ "▁change": 2268,
+ "Size": 2269,
+ "▁ever": 2270,
+ "▁occ": 2271,
+ "ury": 2272,
+ "▁mind": 2273,
+ "order": 2274,
+ "point": 2275,
+ "сто": 2276,
+ "▁whe": 2277,
+ "▁important": 2278,
+ "des": 2279,
+ "▁Not": 2280,
+ "▁writ": 2281,
+ "▁eyes": 2282,
+ "▁desc": 2283,
+ "most": 2284,
+ "ks": 2285,
+ "▁bit": 2286,
+ "▁▁▁": 2287,
+ "▁success": 2288,
+ "ть": 2289,
+ "бо": 2290,
+ "core": 2291,
+ "}(": 2292,
+ "▁array": 2293,
+ "lin": 2294,
+ "lish": 2295,
+ "▁following": 2296,
+ "Field": 2297,
+ "ids": 2298,
+ "hing": 2299,
+ "▁cal": 2300,
+ "Is": 2301,
+ "aring": 2302,
+ "lev": 2303,
+ "alt": 2304,
+ "CH": 2305,
+ "▁dé": 2306,
+ "alpha": 2307,
+ "▁four": 2308,
+ "▁law": 2309,
+ "▁се": 2310,
+ "iron": 2311,
+ "▁disc": 2312,
+ "се": 2313,
+ "ken": 2314,
+ "node": 2315,
+ "▁Par": 2316,
+ "▁Eng": 2317,
+ "▁move": 2318,
+ "▁License": 2319,
+ "cul": 2320,
+ "ione": 2321,
+ ")$": 2322,
+ "▁tw": 2323,
+ "We": 2324,
+ "sel": 2325,
+ "▁With": 2326,
+ "▁once": 2327,
+ "Service": 2328,
+ "bol": 2329,
+ "ured": 2330,
+ "ida": 2331,
+ "▁Qu": 2332,
+ "▁grow": 2333,
+ "▁conne": 2334,
+ "EX": 2335,
+ "▁htt": 2336,
+ "▁};": 2337,
+ "▁walk": 2338,
+ "▁init": 2339,
+ "nal": 2340,
+ "ender": 2341,
+ "cription": 2342,
+ "mber": 2343,
+ "lected": 2344,
+ "po": 2345,
+ "▁nil": 2346,
+ "▁prob": 2347,
+ "чи": 2348,
+ "▁Ste": 2349,
+ "ison": 2350,
+ "ands": 2351,
+ "osed": 2352,
+ "же": 2353,
+ "▁His": 2354,
+ "ür": 2355,
+ "Man": 2356,
+ "Element": 2357,
+ "▁able": 2358,
+ "Index": 2359,
+ "search": 2360,
+ "▁mag": 2361,
+ "ар": 2362,
+ "▁course": 2363,
+ "▁Car": 2364,
+ "▁exp": 2365,
+ "aph": 2366,
+ "▁mit": 2367,
+ "▁doesn": 2368,
+ "▁default": 2369,
+ "/>": 2370,
+ "aim": 2371,
+ "▁service": 2372,
+ "▁within": 2373,
+ "angu": 2374,
+ "▁Д": 2375,
+ "uffer": 2376,
+ "AG": 2377,
+ "▁Do": 2378,
+ "▁incre": 2379,
+ "▁understand": 2380,
+ "}^": 2381,
+ "▁looked": 2382,
+ "gen": 2383,
+ "ailed": 2384,
+ "▁е": 2385,
+ "ayer": 2386,
+ "▁One": 2387,
+ "▁bas": 2388,
+ "▁job": 2389,
+ "mu": 2390,
+ "but": 2391,
+ "elta": 2392,
+ "▁Christ": 2393,
+ "uration": 2394,
+ "▁record": 2395,
+ "▁Univers": 2396,
+ "ivid": 2397,
+ "valid": 2398,
+ "▁Р": 2399,
+ "▁hold": 2400,
+ "▁table": 2401,
+ "ones": 2402,
+ "link": 2403,
+ "▁Ge": 2404,
+ "▁offer": 2405,
+ "ster": 2406,
+ "Form": 2407,
+ "={": 2408,
+ "▁не": 2409,
+ "stance": 2410,
+ "▁govern": 2411,
+ "▁techn": 2412,
+ "▁prim": 2413,
+ "*.": 2414,
+ "cho": 2415,
+ "max": 2416,
+ "▁fore": 2417,
+ "▁Can": 2418,
+ "▁polit": 2419,
+ "ories": 2420,
+ "▁times": 2421,
+ "▁dans": 2422,
+ "▁air": 2423,
+ "▁anything": 2424,
+ "▁sever": 2425,
+ "acy": 2426,
+ "}_": 2427,
+ "He": 2428,
+ "▁least": 2429,
+ "ips": 2430,
+ "ENT": 2431,
+ "do": 2432,
+ "▁от": 2433,
+ "▁cost": 2434,
+ ".”": 2435,
+ "▁children": 2436,
+ "ability": 2437,
+ "But": 2438,
+ "▁path": 2439,
+ "result": 2440,
+ "acter": 2441,
+ "▁element": 2442,
+ "ee": 2443,
+ "▁wait": 2444,
+ "▁money": 2445,
+ "Map": 2446,
+ "td": 2447,
+ "oin": 2448,
+ "iving": 2449,
+ "icht": 2450,
+ "icy": 2451,
+ "sch": 2452,
+ "ste": 2453,
+ "ду": 2454,
+ "ored": 2455,
+ "oud": 2456,
+ "ille": 2457,
+ "ised": 2458,
+ "plication": 2459,
+ "▁custom": 2460,
+ "▁having": 2461,
+ "ponent": 2462,
+ "▁By": 2463,
+ "ules": 2464,
+ "ued": 2465,
+ "atter": 2466,
+ "And": 2467,
+ "itive": 2468,
+ "Def": 2469,
+ "▁moment": 2470,
+ "aterial": 2471,
+ "Class": 2472,
+ "ograph": 2473,
+ "ike": 2474,
+ "▁large": 2475,
+ "▁####": 2476,
+ "▁either": 2477,
+ "duct": 2478,
+ "▁Then": 2479,
+ "▁Gu": 2480,
+ "olean": 2481,
+ "pert": 2482,
+ "▁Get": 2483,
+ "▁Ab": 2484,
+ "▁short": 2485,
+ "On": 2486,
+ "iment": 2487,
+ "▁project": 2488,
+ "cript": 2489,
+ "▁including": 2490,
+ "ния": 2491,
+ "▁making": 2492,
+ "▁someone": 2493,
+ "▁Fl": 2494,
+ "▁sat": 2495,
+ "▁company": 2496,
+ "ocus": 2497,
+ "pu": 2498,
+ "▁God": 2499,
+ "ification": 2500,
+ "No": 2501,
+ "▁sn": 2502,
+ "ano": 2503,
+ "ga": 2504,
+ "▁au": 2505,
+ "▁cou": 2506,
+ "ás": 2507,
+ "ended": 2508,
+ "ту": 2509,
+ "ober": 2510,
+ "▁nothing": 2511,
+ "▁net": 2512,
+ "▁pot": 2513,
+ "▁typ": 2514,
+ "▁item": 2515,
+ "rew": 2516,
+ "Att": 2517,
+ "▁young": 2518,
+ "}\r": 2519,
+ "nder": 2520,
+ "start": 2521,
+ "▁Sc": 2522,
+ "*)": 2523,
+ "▁enc": 2524,
+ "▁women": 2525,
+ "▁looking": 2526,
+ "▁ро": 2527,
+ "▁health": 2528,
+ "Path": 2529,
+ "▁After": 2530,
+ "▁mult": 2531,
+ "▁{\\": 2532,
+ "▁land": 2533,
+ "orld": 2534,
+ "▁Des": 2535,
+ "▁eng": 2536,
+ "input": 2537,
+ "▁Pol": 2538,
+ "\"\"": 2539,
+ "Code": 2540,
+ "▁supp": 2541,
+ "ainer": 2542,
+ "heck": 2543,
+ "▁mor": 2544,
+ "▁mill": 2545,
+ "▁aw": 2546,
+ "fs": 2547,
+ "▁doing": 2548,
+ "tings": 2549,
+ "ades": 2550,
+ "▁toget": 2551,
+ "▁certain": 2552,
+ "▁together": 2553,
+ "CE": 2554,
+ "ideo": 2555,
+ "▁American": 2556,
+ "ony": 2557,
+ "idd": 2558,
+ "II": 2559,
+ "ged": 2560,
+ "ables": 2561,
+ "▁ident": 2562,
+ "iod": 2563,
+ "▁parent": 2564,
+ "For": 2565,
+ "ambda": 2566,
+ "ando": 2567,
+ "=\\": 2568,
+ "aged": 2569,
+ "ending": 2570,
+ "Int": 2571,
+ "▁possible": 2572,
+ "▁со": 2573,
+ "ivity": 2574,
+ "num": 2575,
+ "rt": 2576,
+ "ajor": 2577,
+ "create": 2578,
+ "ride": 2579,
+ "▁knew": 2580,
+ "bit": 2581,
+ "itional": 2582,
+ "▁lik": 2583,
+ "▁Her": 2584,
+ "ension": 2585,
+ "\".": 2586,
+ "oto": 2587,
+ "▁exist": 2588,
+ "aken": 2589,
+ "▁actually": 2590,
+ "ca": 2591,
+ "▁Г": 2592,
+ "хо": 2593,
+ "inn": 2594,
+ "All": 2595,
+ "buf": 2596,
+ "▁Me": 2597,
+ "▁seen": 2598,
+ "ops": 2599,
+ "▁▁▁▁▁▁▁▁▁": 2600,
+ "Not": 2601,
+ "▁control": 2602,
+ "▁respon": 2603,
+ "};": 2604,
+ "ilt": 2605,
+ "isk": 2606,
+ "▁bad": 2607,
+ "▁often": 2608,
+ "▁past": 2609,
+ "aper": 2610,
+ "▁reason": 2611,
+ "eters": 2612,
+ "▁wanted": 2613,
+ "ura": 2614,
+ "table": 2615,
+ "ormal": 2616,
+ "width": 2617,
+ "га": 2618,
+ "ptr": 2619,
+ "▁dest": 2620,
+ "▁design": 2621,
+ "▁sound": 2622,
+ "▁plan": 2623,
+ "▁base": 2624,
+ "hand": 2625,
+ "gs": 2626,
+ "▁says": 2627,
+ "function": 2628,
+ "▁tri": 2629,
+ "mt": 2630,
+ "▁invest": 2631,
+ "▁available": 2632,
+ "ayout": 2633,
+ "▁och": 2634,
+ "▁las": 2635,
+ "illed": 2636,
+ "Val": 2637,
+ "▁ф": 2638,
+ "iety": 2639,
+ "mon": 2640,
+ "Hand": 2641,
+ "Fr": 2642,
+ "iam": 2643,
+ "pace": 2644,
+ "▁Ob": 2645,
+ "▁para": 2646,
+ "▁meet": 2647,
+ "▁sum": 2648,
+ "Message": 2649,
+ "ici": 2650,
+ "▁known": 2651,
+ "▁gen": 2652,
+ "amma": 2653,
+ "arr": 2654,
+ "▁tre": 2655,
+ "oke": 2656,
+ "uth": 2657,
+ "~\\": 2658,
+ "▁experience": 2659,
+ "icle": 2660,
+ "▁Il": 2661,
+ "▁sent": 2662,
+ "▁others": 2663,
+ "▁soft": 2664,
+ "IP": 2665,
+ "▁max": 2666,
+ "ball": 2667,
+ "▁market": 2668,
+ "▁pour": 2669,
+ "pression": 2670,
+ "eps": 2671,
+ "▁saw": 2672,
+ "▁across": 2673,
+ "▁Su": 2674,
+ "Over": 2675,
+ "ние": 2676,
+ "ulation": 2677,
+ "▁Reg": 2678,
+ "▁+=": 2679,
+ "body": 2680,
+ ")\\": 2681,
+ "▁print": 2682,
+ "▁при": 2683,
+ "db": 2684,
+ "ources": 2685,
+ "wards": 2686,
+ "▁black": 2687,
+ "со": 2688,
+ "ili": 2689,
+ "▁Ed": 2690,
+ "▁complet": 2691,
+ "▁single": 2692,
+ "▁IN": 2693,
+ "ached": 2694,
+ "bt": 2695,
+ "▁code": 2696,
+ "▁bool": 2697,
+ "▁area": 2698,
+ "▁require": 2699,
+ "▁problem": 2700,
+ "aced": 2701,
+ "Equ": 2702,
+ "▁config": 2703,
+ "vec": 2704,
+ "ney": 2705,
+ "cy": 2706,
+ "Al": 2707,
+ "▁account": 2708,
+ "ymbol": 2709,
+ "▁ste": 2710,
+ "ges": 2711,
+ "Array": 2712,
+ "empl": 2713,
+ "context": 2714,
+ "Des": 2715,
+ "Result": 2716,
+ "ecut": 2717,
+ "▁target": 2718,
+ "▁getting": 2719,
+ "\"/>": 2720,
+ "ogle": 2721,
+ "▁himself": 2722,
+ "▁wasn": 2723,
+ "▁block": 2724,
+ "▁ant": 2725,
+ "▁York": 2726,
+ "▁become": 2727,
+ "iff": 2728,
+ "ports": 2729,
+ "reate": 2730,
+ "='": 2731,
+ "cd": 2732,
+ "location": 2733,
+ "ет": 2734,
+ "▁access": 2735,
+ "gress": 2736,
+ "ros": 2737,
+ "Up": 2738,
+ "▁working": 2739,
+ "▁Am": 2740,
+ "iqu": 2741,
+ "cer": 2742,
+ "▁((": 2743,
+ "▁Per": 2744,
+ "▁func": 2745,
+ "▁girl": 2746,
+ "▁above": 2747,
+ "pen": 2748,
+ "пи": 2749,
+ "ido": 2750,
+ "▁version": 2751,
+ "TY": 2752,
+ "▁;": 2753,
+ "mary": 2754,
+ "abled": 2755,
+ "annel": 2756,
+ "▁example": 2757,
+ "▁context": 2758,
+ "OP": 2759,
+ "▁red": 2760,
+ "▁cir": 2761,
+ "sm": 2762,
+ "Log": 2763,
+ "▁space": 2764,
+ "▁fut": 2765,
+ "▁Gener": 2766,
+ "ills": 2767,
+ "▁dri": 2768,
+ "_.": 2769,
+ "▁felt": 2770,
+ "▁offic": 2771,
+ "▁===": 2772,
+ "ii": 2773,
+ "▁started": 2774,
+ "▁Т": 2775,
+ "▁});": 2776,
+ "js": 2777,
+ "▁front": 2778,
+ "▁almost": 2779,
+ "irm": 2780,
+ "!\"": 2781,
+ "signed": 2782,
+ "▁yet": 2783,
+ "▁trad": 2784,
+ "ients": 2785,
+ "ama": 2786,
+ "▁input": 2787,
+ "lim": 2788,
+ "па": 2789,
+ "▁ка": 2790,
+ "▁camp": 2791,
+ "ibr": 2792,
+ "fect": 2793,
+ "unt": 2794,
+ "▁half": 2795,
+ "▁cover": 2796,
+ "anguage": 2797,
+ "▁ben": 2798,
+ "ha": 2799,
+ "▁diff": 2800,
+ "_\\": 2801,
+ "▁об": 2802,
+ "])": 2803,
+ "odes": 2804,
+ "hel": 2805,
+ "ios": 2806,
+ "▁О": 2807,
+ "▁mot": 2808,
+ "▁social": 2809,
+ "////////": 2810,
+ "▁stre": 2811,
+ "ground": 2812,
+ "ів": 2813,
+ "object": 2814,
+ "ples": 2815,
+ "reed": 2816,
+ "▁een": 2817,
+ "▁based": 2818,
+ "▁range": 2819,
+ "An": 2820,
+ "urg": 2821,
+ "▁learn": 2822,
+ "▁exc": 2823,
+ "▁imp": 2824,
+ "▁means": 2825,
+ "▁wur": 2826,
+ "ends": 2827,
+ "void": 2828,
+ "▁std": 2829,
+ "▁particular": 2830,
+ "ja": 2831,
+ "▁source": 2832,
+ "default": 2833,
+ "py": 2834,
+ "▁als": 2835,
+ "scri": 2836,
+ "status": 2837,
+ "▁story": 2838,
+ "▁begin": 2839,
+ "▁position": 2840,
+ "▁special": 2841,
+ "php": 2842,
+ "▁bar": 2843,
+ "▁pract": 2844,
+ "call": 2845,
+ "▁das": 2846,
+ "▁rad": 2847,
+ "▁close": 2848,
+ "www": 2849,
+ "ере": 2850,
+ "gu": 2851,
+ "▁Er": 2852,
+ "▁dom": 2853,
+ "AM": 2854,
+ "▁bed": 2855,
+ "▁several": 2856,
+ "aul": 2857,
+ "box": 2858,
+ "▁low": 2859,
+ "pack": 2860,
+ "Reg": 2861,
+ "Of": 2862,
+ "atures": 2863,
+ "én": 2864,
+ "eder": 2865,
+ "uilder": 2866,
+ "cast": 2867,
+ "conom": 2868,
+ "raft": 2869,
+ "▁makes": 2870,
+ "Loc": 2871,
+ "http": 2872,
+ "▁abs": 2873,
+ "resh": 2874,
+ "▁Will": 2875,
+ "break": 2876,
+ "▁options": 2877,
+ "fort": 2878,
+ "▁из": 2879,
+ "▁anal": 2880,
+ "▁env": 2881,
+ "({": 2882,
+ "event": 2883,
+ "▁page": 2884,
+ "ternal": 2885,
+ "▁distribut": 2886,
+ "▁food": 2887,
+ "check": 2888,
+ "CK": 2889,
+ "▁во": 2890,
+ "assert": 2891,
+ "án": 2892,
+ "base": 2893,
+ "▁whole": 2894,
+ "ación": 2895,
+ "OD": 2896,
+ "▁turned": 2897,
+ "igma": 2898,
+ "▁response": 2899,
+ "▁University": 2900,
+ "▁div": 2901,
+ "apter": 2902,
+ "▁results": 2903,
+ "▁represent": 2904,
+ "▁everything": 2905,
+ "▁Cent": 2906,
+ "utes": 2907,
+ "rix": 2908,
+ "▁Some": 2909,
+ "▁behind": 2910,
+ "▁creat": 2911,
+ "place": 2912,
+ "su": 2913,
+ "▁Part": 2914,
+ "umb": 2915,
+ "mathbb": 2916,
+ "ping": 2917,
+ "▁match": 2918,
+ "Out": 2919,
+ "dom": 2920,
+ "▁situ": 2921,
+ "dr": 2922,
+ "ara": 2923,
+ "▁window": 2924,
+ "ns": 2925,
+ "lished": 2926,
+ "▁Ver": 2927,
+ "▁message": 2928,
+ "▁Em": 2929,
+ "▁human": 2930,
+ "perties": 2931,
+ "лу": 2932,
+ "lem": 2933,
+ "ORT": 2934,
+ "▁early": 2935,
+ "▁quick": 2936,
+ "▁та": 2937,
+ "roid": 2938,
+ "▁country": 2939,
+ "▁due": 2940,
+ "▁Die": 2941,
+ "▁trying": 2942,
+ "▁live": 2943,
+ "▁press": 2944,
+ "INT": 2945,
+ "With": 2946,
+ "oved": 2947,
+ "▁specific": 2948,
+ "▁fall": 2949,
+ "uk": 2950,
+ "yl": 2951,
+ "▁general": 2952,
+ "му": 2953,
+ "ну": 2954,
+ "▁names": 2955,
+ "where": 2956,
+ "▁These": 2957,
+ "▁sil": 2958,
+ "ét": 2959,
+ "▁ener": 2960,
+ "▁Now": 2961,
+ "▁address": 2962,
+ "Response": 2963,
+ "▁Mr": 2964,
+ "▁answ": 2965,
+ "▁film": 2966,
+ "▁strong": 2967,
+ "▁bring": 2968,
+ "▁United": 2969,
+ "▁ge": 2970,
+ "▁woman": 2971,
+ "New": 2972,
+ "ett": 2973,
+ ".)": 2974,
+ "ename": 2975,
+ "▁AN": 2976,
+ "▁describ": 2977,
+ "за": 2978,
+ "ising": 2979,
+ "EL": 2980,
+ "ql": 2981,
+ "▁fur": 2982,
+ "ying": 2983,
+ "▁Cal": 2984,
+ "▁Dr": 2985,
+ "ERR": 2986,
+ "▁\\\\": 2987,
+ "angle": 2988,
+ "urope": 2989,
+ "▁city": 2990,
+ "▁index": 2991,
+ "▁action": 2992,
+ "▁However": 2993,
+ "▁fig": 2994,
+ "ias": 2995,
+ "▁question": 2996,
+ "▁Jan": 2997,
+ "▁Med": 2998,
+ "▁Cont": 2999,
+ "amed": 3000,
+ "Call": 3001,
+ "plied": 3002,
+ "tty": 3003,
+ "▁individ": 3004,
+ "page": 3005,
+ "▁comb": 3006,
+ "section": 3007,
+ "▁Comm": 3008,
+ "uel": 3009,
+ "▁het": 3010,
+ "▁Bar": 3011,
+ "agement": 3012,
+ "fin": 3013,
+ "▁major": 3014,
+ "oper": 3015,
+ "api": 3016,
+ "room": 3017,
+ "▁„": 3018,
+ "▁hab": 3019,
+ "зи": 3020,
+ "▁auf": 3021,
+ "current": 3022,
+ "ni": 3023,
+ "▁include": 3024,
+ "▁qui": 3025,
+ "va": 3026,
+ "UE": 3027,
+ "▁idea": 3028,
+ ",'": 3029,
+ "▁required": 3030,
+ "▁heart": 3031,
+ "ibility": 3032,
+ "iction": 3033,
+ "Model": 3034,
+ "write": 3035,
+ "▁content": 3036,
+ "▁wer": 3037,
+ "▁hands": 3038,
+ "zen": 3039,
+ "char": 3040,
+ "}^{": 3041,
+ "▁mass": 3042,
+ "ply": 3043,
+ "▁nat": 3044,
+ "rel": 3045,
+ "▁dat": 3046,
+ "================": 3047,
+ "imal": 3048,
+ "▁probably": 3049,
+ "unch": 3050,
+ "▁mer": 3051,
+ "ilar": 3052,
+ "ires": 3053,
+ "▁watch": 3054,
+ "SI": 3055,
+ "▁cult": 3056,
+ "▁mother": 3057,
+ "▁government": 3058,
+ "ording": 3059,
+ "▁()": 3060,
+ "▁pri": 3061,
+ "▁link": 3062,
+ "group": 3063,
+ "OL": 3064,
+ "▁near": 3065,
+ "▁Ser": 3066,
+ "Ser": 3067,
+ "ito": 3068,
+ "▁values": 3069,
+ "▁java": 3070,
+ "fully": 3071,
+ "Count": 3072,
+ "++)": 3073,
+ "▁vi": 3074,
+ "▁white": 3075,
+ "mat": 3076,
+ "ctx": 3077,
+ "▁conc": 3078,
+ "▁stay": 3079,
+ "ging": 3080,
+ "▁clear": 3081,
+ "▁copy": 3082,
+ "selves": 3083,
+ "▁provide": 3084,
+ "▁words": 3085,
+ "comp": 3086,
+ "args": 3087,
+ "▁pick": 3088,
+ "uly": 3089,
+ "▁vari": 3090,
+ "▁believe": 3091,
+ "▁Co": 3092,
+ "Property": 3093,
+ "Group": 3094,
+ "▁ten": 3095,
+ "ischen": 3096,
+ "eturn": 3097,
+ "ival": 3098,
+ "System": 3099,
+ "CL": 3100,
+ "bed": 3101,
+ "▁total": 3102,
+ "▁ist": 3103,
+ "Input": 3104,
+ "uments": 3105,
+ "Manager": 3106,
+ "ши": 3107,
+ "▁win": 3108,
+ "leep": 3109,
+ "PI": 3110,
+ "ного": 3111,
+ "ruction": 3112,
+ "▁inte": 3113,
+ "App": 3114,
+ "avor": 3115,
+ "▁respect": 3116,
+ "ators": 3117,
+ "▁como": 3118,
+ "▁cut": 3119,
+ "FA": 3120,
+ "▁sus": 3121,
+ "▁App": 3122,
+ "rect": 3123,
+ "FI": 3124,
+ "▁began": 3125,
+ "oph": 3126,
+ "▁sort": 3127,
+ "though": 3128,
+ "је": 3129,
+ "icro": 3130,
+ "Trans": 3131,
+ "лі": 3132,
+ "▁Inst": 3133,
+ "request": 3134,
+ "ор": 3135,
+ "▁relations": 3136,
+ "-\\": 3137,
+ "Status": 3138,
+ "жи": 3139,
+ "▁father": 3140,
+ "cs": 3141,
+ "▁sex": 3142,
+ "isch": 3143,
+ "vo": 3144,
+ "}_{": 3145,
+ "aven": 3146,
+ "▁Ne": 3147,
+ "ATE": 3148,
+ "itten": 3149,
+ "▁ess": 3150,
+ "TH": 3151,
+ "ights": 3152,
+ "▁hom": 3153,
+ "▁today": 3154,
+ "▁zu": 3155,
+ "ita": 3156,
+ "▁isn": 3157,
+ "▁opt": 3158,
+ "ogn": 3159,
+ "ér": 3160,
+ "▁whether": 3161,
+ "ixed": 3162,
+ "phi": 3163,
+ "idence": 3164,
+ "ald": 3165,
+ "Client": 3166,
+ "At": 3167,
+ "▁death": 3168,
+ "▁Let": 3169,
+ "ius": 3170,
+ "ги": 3171,
+ "▁ре": 3172,
+ "ben": 3173,
+ ")\r": 3174,
+ "ba": 3175,
+ ">": 3176,
+ "avel": 3177,
+ "▁miss": 3178,
+ "▁node": 3179,
+ "▁($": 3180,
+ "▁color": 3181,
+ "▁obt": 3182,
+ "tot": 3183,
+ "▁пре": 3184,
+ "CON": 3185,
+ "ette": 3186,
+ "▁Go": 3187,
+ "Fl": 3188,
+ "▁Don": 3189,
+ "▁crit": 3190,
+ "▁ri": 3191,
+ "post": 3192,
+ "▁->": 3193,
+ "▁Just": 3194,
+ "What": 3195,
+ "atal": 3196,
+ "▁Min": 3197,
+ "▁Cor": 3198,
+ "▁dark": 3199,
+ "rl": 3200,
+ "▁larg": 3201,
+ "ding": 3202,
+ "ón": 3203,
+ "ouch": 3204,
+ "▁um": 3205,
+ "▁elect": 3206,
+ "▁dam": 3207,
+ "▁needs": 3208,
+ "▁matter": 3209,
+ "▁rather": 3210,
+ "from": 3211,
+ "ram": 3212,
+ "▁і": 3213,
+ "▁taken": 3214,
+ "▁deal": 3215,
+ "▁period": 3216,
+ "▁Mon": 3217,
+ "▁Л": 3218,
+ "▁Aug": 3219,
+ "run": 3220,
+ "mm": 3221,
+ "elle": 3222,
+ "▁export": 3223,
+ "Sc": 3224,
+ "vis": 3225,
+ "abor": 3226,
+ "▁author": 3227,
+ "ère": 3228,
+ "▁remember": 3229,
+ "▁redu": 3230,
+ "▁List": 3231,
+ "▁focus": 3232,
+ "▁character": 3233,
+ "Table": 3234,
+ "▁individual": 3235,
+ "▁needed": 3236,
+ "bum": 3237,
+ "▁style": 3238,
+ "inary": 3239,
+ "ersion": 3240,
+ "oute": 3241,
+ "▁Pe": 3242,
+ "▁hon": 3243,
+ "mut": 3244,
+ "see": 3245,
+ "▁became": 3246,
+ "▁dire": 3247,
+ "▁document": 3248,
+ "sec": 3249,
+ "ening": 3250,
+ "▁visit": 3251,
+ "▁fac": 3252,
+ "tx": 3253,
+ "down": 3254,
+ "plit": 3255,
+ "▁phys": 3256,
+ "itting": 3257,
+ "joy": 3258,
+ "▁hig": 3259,
+ "This": 3260,
+ "Ad": 3261,
+ "▁Brit": 3262,
+ "▁employ": 3263,
+ "▁ré": 3264,
+ "▁т": 3265,
+ "lambda": 3266,
+ "▁impro": 3267,
+ "▁Bo": 3268,
+ "iding": 3269,
+ "▁online": 3270,
+ "mem": 3271,
+ "atform": 3272,
+ "▁War": 3273,
+ "▁cas": 3274,
+ "asure": 3275,
+ "▁pur": 3276,
+ "medi": 3277,
+ "Dis": 3278,
+ "▁Germ": 3279,
+ "pc": 3280,
+ "са": 3281,
+ "▁friends": 3282,
+ "▁Mc": 3283,
+ "DI": 3284,
+ "▁plus": 3285,
+ "▁Set": 3286,
+ "iddle": 3287,
+ "itut": 3288,
+ "▁depend": 3289,
+ "rest": 3290,
+ "▁Je": 3291,
+ "▁hor": 3292,
+ "▁entire": 3293,
+ "Query": 3294,
+ "▁refer": 3295,
+ "▁hot": 3296,
+ "▁Aust": 3297,
+ "▁common": 3298,
+ "ці": 3299,
+ "▁pull": 3300,
+ "▁Add": 3301,
+ "▁season": 3302,
+ "▁invol": 3303,
+ "▁World": 3304,
+ "client": 3305,
+ "now": 3306,
+ "true": 3307,
+ "append": 3308,
+ "itted": 3309,
+ "empt": 3310,
+ "){": 3311,
+ "///": 3312,
+ "▁prop": 3313,
+ "imate": 3314,
+ "SC": 3315,
+ "▁hours": 3316,
+ "▁hope": 3317,
+ "andom": 3318,
+ "ід": 3319,
+ "istic": 3320,
+ "▁property": 3321,
+ "sg": 3322,
+ ">(": 3323,
+ "▁write": 3324,
+ "mark": 3325,
+ "find": 3326,
+ "▁personal": 3327,
+ "][": 3328,
+ "rown": 3329,
+ "Ph": 3330,
+ "▁foot": 3331,
+ "▁research": 3332,
+ "ironment": 3333,
+ "▁nom": 3334,
+ "▁instance": 3335,
+ "▁held": 3336,
+ "De": 3337,
+ "▁members": 3338,
+ "▁fire": 3339,
+ "▁history": 3340,
+ "▁map": 3341,
+ "▁discuss": 3342,
+ "▁espec": 3343,
+ "▁taking": 3344,
+ "▁services": 3345,
+ "▁indust": 3346,
+ "igen": 3347,
+ "▁Ass": 3348,
+ "▁expected": 3349,
+ "▁wurde": 3350,
+ "dir": 3351,
+ "▁among": 3352,
+ "▁sugg": 3353,
+ "rec": 3354,
+ "Inter": 3355,
+ "block": 3356,
+ "▁Rep": 3357,
+ "▁pain": 3358,
+ "▁five": 3359,
+ "▁fund": 3360,
+ "rid": 3361,
+ "arrow": 3362,
+ "▁treat": 3363,
+ "▁heard": 3364,
+ "▁determ": 3365,
+ "icult": 3366,
+ "▁sense": 3367,
+ "ese": 3368,
+ "Fun": 3369,
+ "▁months": 3370,
+ "json": 3371,
+ ",”": 3372,
+ "TI": 3373,
+ "orage": 3374,
+ "▁У": 3375,
+ "▁everyone": 3376,
+ "▁clos": 3377,
+ "iers": 3378,
+ "airs": 3379,
+ "define": 3380,
+ "If": 3381,
+ "osp": 3382,
+ "▁wonder": 3383,
+ "NA": 3384,
+ "query": 3385,
+ "pg": 3386,
+ "ites": 3387,
+ "▁material": 3388,
+ "yd": 3389,
+ "Read": 3390,
+ "html": 3391,
+ "TE": 3392,
+ "Pr": 3393,
+ "^{\\": 3394,
+ "▁gave": 3395,
+ "▁IS": 3396,
+ "▁suggest": 3397,
+ "Override": 3398,
+ "rodu": 3399,
+ "From": 3400,
+ "▁Europe": 3401,
+ "PO": 3402,
+ "▁soon": 3403,
+ "host": 3404,
+ "▁Ber": 3405,
+ "....": 3406,
+ "▁Har": 3407,
+ "▁energy": 3408,
+ "><": 3409,
+ "aves": 3410,
+ "▁easy": 3411,
+ "▁bre": 3412,
+ "frame": 3413,
+ "▁ground": 3414,
+ "with": 3415,
+ "▁inside": 3416,
+ "ief": 3417,
+ "▁mo": 3418,
+ "pm": 3419,
+ "pan": 3420,
+ "igr": 3421,
+ "▁om": 3422,
+ "next": 3423,
+ "omet": 3424,
+ "▁status": 3425,
+ "▁}\r": 3426,
+ "▁music": 3427,
+ "ora": 3428,
+ "iles": 3429,
+ "ki": 3430,
+ "▁esc": 3431,
+ "▁bes": 3432,
+ "▁Dis": 3433,
+ "▁host": 3434,
+ "▁comes": 3435,
+ "used": 3436,
+ "▁future": 3437,
+ "lick": 3438,
+ "aid": 3439,
+ "▁compet": 3440,
+ "▁voice": 3441,
+ "▁load": 3442,
+ "evel": 3443,
+ "▁neg": 3444,
+ "▁command": 3445,
+ "▁für": 3446,
+ "▁pie": 3447,
+ "▁quite": 3448,
+ "▁blo": 3449,
+ "agn": 3450,
+ "ilon": 3451,
+ "▁claim": 3452,
+ "▁teach": 3453,
+ "▁previous": 3454,
+ "▁site": 3455,
+ "color": 3456,
+ "attr": 3457,
+ "▁accept": 3458,
+ "▁exact": 3459,
+ ")}": 3460,
+ "aft": 3461,
+ "roller": 3462,
+ "он": 3463,
+ "oo": 3464,
+ "Date": 3465,
+ "▁ou": 3466,
+ "sy": 3467,
+ "▁pretty": 3468,
+ "▁image": 3469,
+ "BU": 3470,
+ "▁terms": 3471,
+ "▁search": 3472,
+ "▁è": 3473,
+ "▁Val": 3474,
+ "▁‘": 3475,
+ "▁Dav": 3476,
+ "MS": 3477,
+ "src": 3478,
+ "mar": 3479,
+ "incip": 3480,
+ "▁couldn": 3481,
+ "ados": 3482,
+ "▁dro": 3483,
+ "beta": 3484,
+ "imum": 3485,
+ "▁minutes": 3486,
+ "▁grand": 3487,
+ "▁»": 3488,
+ "▁Our": 3489,
+ "Str": 3490,
+ "VER": 3491,
+ "maz": 3492,
+ "▁original": 3493,
+ "ini": 3494,
+ "▁coll": 3495,
+ "loat": 3496,
+ "▁os": 3497,
+ "});": 3498,
+ "summary": 3499,
+ "▁wall": 3500,
+ "Color": 3501,
+ "▁vers": 3502,
+ "▁della": 3503,
+ "▁\"\"\"": 3504,
+ "mathbf": 3505,
+ "zer": 3506,
+ "aur": 3507,
+ "▁track": 3508,
+ "▁associ": 3509,
+ "▁suff": 3510,
+ "▁inde": 3511,
+ "ague": 3512,
+ "▁Apr": 3513,
+ "Le": 3514,
+ "roups": 3515,
+ "board": 3516,
+ "▁attack": 3517,
+ "▁series": 3518,
+ "▁instead": 3519,
+ "ham": 3520,
+ "book": 3521,
+ "▁six": 3522,
+ "▁Rec": 3523,
+ "▁coming": 3524,
+ "urt": 3525,
+ "▁global": 3526,
+ "▁necess": 3527,
+ "lege": 3528,
+ "Pos": 3529,
+ "▁leave": 3530,
+ "▁pod": 3531,
+ "ategory": 3532,
+ "uz": 3533,
+ "▁deep": 3534,
+ "▁km": 3535,
+ "▁outside": 3536,
+ "has": 3537,
+ "options": 3538,
+ "▁Sm": 3539,
+ "Sub": 3540,
+ "rows": 3541,
+ "▁ви": 3542,
+ "▁States": 3543,
+ "▁wrong": 3544,
+ "▁however": 3545,
+ "▁sem": 3546,
+ "▁catch": 3547,
+ "\"),": 3548,
+ "model": 3549,
+ "▁http": 3550,
+ "▁option": 3551,
+ "rie": 3552,
+ "▁ста": 3553,
+ "▁är": 3554,
+ "▁enjoy": 3555,
+ "nu": 3556,
+ "▁pas": 3557,
+ "▁amount": 3558,
+ "▁respons": 3559,
+ "▁Intern": 3560,
+ "▁myself": 3561,
+ "▁opp": 3562,
+ "▁Sim": 3563,
+ "▁sens": 3564,
+ "Ed": 3565,
+ "▁(\\": 3566,
+ "▁students": 3567,
+ "нов": 3568,
+ "▁points": 3569,
+ "arning": 3570,
+ "UP": 3571,
+ "elling": 3572,
+ "▁cannot": 3573,
+ "Be": 3574,
+ "▁length": 3575,
+ "null": 3576,
+ "uint": 3577,
+ "wise": 3578,
+ "▁double": 3579,
+ "ige": 3580,
+ "ista": 3581,
+ "▁estab": 3582,
+ "anch": 3583,
+ "▁ago": 3584,
+ "▁bound": 3585,
+ "▁fa": 3586,
+ "▁clean": 3587,
+ "▁simple": 3588,
+ "mi": 3589,
+ "########": 3590,
+ "ifier": 3591,
+ "▁General": 3592,
+ "▁seemed": 3593,
+ "ena": 3594,
+ "▁age": 3595,
+ "ной": 3596,
+ "endif": 3597,
+ "AA": 3598,
+ "▁caus": 3599,
+ "▁educ": 3600,
+ "▁cell": 3601,
+ "Gener": 3602,
+ "space": 3603,
+ "▁Your": 3604,
+ "▁beaut": 3605,
+ "gt": 3606,
+ "▁limit": 3607,
+ "▁date": 3608,
+ "Util": 3609,
+ "▁National": 3610,
+ "ows": 3611,
+ "pat": 3612,
+ "quad": 3613,
+ "▁ok": 3614,
+ "▁И": 3615,
+ "arth": 3616,
+ "hat": 3617,
+ "▁community": 3618,
+ "oul": 3619,
+ "▁econom": 3620,
+ "Component": 3621,
+ "bor": 3622,
+ "usion": 3623,
+ "▁below": 3624,
+ "earch": 3625,
+ "ores": 3626,
+ "ban": 3627,
+ "▁August": 3628,
+ "▁further": 3629,
+ "sigma": 3630,
+ "▁ha": 3631,
+ "ji": 3632,
+ "▁comput": 3633,
+ "гра": 3634,
+ "▁None": 3635,
+ "▁ter": 3636,
+ "▁anyone": 3637,
+ "▁task": 3638,
+ "ente": 3639,
+ "position": 3640,
+ "pped": 3641,
+ "▁aus": 3642,
+ "Attribute": 3643,
+ "req": 3644,
+ "addr": 3645,
+ "light": 3646,
+ "ше": 3647,
+ "▁arm": 3648,
+ "cover": 3649,
+ "upport": 3650,
+ "▁Gl": 3651,
+ "▁San": 3652,
+ "▁writing": 3653,
+ "▁lost": 3654,
+ "▁Mark": 3655,
+ "▁gre": 3656,
+ "TYPE": 3657,
+ "▁South": 3658,
+ "▁perfect": 3659,
+ "▁package": 3660,
+ "▁infl": 3661,
+ "haps": 3662,
+ "▁Ang": 3663,
+ "respon": 3664,
+ "ris": 3665,
+ "ptember": 3666,
+ "▁building": 3667,
+ "VAL": 3668,
+ "free": 3669,
+ "▁ce": 3670,
+ "HT": 3671,
+ "▁From": 3672,
+ "ds": 3673,
+ "roy": 3674,
+ "achine": 3675,
+ "nown": 3676,
+ "▁saying": 3677,
+ "▁бы": 3678,
+ "oe": 3679,
+ "Ref": 3680,
+ "▁network": 3681,
+ "parent": 3682,
+ "uge": 3683,
+ "▁similar": 3684,
+ ">\r": 3685,
+ "Builder": 3686,
+ "▁living": 3687,
+ "▁continue": 3688,
+ "anger": 3689,
+ "▁Red": 3690,
+ "▁hair": 3691,
+ "anced": 3692,
+ "ians": 3693,
+ "▁dead": 3694,
+ "▁boolean": 3695,
+ "ication": 3696,
+ "▁де": 3697,
+ "▁client": 3698,
+ "uct": 3699,
+ "▁•": 3700,
+ "SP": 3701,
+ "older": 3702,
+ "пе": 3703,
+ "udio": 3704,
+ "▁deg": 3705,
+ "asing": 3706,
+ "▁step": 3707,
+ "▁pers": 3708,
+ "ção": 3709,
+ "obj": 3710,
+ "oz": 3711,
+ "ula": 3712,
+ "▁round": 3713,
+ "▁upon": 3714,
+ "▁resource": 3715,
+ "▁valid": 3716,
+ "▁II": 3717,
+ "bug": 3718,
+ "std": 3719,
+ "▁ang": 3720,
+ "span": 3721,
+ "pol": 3722,
+ "ialog": 3723,
+ "▁phot": 3724,
+ "?'": 3725,
+ "DB": 3726,
+ "▁Fin": 3727,
+ "VE": 3728,
+ "Em": 3729,
+ "▁cam": 3730,
+ "target": 3731,
+ "pected": 3732,
+ "Hel": 3733,
+ "▁ut": 3734,
+ "▁Test": 3735,
+ "▁town": 3736,
+ "align": 3737,
+ "▁webs": 3738,
+ "inner": 3739,
+ "augh": 3740,
+ "▁except": 3741,
+ "▁initial": 3742,
+ "enty": 3743,
+ "lich": 3744,
+ "▁Aut": 3745,
+ "top": 3746,
+ "▁fail": 3747,
+ "ona": 3748,
+ "▁benef": 3749,
+ "anks": 3750,
+ "ische": 3751,
+ ".*": 3752,
+ "▁signific": 3753,
+ "▁contact": 3754,
+ "Rec": 3755,
+ "ario": 3756,
+ "ottom": 3757,
+ "▁relationship": 3758,
+ "]);": 3759,
+ "▁На": 3760,
+ "Head": 3761,
+ "format": 3762,
+ "▁ét": 3763,
+ "▁More": 3764,
+ "actory": 3765,
+ "portun": 3766,
+ "+\\": 3767,
+ "▁simply": 3768,
+ "▁ep": 3769,
+ "▁Russ": 3770,
+ "ní": 3771,
+ "ua": 3772,
+ "erc": 3773,
+ "▁longer": 3774,
+ "inition": 3775,
+ "ector": 3776,
+ "aption": 3777,
+ "▁profess": 3778,
+ "▁Mus": 3779,
+ "ilities": 3780,
+ "ès": 3781,
+ "▁Act": 3782,
+ "offset": 3783,
+ "▁ill": 3784,
+ "band": 3785,
+ "▁Ag": 3786,
+ "▁По": 3787,
+ "би": 3788,
+ "content": 3789,
+ "icon": 3790,
+ "▁works": 3791,
+ "ynam": 3792,
+ "plement": 3793,
+ "Resource": 3794,
+ "Action": 3795,
+ "▁difficult": 3796,
+ "▁West": 3797,
+ "▁video": 3798,
+ "▁THE": 3799,
+ "▁decl": 3800,
+ "ondon": 3801,
+ "ded": 3802,
+ "}{\\": 3803,
+ "ocr": 3804,
+ "▁City": 3805,
+ "▁я": 3806,
+ "uer": 3807,
+ "cz": 3808,
+ "▁imag": 3809,
+ "cr": 3810,
+ "ete": 3811,
+ "idget": 3812,
+ "▁Mod": 3813,
+ "▁forward": 3814,
+ "▁pict": 3815,
+ "orge": 3816,
+ "▁subject": 3817,
+ "update": 3818,
+ "attle": 3819,
+ "sa": 3820,
+ "▁Ant": 3821,
+ "▁running": 3822,
+ "▁sal": 3823,
+ "conne": 3824,
+ "▁output": 3825,
+ "adata": 3826,
+ "ML": 3827,
+ "Check": 3828,
+ "ledge": 3829,
+ "▁paper": 3830,
+ "params": 3831,
+ "avy": 3832,
+ "▁af": 3833,
+ "▁eine": 3834,
+ "▁jour": 3835,
+ "AY": 3836,
+ "▁itself": 3837,
+ "▁Str": 3838,
+ "style": 3839,
+ "That": 3840,
+ "▁million": 3841,
+ "▁language": 3842,
+ "OS": 3843,
+ "ving": 3844,
+ "▁ма": 3845,
+ "▁то": 3846,
+ ")(": 3847,
+ "▁buy": 3848,
+ "./": 3849,
+ "▁...": 3850,
+ "▁tried": 3851,
+ "▁compl": 3852,
+ "▁activ": 3853,
+ "apped": 3854,
+ "Button": 3855,
+ "Token": 3856,
+ "▁provided": 3857,
+ "iber": 3858,
+ "▁created": 3859,
+ "curity": 3860,
+ "End": 3861,
+ "ał": 3862,
+ "uster": 3863,
+ "izing": 3864,
+ "omb": 3865,
+ "▁sich": 3866,
+ "▁compon": 3867,
+ "▁See": 3868,
+ "▁uint": 3869,
+ "▁label": 3870,
+ "vol": 3871,
+ "ów": 3872,
+ "ocol": 3873,
+ "▁received": 3874,
+ "▁intern": 3875,
+ "це": 3876,
+ "Run": 3877,
+ "▁road": 3878,
+ "▁Oct": 3879,
+ "▁Comp": 3880,
+ "▁study": 3881,
+ "▁те": 3882,
+ "Act": 3883,
+ "▁tour": 3884,
+ "▁State": 3885,
+ "▁added": 3886,
+ "https": 3887,
+ "stream": 3888,
+ "▁lower": 3889,
+ "▁box": 3890,
+ "▁Sk": 3891,
+ "▁themselves": 3892,
+ "▁cross": 3893,
+ "▁echo": 3894,
+ "▁device": 3895,
+ "pose": 3896,
+ "▁games": 3897,
+ "PL": 3898,
+ "Window": 3899,
+ "ises": 3900,
+ "title": 3901,
+ "Stream": 3902,
+ "zt": 3903,
+ "▁Sw": 3904,
+ "▁role": 3905,
+ "iant": 3906,
+ "ku": 3907,
+ "sequ": 3908,
+ "▁late": 3909,
+ "▁sold": 3910,
+ "ря": 3911,
+ "Comm": 3912,
+ "▁entre": 3913,
+ "▁dog": 3914,
+ "device": 3915,
+ "Par": 3916,
+ "▁likely": 3917,
+ "^{-": 3918,
+ "▁len": 3919,
+ "▁Paul": 3920,
+ "▁tool": 3921,
+ "Off": 3922,
+ "▁famil": 3923,
+ "▁draw": 3924,
+ "apping": 3925,
+ "▁events": 3926,
+ "cret": 3927,
+ "rought": 3928,
+ "Content": 3929,
+ "▁software": 3930,
+ "ria": 3931,
+ "msg": 3932,
+ "gamma": 3933,
+ "▁hear": 3934,
+ "Oper": 3935,
+ "▁yourself": 3936,
+ "▁liter": 3937,
+ "emp": 3938,
+ "▁separ": 3939,
+ "▁З": 3940,
+ "▁title": 3941,
+ "Method": 3942,
+ "mathrm": 3943,
+ "▁slow": 3944,
+ "▁Rom": 3945,
+ "!!": 3946,
+ "▁tax": 3947,
+ "ска": 3948,
+ "emplate": 3949,
+ "oi": 3950,
+ "▁Art": 3951,
+ "false": 3952,
+ "astic": 3953,
+ "сть": 3954,
+ "ocket": 3955,
+ "▁ens": 3956,
+ "TO": 3957,
+ "amente": 3958,
+ "local": 3959,
+ "chie": 3960,
+ "▁pan": 3961,
+ "ний": 3962,
+ "chema": 3963,
+ "▁North": 3964,
+ "зо": 3965,
+ "▁>=": 3966,
+ "Aut": 3967,
+ "▁dig": 3968,
+ "▁seems": 3969,
+ "▁morning": 3970,
+ "sole": 3971,
+ "umer": 3972,
+ "delta": 3973,
+ "ité": 3974,
+ "abase": 3975,
+ "raf": 3976,
+ "▁observ": 3977,
+ "▁Est": 3978,
+ "▁seg": 3979,
+ "▁[]": 3980,
+ "▁Pres": 3981,
+ "iful": 3982,
+ "push": 3983,
+ "▁Off": 3984,
+ "ipe": 3985,
+ "ati": 3986,
+ "▁dim": 3987,
+ "ceed": 3988,
+ "Ent": 3989,
+ "____": 3990,
+ "entry": 3991,
+ "▁fight": 3992,
+ "▁cred": 3993,
+ "▁OR": 3994,
+ "▁Dep": 3995,
+ "${": 3996,
+ "лен": 3997,
+ "Create": 3998,
+ "▁April": 3999,
+ "ministr": 4000,
+ "FL": 4001,
+ "▁Ap": 4002,
+ "▁Here": 4003,
+ "private": 4004,
+ "Instance": 4005,
+ "iem": 4006,
+ "▁office": 4007,
+ "▁third": 4008,
+ "▁update": 4009,
+ "Line": 4010,
+ "tag": 4011,
+ "▁especially": 4012,
+ "▁года": 4013,
+ "▁cu": 4014,
+ "▁kill": 4015,
+ "aught": 4016,
+ "▁swe": 4017,
+ "Options": 4018,
+ "IM": 4019,
+ "CC": 4020,
+ "▁compan": 4021,
+ "just": 4022,
+ "▁While": 4023,
+ "izer": 4024,
+ "▁мо": 4025,
+ "ке": 4026,
+ "▁auto": 4027,
+ "▁band": 4028,
+ "мен": 4029,
+ "iques": 4030,
+ "▁ple": 4031,
+ "NO": 4032,
+ "▁OF": 4033,
+ "▁song": 4034,
+ "▁Acc": 4035,
+ "EXT": 4036,
+ "ensor": 4037,
+ "ining": 4038,
+ "▁lat": 4039,
+ "big": 4040,
+ "▁King": 4041,
+ "och": 4042,
+ "si": 4043,
+ "▁Hist": 4044,
+ "▁quality": 4045,
+ "mode": 4046,
+ "▁opportun": 4047,
+ "▁wouldn": 4048,
+ ":**": 4049,
+ "output": 4050,
+ "▁feet": 4051,
+ "▁mis": 4052,
+ "df": 4053,
+ "aging": 4054,
+ "▁ме": 4055,
+ "▁tro": 4056,
+ "▁defined": 4057,
+ "▁review": 4058,
+ "▁Fil": 4059,
+ ">>": 4060,
+ "▁princip": 4061,
+ "Base": 4062,
+ "dict": 4063,
+ "verage": 4064,
+ "icient": 4065,
+ "IF": 4066,
+ "▁hit": 4067,
+ "Page": 4068,
+ "▁perm": 4069,
+ "cel": 4070,
+ "ít": 4071,
+ "▁express": 4072,
+ "▁indic": 4073,
+ "▁September": 4074,
+ "image": 4075,
+ "▁products": 4076,
+ "▁media": 4077,
+ "change": 4078,
+ "igger": 4079,
+ "▁send": 4080,
+ "last": 4081,
+ "ming": 4082,
+ "pa": 4083,
+ "uary": 4084,
+ "▁speak": 4085,
+ "ный": 4086,
+ "ще": 4087,
+ "ysis": 4088,
+ "lying": 4089,
+ "▁ч": 4090,
+ "like": 4091,
+ "ры": 4092,
+ "ві": 4093,
+ "▁Mich": 4094,
+ "MO": 4095,
+ "▁Jah": 4096,
+ "ensive": 4097,
+ "▁share": 4098,
+ "▁development": 4099,
+ "CP": 4100,
+ "spec": 4101,
+ "▁fast": 4102,
+ "het": 4103,
+ "HO": 4104,
+ "▁particip": 4105,
+ "Block": 4106,
+ "▁viol": 4107,
+ "▁frame": 4108,
+ "▁qual": 4109,
+ "tre": 4110,
+ "▁Ф": 4111,
+ "▁toward": 4112,
+ "fg": 4113,
+ "Box": 4114,
+ "Column": 4115,
+ "▁milit": 4116,
+ "▁March": 4117,
+ "▁various": 4118,
+ "pass": 4119,
+ "▁Park": 4120,
+ "▁Ben": 4121,
+ "Frame": 4122,
+ "▁normal": 4123,
+ "open": 4124,
+ "px": 4125,
+ "▁phone": 4126,
+ "▁Even": 4127,
+ "▁ma": 4128,
+ "ibrary": 4129,
+ "Start": 4130,
+ "idden": 4131,
+ "rho": 4132,
+ "graph": 4133,
+ "acing": 4134,
+ "'.": 4135,
+ "arter": 4136,
+ "mes": 4137,
+ "inst": 4138,
+ "▁ir": 4139,
+ "active": 4140,
+ "▁fem": 4141,
+ "▁moved": 4142,
+ "▁store": 4143,
+ "▁price": 4144,
+ "\").": 4145,
+ "berg": 4146,
+ "▁nov": 4147,
+ "▁card": 4148,
+ "ellow": 4149,
+ "▁party": 4150,
+ "▁Mor": 4151,
+ "ael": 4152,
+ "▁percent": 4153,
+ "▁training": 4154,
+ "▁ing": 4155,
+ "imer": 4156,
+ "▁Sam": 4157,
+ "Default": 4158,
+ "▁fuck": 4159,
+ "▁complete": 4160,
+ "uid": 4161,
+ "▁details": 4162,
+ "▁led": 4163,
+ "Point": 4164,
+ "▁Count": 4165,
+ "▁regard": 4166,
+ "zo": 4167,
+ "▁Bro": 4168,
+ "▁recogn": 4169,
+ "▁Hol": 4170,
+ "UM": 4171,
+ "element": 4172,
+ "Mode": 4173,
+ "▁exam": 4174,
+ "▁EX": 4175,
+ "Image": 4176,
+ "verse": 4177,
+ "riter": 4178,
+ "soft": 4179,
+ "▁introdu": 4180,
+ "▁surpr": 4181,
+ "Buffer": 4182,
+ "lector": 4183,
+ "aren": 4184,
+ "anged": 4185,
+ "▁Pat": 4186,
+ "▁Pal": 4187,
+ "▁contr": 4188,
+ "Handler": 4189,
+ "▁features": 4190,
+ "iple": 4191,
+ "▁CON": 4192,
+ "Fil": 4193,
+ "▁Port": 4194,
+ "▁thinking": 4195,
+ "doc": 4196,
+ "wer": 4197,
+ "▁worked": 4198,
+ "PC": 4199,
+ "cm": 4200,
+ "dat": 4201,
+ "PRO": 4202,
+ "▁Every": 4203,
+ "▁era": 4204,
+ "▁First": 4205,
+ "gn": 4206,
+ "▁immedi": 4207,
+ "ovember": 4208,
+ "apan": 4209,
+ "▁extra": 4210,
+ "▁section": 4211,
+ "▁June": 4212,
+ "▁via": 4213,
+ "▁gone": 4214,
+ "come": 4215,
+ "▁stri": 4216,
+ "^\\": 4217,
+ "antly": 4218,
+ "▁arch": 4219,
+ "Source": 4220,
+ "▁conv": 4221,
+ "▁London": 4222,
+ "Number": 4223,
+ "▁questions": 4224,
+ "andid": 4225,
+ "▁played": 4226,
+ "env": 4227,
+ "▁School": 4228,
+ "▁natural": 4229,
+ "can": 4230,
+ "▁news": 4231,
+ "DR": 4232,
+ "▁chall": 4233,
+ "▁Soc": 4234,
+ "▁э": 4235,
+ "▁attempt": 4236,
+ "*}": 4237,
+ "Null": 4238,
+ "rote": 4239,
+ "▁bi": 4240,
+ "▁written": 4241,
+ "▁blood": 4242,
+ "▁happened": 4243,
+ "▁cause": 4244,
+ "ashing": 4245,
+ "▁William": 4246,
+ "adem": 4247,
+ "▁brought": 4248,
+ "▁display": 4249,
+ "ima": 4250,
+ "▁finally": 4251,
+ "tab": 4252,
+ "▁returned": 4253,
+ "ных": 4254,
+ "nie": 4255,
+ "▁q": 4256,
+ "▁hers": 4257,
+ "▁Pre": 4258,
+ "▁dou": 4259,
+ "buffer": 4260,
+ "▁effort": 4261,
+ "aine": 4262,
+ "xy": 4263,
+ "▁histor": 4264,
+ "enu": 4265,
+ "▁arriv": 4266,
+ "▁Dem": 4267,
+ "▁favor": 4268,
+ "▁handle": 4269,
+ "SET": 4270,
+ "▁Public": 4271,
+ "rupt": 4272,
+ "▁ur": 4273,
+ "▁force": 4274,
+ "▁és": 4275,
+ "ube": 4276,
+ "Pre": 4277,
+ "рі": 4278,
+ "iny": 4279,
+ "theta": 4280,
+ "isf": 4281,
+ "▁national": 4282,
+ "Equal": 4283,
+ "rench": 4284,
+ "▁wife": 4285,
+ "▁capt": 4286,
+ "▁Inter": 4287,
+ "tau": 4288,
+ "▁sleep": 4289,
+ "../../": 4290,
+ "▁issue": 4291,
+ "▁member": 4292,
+ "▁await": 4293,
+ "▁Dan": 4294,
+ "zi": 4295,
+ "inate": 4296,
+ "▁sym": 4297,
+ "chan": 4298,
+ "▁Jack": 4299,
+ "▁English": 4300,
+ "▁sz": 4301,
+ "ributes": 4302,
+ "▁ign": 4303,
+ "ál": 4304,
+ "▁appear": 4305,
+ "rad": 4306,
+ "idge": 4307,
+ "▁couple": 4308,
+ "▁ship": 4309,
+ "lig": 4310,
+ "web": 4311,
+ "▁usually": 4312,
+ "▁ready": 4313,
+ "▁vill": 4314,
+ "▁Why": 4315,
+ "ebru": 4316,
+ "▁grad": 4317,
+ "ords": 4318,
+ "▁inf": 4319,
+ "▁loss": 4320,
+ "▁od": 4321,
+ "▁Phil": 4322,
+ "server": 4323,
+ "▁Up": 4324,
+ "▁buff": 4325,
+ "▁filename": 4326,
+ "ABLE": 4327,
+ "iting": 4328,
+ "efore": 4329,
+ "()->": 4330,
+ "▁conditions": 4331,
+ "vm": 4332,
+ "eld": 4333,
+ "itz": 4334,
+ "▁Trans": 4335,
+ "▁weight": 4336,
+ "▁higher": 4337,
+ "▁rate": 4338,
+ "▁accom": 4339,
+ "vider": 4340,
+ "OM": 4341,
+ "▁ways": 4342,
+ "coming": 4343,
+ "▁lock": 4344,
+ "▁etc": 4345,
+ "▁avec": 4346,
+ "▁takes": 4347,
+ "▁Char": 4348,
+ "▁November": 4349,
+ "method": 4350,
+ "▁Austral": 4351,
+ "▁America": 4352,
+ "long": 4353,
+ "cember": 4354,
+ "▁political": 4355,
+ "flow": 4356,
+ "▁maybe": 4357,
+ "▁amb": 4358,
+ "Layout": 4359,
+ "iled": 4360,
+ "omen": 4361,
+ "ola": 4362,
+ "icip": 4363,
+ "partial": 4364,
+ "True": 4365,
+ "▁floor": 4366,
+ "▁Def": 4367,
+ "▁concern": 4368,
+ "yr": 4369,
+ "▁shows": 4370,
+ "ih": 4371,
+ "▁answer": 4372,
+ "acc": 4373,
+ "▁ball": 4374,
+ "▁Rev": 4375,
+ "▁sun": 4376,
+ "▁quickly": 4377,
+ "▁somet": 4378,
+ "mente": 4379,
+ "▁Mal": 4380,
+ "undred": 4381,
+ "▁issues": 4382,
+ "ecause": 4383,
+ "pes": 4384,
+ "▁player": 4385,
+ "▁parents": 4386,
+ "▁popular": 4387,
+ "▁mode": 4388,
+ "▁mention": 4389,
+ "NE": 4390,
+ "Load": 4391,
+ "▁regular": 4392,
+ "aved": 4393,
+ "?:": 4394,
+ "year": 4395,
+ "func": 4396,
+ "▁performance": 4397,
+ "▁July": 4398,
+ "thern": 4399,
+ "▁website": 4400,
+ "ford": 4401,
+ "PR": 4402,
+ "ela": 4403,
+ "level": 4404,
+ "uit": 4405,
+ "flags": 4406,
+ "▁worth": 4407,
+ "▁correspon": 4408,
+ "▁British": 4409,
+ "sim": 4410,
+ "▁alone": 4411,
+ "▁har": 4412,
+ "▁ones": 4413,
+ "obile": 4414,
+ "▁dru": 4415,
+ "chi": 4416,
+ "▁David": 4417,
+ "▁problems": 4418,
+ "▁column": 4419,
+ "();\r": 4420,
+ "ZE": 4421,
+ "▁relig": 4422,
+ "ological": 4423,
+ "▁region": 4424,
+ "ady": 4425,
+ "IO": 4426,
+ "ander": 4427,
+ "Net": 4428,
+ "▁built": 4429,
+ "▁install": 4430,
+ "▁approach": 4431,
+ "Cur": 4432,
+ "▁fine": 4433,
+ "▁talking": 4434,
+ "▁changes": 4435,
+ "Style": 4436,
+ "▁Mart": 4437,
+ "лю": 4438,
+ "response": 4439,
+ "teger": 4440,
+ "{\r": 4441,
+ "irit": 4442,
+ "▁protected": 4443,
+ "▁rele": 4444,
+ "ership": 4445,
+ "тель": 4446,
+ "unsigned": 4447,
+ "ialize": 4448,
+ "▁https": 4449,
+ "Tag": 4450,
+ "▁$(": 4451,
+ "more": 4452,
+ "ypes": 4453,
+ "▁stream": 4454,
+ "etch": 4455,
+ "▁engine": 4456,
+ "KE": 4457,
+ "cmd": 4458,
+ "script": 4459,
+ "ttp": 4460,
+ "▁avoid": 4461,
+ "▁terr": 4462,
+ "▁rock": 4463,
+ "▁ful": 4464,
+ "Update": 4465,
+ "▁environment": 4466,
+ "▁prec": 4467,
+ "▁са": 4468,
+ "▁cases": 4469,
+ "▁offset": 4470,
+ "▁rais": 4471,
+ "lib": 4472,
+ "ées": 4473,
+ "aa": 4474,
+ "yt": 4475,
+ "▁arr": 4476,
+ "opyright": 4477,
+ "first": 4478,
+ "▁util": 4479,
+ "▁feature": 4480,
+ "posed": 4481,
+ "ffect": 4482,
+ "жа": 4483,
+ "itude": 4484,
+ "ements": 4485,
+ "asc": 4486,
+ "ador": 4487,
+ "lections": 4488,
+ "▁club": 4489,
+ "]{": 4490,
+ "▁*)": 4491,
+ "ство": 4492,
+ "▁imm": 4493,
+ "▁former": 4494,
+ "▁rights": 4495,
+ "▁decided": 4496,
+ "▁rev": 4497,
+ "▁ment": 4498,
+ "ani": 4499,
+ "▁stru": 4500,
+ "▁attention": 4501,
+ "artment": 4502,
+ "▁Ital": 4503,
+ "alle": 4504,
+ "▁bis": 4505,
+ "gener": 4506,
+ "▁integr": 4507,
+ "ello": 4508,
+ "rypt": 4509,
+ "▁achie": 4510,
+ "nes": 4511,
+ "▁stra": 4512,
+ "sb": 4513,
+ "▁types": 4514,
+ "▁RE": 4515,
+ "Init": 4516,
+ "▁comment": 4517,
+ "▁addition": 4518,
+ "▁ID": 4519,
+ "ART": 4520,
+ "FO": 4521,
+ "щи": 4522,
+ "Conne": 4523,
+ "▁squ": 4524,
+ "▁considered": 4525,
+ "idad": 4526,
+ "▁October": 4527,
+ "cial": 4528,
+ "▁Of": 4529,
+ "▁travel": 4530,
+ "▁boy": 4531,
+ "').": 4532,
+ "uy": 4533,
+ "illa": 4534,
+ "istry": 4535,
+ "▁va": 4536,
+ "▁Che": 4537,
+ "ERT": 4538,
+ "ende": 4539,
+ "ungen": 4540,
+ "aby": 4541,
+ "▁Rober": 4542,
+ "▁playing": 4543,
+ "ils": 4544,
+ "▁sam": 4545,
+ "▁execut": 4546,
+ "▁Us": 4547,
+ "▁mut": 4548,
+ "▁bal": 4549,
+ "asse": 4550,
+ "▁kids": 4551,
+ "▁financ": 4552,
+ "gor": 4553,
+ "▁Sec": 4554,
+ "bert": 4555,
+ "▁High": 4556,
+ "▁је": 4557,
+ "▁kept": 4558,
+ "button": 4559,
+ "itory": 4560,
+ "▁Rem": 4561,
+ "▁DE": 4562,
+ "▁reach": 4563,
+ "▁bur": 4564,
+ "Label": 4565,
+ "át": 4566,
+ "ago": 4567,
+ "▁passed": 4568,
+ "▁behav": 4569,
+ "xFF": 4570,
+ "▁Return": 4571,
+ "STR": 4572,
+ "▁Les": 4573,
+ "▁ord": 4574,
+ "ala": 4575,
+ "inger": 4576,
+ "▁Since": 4577,
+ "▁experi": 4578,
+ "▁shall": 4579,
+ "▁star": 4580,
+ "non": 4581,
+ "▁gun": 4582,
+ "▁Bel": 4583,
+ "▁obj": 4584,
+ "ares": 4585,
+ "rs": 4586,
+ "▁weeks": 4587,
+ "nen": 4588,
+ "▁Stre": 4589,
+ "oring": 4590,
+ "▁î": 4591,
+ "▁serious": 4592,
+ "times": 4593,
+ "▁House": 4594,
+ "▁roll": 4595,
+ "▁register": 4596,
+ "▁module": 4597,
+ "▁applic": 4598,
+ "IR": 4599,
+ "▁cook": 4600,
+ "aux": 4601,
+ "▁save": 4602,
+ "▁Cr": 4603,
+ ",\r": 4604,
+ "▁states": 4605,
+ "▁empty": 4606,
+ "▁autom": 4607,
+ "figure": 4608,
+ "iance": 4609,
+ "▁happy": 4610,
+ "▁fn": 4611,
+ "▁jud": 4612,
+ "▁hat": 4613,
+ "ACK": 4614,
+ "▁Fe": 4615,
+ "$-": 4616,
+ "ivil": 4617,
+ "oted": 4618,
+ "▁sizeof": 4619,
+ "▁situation": 4620,
+ "▁lives": 4621,
+ "▁feeling": 4622,
+ "▁risk": 4623,
+ "▁January": 4624,
+ "▁Object": 4625,
+ "▁recomm": 4626,
+ "▁вы": 4627,
+ "▁potential": 4628,
+ "eah": 4629,
+ "▁complex": 4630,
+ "printf": 4631,
+ "istance": 4632,
+ "irth": 4633,
+ "lik": 4634,
+ "aste": 4635,
+ "▁whose": 4636,
+ "Arg": 4637,
+ "▁modern": 4638,
+ "iones": 4639,
+ "▁че": 4640,
+ "▁sett": 4641,
+ "▁Mag": 4642,
+ "ae": 4643,
+ "▁condition": 4644,
+ "Length": 4645,
+ "▁fit": 4646,
+ "ounds": 4647,
+ "▁changed": 4648,
+ "▁guy": 4649,
+ "filter": 4650,
+ "atever": 4651,
+ "éd": 4652,
+ "remove": 4653,
+ "▁hop": 4654,
+ "▁Out": 4655,
+ "▁Rich": 4656,
+ "child": 4657,
+ "▁included": 4658,
+ "$\\": 4659,
+ "▁Tom": 4660,
+ "eline": 4661,
+ "▁sometimes": 4662,
+ "▁drink": 4663,
+ "▁quant": 4664,
+ "▁please": 4665,
+ "▁Int": 4666,
+ "rief": 4667,
+ "▁exactly": 4668,
+ "cing": 4669,
+ "▁allowed": 4670,
+ "build": 4671,
+ "▁beautiful": 4672,
+ "▁Well": 4673,
+ "▁looks": 4674,
+ "▁ü": 4675,
+ "▁chance": 4676,
+ "▁wrote": 4677,
+ "▁nor": 4678,
+ "▁failed": 4679,
+ "Met": 4680,
+ "▁prior": 4681,
+ "▁hundred": 4682,
+ "ской": 4683,
+ "oria": 4684,
+ "▁cy": 4685,
+ "▁web": 4686,
+ "▁mess": 4687,
+ "leq": 4688,
+ "dy": 4689,
+ "tex": 4690,
+ "▁anim": 4691,
+ "atur": 4692,
+ "▁structure": 4693,
+ "option": 4694,
+ "▁actual": 4695,
+ "▁Franc": 4696,
+ "enced": 4697,
+ ".": 4698,
+ "▁flow": 4699,
+ "▁Afr": 4700,
+ "det": 4701,
+ "▁Ke": 4702,
+ "ety": 4703,
+ "ский": 4704,
+ "▁stuff": 4705,
+ "itter": 4706,
+ "▁args": 4707,
+ "▁album": 4708,
+ "▁]": 4709,
+ "ugin": 4710,
+ "SU": 4711,
+ "Per": 4712,
+ "▁circ": 4713,
+ "▁correct": 4714,
+ "▁lines": 4715,
+ "▁completely": 4716,
+ "known": 4717,
+ "▁tree": 4718,
+ "root": 4719,
+ "▁Japan": 4720,
+ "oles": 4721,
+ "endo": 4722,
+ "▁location": 4723,
+ "▁Х": 4724,
+ "▁mid": 4725,
+ "aling": 4726,
+ "GL": 4727,
+ "iano": 4728,
+ "▁{}": 4729,
+ "lang": 4730,
+ "▁equip": 4731,
+ "ERROR": 4732,
+ "▁memory": 4733,
+ "▁(\"": 4734,
+ "▁nature": 4735,
+ "google": 4736,
+ "abs": 4737,
+ "BC": 4738,
+ "▁gets": 4739,
+ "Command": 4740,
+ "TER": 4741,
+ "aled": 4742,
+ "cp": 4743,
+ "▁purch": 4744,
+ "▁Den": 4745,
+ "▁herself": 4746,
+ "▁Ir": 4747,
+ "▁sie": 4748,
+ "gar": 4749,
+ "Ap": 4750,
+ "▁nel": 4751,
+ "ota": 4752,
+ ")]": 4753,
+ "cor": 4754,
+ "acht": 4755,
+ "(*": 4756,
+ "irtual": 4757,
+ "▁police": 4758,
+ "▁skin": 4759,
+ "ship": 4760,
+ "efined": 4761,
+ "aughter": 4762,
+ "inding": 4763,
+ "▁Sl": 4764,
+ "▁influ": 4765,
+ "▁mount": 4766,
+ "▁az": 4767,
+ "▁wood": 4768,
+ "otes": 4769,
+ "ega": 4770,
+ "▁according": 4771,
+ "▁namespace": 4772,
+ "Delta": 4773,
+ "stant": 4774,
+ "▁published": 4775,
+ "aker": 4776,
+ "▁Black": 4777,
+ "ln": 4778,
+ "▁industry": 4779,
+ "SON": 4780,
+ "Rep": 4781,
+ "▁choice": 4782,
+ "▁inn": 4783,
+ "kl": 4784,
+ "▁pal": 4785,
+ "▁aud": 4786,
+ "▁standard": 4787,
+ "▁knowledge": 4788,
+ "**,": 4789,
+ "▁Frank": 4790,
+ "sq": 4791,
+ "Output": 4792,
+ "▁för": 4793,
+ "Valid": 4794,
+ "ugh": 4795,
+ "▁books": 4796,
+ "▁James": 4797,
+ "ko": 4798,
+ "▁companies": 4799,
+ "anning": 4800,
+ "▁vict": 4801,
+ "▁repl": 4802,
+ "▁sche": 4803,
+ "▁happen": 4804,
+ "fty": 4805,
+ "acity": 4806,
+ "ira": 4807,
+ "▁implement": 4808,
+ "ского": 4809,
+ "number": 4810,
+ "SH": 4811,
+ "iro": 4812,
+ "▁fear": 4813,
+ "▁touch": 4814,
+ "▁cast": 4815,
+ "ASS": 4816,
+ "▁consist": 4817,
+ "Task": 4818,
+ "▁sig": 4819,
+ "ба": 4820,
+ "igation": 4821,
+ "▁Most": 4822,
+ "▁Der": 4823,
+ "}(\\": 4824,
+ ":\"": 4825,
+ "▁Fig": 4826,
+ "ali": 4827,
+ "iner": 4828,
+ "'),": 4829,
+ "▁Coun": 4830,
+ "(_": 4831,
+ "▁distributed": 4832,
+ "NAME": 4833,
+ "▁mur": 4834,
+ "▁career": 4835,
+ "~~": 4836,
+ "pers": 4837,
+ "aries": 4838,
+ "enses": 4839,
+ "▁Also": 4840,
+ "Version": 4841,
+ "▁unique": 4842,
+ "▁France": 4843,
+ "BA": 4844,
+ "ky": 4845,
+ "▁Febru": 4846,
+ "▁died": 4847,
+ "omega": 4848,
+ "▁Form": 4849,
+ "▁width": 4850,
+ "tocol": 4851,
+ "▁lie": 4852,
+ "She": 4853,
+ "ém": 4854,
+ "▁straight": 4855,
+ "▁nach": 4856,
+ "▁stood": 4857,
+ "olds": 4858,
+ "▁goes": 4859,
+ "cell": 4860,
+ "▁till": 4861,
+ "LI": 4862,
+ "draw": 4863,
+ "▁satisf": 4864,
+ "▁reading": 4865,
+ "ATION": 4866,
+ "▁Are": 4867,
+ "▁Ac": 4868,
+ ")*": 4869,
+ "▁additional": 4870,
+ "wood": 4871,
+ "cil": 4872,
+ "пу": 4873,
+ "ULT": 4874,
+ "▁bill": 4875,
+ "mas": 4876,
+ "ania": 4877,
+ "су": 4878,
+ "anz": 4879,
+ "height": 4880,
+ "jo": 4881,
+ "▁dos": 4882,
+ "\\\"": 4883,
+ "▁/>": 4884,
+ "▁production": 4885,
+ "iger": 4886,
+ "▁ст": 4887,
+ "show": 4888,
+ "▁population": 4889,
+ "▁park": 4890,
+ "▁Ze": 4891,
+ "▁necessary": 4892,
+ "▁trust": 4893,
+ "▁shown": 4894,
+ "module": 4895,
+ "GE": 4896,
+ "▁lay": 4897,
+ "▁announ": 4898,
+ "▁className": 4899,
+ "▁calcul": 4900,
+ "Function": 4901,
+ "▁Sal": 4902,
+ "OK": 4903,
+ "TP": 4904,
+ "▁entry": 4905,
+ "▁Stud": 4906,
+ "▁items": 4907,
+ "▁security": 4908,
+ "Entry": 4909,
+ "float": 4910,
+ "ls": 4911,
+ "ibly": 4912,
+ "▁contribut": 4913,
+ "▁Check": 4914,
+ "MD": 4915,
+ "▁improve": 4916,
+ "Part": 4917,
+ "▁systems": 4918,
+ "Bl": 4919,
+ "▁policy": 4920,
+ "▁screen": 4921,
+ "▁Any": 4922,
+ "▁opened": 4923,
+ "alloc": 4924,
+ "▁December": 4925,
+ "▁É": 4926,
+ "▁email": 4927,
+ "ader": 4928,
+ "=>": 4929,
+ "▁Hen": 4930,
+ "▁info": 4931,
+ "▁float": 4932,
+ "▁switch": 4933,
+ "ран": 4934,
+ "urance": 4935,
+ "▁assum": 4936,
+ "ustr": 4937,
+ "▁groups": 4938,
+ "▁Read": 4939,
+ "▁wat": 4940,
+ "Sp": 4941,
+ "вер": 4942,
+ "RAN": 4943,
+ "hib": 4944,
+ "ALL": 4945,
+ "▁hus": 4946,
+ "Spec": 4947,
+ "\"))": 4948,
+ "▁French": 4949,
+ "▁Class": 4950,
+ "▁president": 4951,
+ "▁definit": 4952,
+ "▁Nor": 4953,
+ "▁Thom": 4954,
+ "aign": 4955,
+ "Width": 4956,
+ "Do": 4957,
+ "▁{@": 4958,
+ "agon": 4959,
+ "▁Lu": 4960,
+ "▁followed": 4961,
+ "MM": 4962,
+ "asons": 4963,
+ "tmp": 4964,
+ "▁throws": 4965,
+ "ITY": 4966,
+ "ном": 4967,
+ "▁fair": 4968,
+ "▁pen": 4969,
+ "ég": 4970,
+ "▁interface": 4971,
+ "▁saf": 4972,
+ "oon": 4973,
+ "Back": 4974,
+ "▁speed": 4975,
+ "▁extends": 4976,
+ "empty": 4977,
+ "▁пере": 4978,
+ "▁proper": 4979,
+ "▁driv": 4980,
+ "фи": 4981,
+ "▁center": 4982,
+ "header": 4983,
+ "▁})": 4984,
+ "wa": 4985,
+ "▁middle": 4986,
+ "▁choose": 4987,
+ "▁Stad": 4988,
+ "SO": 4989,
+ "Factory": 4990,
+ "Dev": 4991,
+ "icles": 4992,
+ "▁application": 4993,
+ "▁models": 4994,
+ "pite": 4995,
+ "cap": 4996,
+ "xi": 4997,
+ "ospital": 4998,
+ "▁dream": 4999,
+ "END": 5000,
+ "▁contract": 5001,
+ "icrosoft": 5002,
+ "▁thous": 5003,
+ "izes": 5004,
+ "▁да": 5005,
+ "▁CO": 5006,
+ "▁direction": 5007,
+ "▁``": 5008,
+ "▁drive": 5009,
+ "Max": 5010,
+ "cia": 5011,
+ "▁continu": 5012,
+ "▁Alex": 5013,
+ "▁gold": 5014,
+ "▁prep": 5015,
+ "▁origin": 5016,
+ "▁rap": 5017,
+ "Op": 5018,
+ "ously": 5019,
+ "▁areas": 5020,
+ "PORT": 5021,
+ "она": 5022,
+ "▁safe": 5023,
+ "▁professional": 5024,
+ "apache": 5025,
+ "▁temper": 5026,
+ "sz": 5027,
+ "▁unit": 5028,
+ "▁cop": 5029,
+ "eqn": 5030,
+ "Listener": 5031,
+ "▁format": 5032,
+ "select": 5033,
+ "▁comfort": 5034,
+ "▁meant": 5035,
+ "iday": 5036,
+ "eme": 5037,
+ "▁active": 5038,
+ "▁note": 5039,
+ "▁Mil": 5040,
+ "only": 5041,
+ "▁<=": 5042,
+ "▁neigh": 5043,
+ "ao": 5044,
+ "▁blue": 5045,
+ "▁TV": 5046,
+ "Child": 5047,
+ "▁reached": 5048,
+ "Address": 5049,
+ "ств": 5050,
+ "▁closed": 5051,
+ "inder": 5052,
+ "olo": 5053,
+ "▁alt": 5054,
+ "▁adm": 5055,
+ "Format": 5056,
+ "UI": 5057,
+ "▁Ham": 5058,
+ "▁frequ": 5059,
+ "▁independ": 5060,
+ "▁easily": 5061,
+ "▁Land": 5062,
+ "▁tor": 5063,
+ "ography": 5064,
+ "infty": 5065,
+ "▁Work": 5066,
+ "iven": 5067,
+ "▁County": 5068,
+ "▁src": 5069,
+ "}$,": 5070,
+ "parse": 5071,
+ "CD": 5072,
+ "▁Cour": 5073,
+ "▁fol": 5074,
+ "Entity": 5075,
+ "pgf": 5076,
+ "▁China": 5077,
+ "▁Sub": 5078,
+ "hood": 5079,
+ "▁fields": 5080,
+ "▁yes": 5081,
+ "rend": 5082,
+ "▁towards": 5083,
+ "▁staff": 5084,
+ "▁Air": 5085,
+ "▁station": 5086,
+ "atives": 5087,
+ "▁impact": 5088,
+ "вы": 5089,
+ "▁directly": 5090,
+ "issions": 5091,
+ "iva": 5092,
+ "|\\": 5093,
+ "Ptr": 5094,
+ "▁Sant": 5095,
+ "Pol": 5096,
+ "▁progress": 5097,
+ "itar": 5098,
+ "▁parts": 5099,
+ "▁plant": 5100,
+ "▁absolut": 5101,
+ "▁guess": 5102,
+ "eqref": 5103,
+ "▁tim": 5104,
+ "▁Lou": 5105,
+ "▁cool": 5106,
+ "alu": 5107,
+ "▁mouth": 5108,
+ "них": 5109,
+ "▁height": 5110,
+ "gest": 5111,
+ "▁Post": 5112,
+ "▁board": 5113,
+ "▁tit": 5114,
+ "▁hour": 5115,
+ "▁server": 5116,
+ "▁players": 5117,
+ "rier": 5118,
+ "Link": 5119,
+ "▁President": 5120,
+ "](": 5121,
+ "▁construct": 5122,
+ "handle": 5123,
+ "}$.": 5124,
+ "rying": 5125,
+ "▁shop": 5126,
+ "iana": 5127,
+ "exp": 5128,
+ "Helper": 5129,
+ "Offset": 5130,
+ "aches": 5131,
+ "▁connection": 5132,
+ "▁difference": 5133,
+ "service": 5134,
+ "▁gas": 5135,
+ "▁priv": 5136,
+ "▁univers": 5137,
+ "▁wish": 5138,
+ "Rem": 5139,
+ "Url": 5140,
+ "geb": 5141,
+ "So": 5142,
+ "ensions": 5143,
+ "Module": 5144,
+ "SIZE": 5145,
+ "▁prem": 5146,
+ "window": 5147,
+ "▁dies": 5148,
+ "del": 5149,
+ "▁row": 5150,
+ "▁average": 5151,
+ "xim": 5152,
+ "▁pu": 5153,
+ "anç": 5154,
+ "Det": 5155,
+ "ker": 5156,
+ "ya": 5157,
+ "▁Det": 5158,
+ "▁på": 5159,
+ "▁named": 5160,
+ "▁decision": 5161,
+ "win": 5162,
+ "▁George": 5163,
+ "arily": 5164,
+ "▁solution": 5165,
+ "▁multiple": 5166,
+ "ategy": 5167,
+ "▁learning": 5168,
+ "▁secret": 5169,
+ "DO": 5170,
+ "▁nice": 5171,
+ "////////////////": 5172,
+ "Su": 5173,
+ "itation": 5174,
+ "▁join": 5175,
+ "▁elements": 5176,
+ "▁emer": 5177,
+ "tilde": 5178,
+ "▁dep": 5179,
+ "▁shot": 5180,
+ "▁platform": 5181,
+ "othing": 5182,
+ "My": 5183,
+ "edia": 5184,
+ "oms": 5185,
+ "aily": 5186,
+ "([": 5187,
+ "▁dress": 5188,
+ "▁official": 5189,
+ "estern": 5190,
+ "▁discover": 5191,
+ "▁mi": 5192,
+ "ные": 5193,
+ "CA": 5194,
+ "oding": 5195,
+ "▁Found": 5196,
+ "▁affect": 5197,
+ "Vis": 5198,
+ "stract": 5199,
+ "iced": 5200,
+ "debug": 5201,
+ "▁related": 5202,
+ "▁spect": 5203,
+ "ushed": 5204,
+ "сько": 5205,
+ "▁bank": 5206,
+ "▁cele": 5207,
+ "AND": 5208,
+ "olf": 5209,
+ "ем": 5210,
+ "▁fill": 5211,
+ "▁gives": 5212,
+ "▁бу": 5213,
+ "aron": 5214,
+ "▁Jes": 5215,
+ "REG": 5216,
+ "▁sudd": 5217,
+ "dated": 5218,
+ "vi": 5219,
+ "▁gi": 5220,
+ "send": 5221,
+ "cpp": 5222,
+ "▁spent": 5223,
+ "ande": 5224,
+ "▁operation": 5225,
+ "process": 5226,
+ "▁inform": 5227,
+ "▁Free": 5228,
+ "yond": 5229,
+ "▁perhaps": 5230,
+ "▁surv": 5231,
+ "▁Loc": 5232,
+ "▁concl": 5233,
+ "▁раз": 5234,
+ "▁Over": 5235,
+ "hol": 5236,
+ "raz": 5237,
+ "Write": 5238,
+ "▁giving": 5239,
+ "rd": 5240,
+ "instance": 5241,
+ "▁released": 5242,
+ "▁Ro": 5243,
+ "RA": 5244,
+ "▁practice": 5245,
+ "▁graph": 5246,
+ "▁increase": 5247,
+ "▁figure": 5248,
+ "Filter": 5249,
+ "HECK": 5250,
+ "idx": 5251,
+ "▁glass": 5252,
+ "ski": 5253,
+ "comes": 5254,
+ "▁cat": 5255,
+ "▁cold": 5256,
+ "goto": 5257,
+ "ufact": 5258,
+ "▁Copyright": 5259,
+ "}}\\": 5260,
+ "▁streng": 5261,
+ "▁dir": 5262,
+ "token": 5263,
+ "▁occur": 5264,
+ "arlier": 5265,
+ "▁measure": 5266,
+ "▁sec": 5267,
+ "▁más": 5268,
+ "▁Net": 5269,
+ "▁argument": 5270,
+ "▁sou": 5271,
+ "▁moving": 5272,
+ "▁prefer": 5273,
+ "mask": 5274,
+ "<<": 5275,
+ "▁breath": 5276,
+ "▁physical": 5277,
+ "▁positive": 5278,
+ "▁sor": 5279,
+ "▁depart": 5280,
+ "▁remove": 5281,
+ "▁kit": 5282,
+ "▁meeting": 5283,
+ "▁Data": 5284,
+ "ograf": 5285,
+ "actions": 5286,
+ "▁parameters": 5287,
+ "▁Att": 5288,
+ "esch": 5289,
+ "▁involved": 5290,
+ "ät": 5291,
+ "LL": 5292,
+ "Bar": 5293,
+ "▁си": 5294,
+ "ech": 5295,
+ "GET": 5296,
+ "▁prevent": 5297,
+ "▁beyond": 5298,
+ "▁Other": 5299,
+ "än": 5300,
+ "byte": 5301,
+ "▁sudden": 5302,
+ "olve": 5303,
+ "▁но": 5304,
+ "LOG": 5305,
+ "unit": 5306,
+ "▁truth": 5307,
+ "rat": 5308,
+ "SD": 5309,
+ "▁eat": 5310,
+ "▁Mad": 5311,
+ "▁provides": 5312,
+ "▁session": 5313,
+ "Dele": 5314,
+ "▁convers": 5315,
+ "center": 5316,
+ "▁continued": 5317,
+ "otion": 5318,
+ "cache": 5319,
+ "display": 5320,
+ "▁protect": 5321,
+ "ams": 5322,
+ "▁pow": 5323,
+ "CTION": 5324,
+ "▁Mac": 5325,
+ "mo": 5326,
+ "ха": 5327,
+ "▁distance": 5328,
+ "▁Time": 5329,
+ "gi": 5330,
+ "▁sequ": 5331,
+ "Target": 5332,
+ "сле": 5333,
+ "Server": 5334,
+ "▁wide": 5335,
+ "close": 5336,
+ "▁cru": 5337,
+ "Ext": 5338,
+ "▁select": 5339,
+ "▁pattern": 5340,
+ "\"));": 5341,
+ "Provider": 5342,
+ "URL": 5343,
+ "▁green": 5344,
+ "▁waiting": 5345,
+ "proto": 5346,
+ "▁immediately": 5347,
+ "common": 5348,
+ "azione": 5349,
+ "river": 5350,
+ "▁sen": 5351,
+ "▁!==": 5352,
+ "▁February": 5353,
+ "urb": 5354,
+ "▁Sen": 5355,
+ "dest": 5356,
+ "": 5357,
+ "▁edge": 5358,
+ "▁mais": 5359,
+ "gorith": 5360,
+ "cpu": 5361,
+ "▁education": 5362,
+ "▁associated": 5363,
+ "None": 5364,
+ "hi": 5365,
+ "▁poor": 5366,
+ "sem": 5367,
+ "▁Wil": 5368,
+ "▁bud": 5369,
+ "▁auch": 5370,
+ "eller": 5371,
+ "▁Life": 5372,
+ "▁files": 5373,
+ "▁leading": 5374,
+ "▁obtain": 5375,
+ "▁Jul": 5376,
+ "atory": 5377,
+ "гу": 5378,
+ "itable": 5379,
+ "▁onto": 5380,
+ "▁born": 5381,
+ "orem": 5382,
+ "▁Street": 5383,
+ "▁maint": 5384,
+ "Params": 5385,
+ "rip": 5386,
+ "▁ST": 5387,
+ "uv": 5388,
+ "main": 5389,
+ "▁▁▁▁▁▁▁": 5390,
+ "▁recent": 5391,
+ "Web": 5392,
+ "ova": 5393,
+ "ца": 5394,
+ "aise": 5395,
+ "yles": 5396,
+ "▁described": 5397,
+ "▁beginning": 5398,
+ "▁Day": 5399,
+ "▁Vol": 5400,
+ "▁huge": 5401,
+ "Has": 5402,
+ "ancy": 5403,
+ "Header": 5404,
+ "▁aren": 5405,
+ "ван": 5406,
+ "▁ensure": 5407,
+ "▁pet": 5408,
+ "mult": 5409,
+ "▁Like": 5410,
+ "▁management": 5411,
+ "PS": 5412,
+ "while": 5413,
+ "▁background": 5414,
+ "ounter": 5415,
+ "bool": 5416,
+ "FC": 5417,
+ "Num": 5418,
+ "RL": 5419,
+ "▁excl": 5420,
+ "▁eye": 5421,
+ "img": 5422,
+ "▁rom": 5423,
+ "▁Hel": 5424,
+ "Option": 5425,
+ "▁stopped": 5426,
+ "▁thread": 5427,
+ "totype": 5428,
+ ")))": 5429,
+ "▁stage": 5430,
+ "▁über": 5431,
+ "▁although": 5432,
+ "Types": 5433,
+ "▁Oh": 5434,
+ "▁eight": 5435,
+ "▁description": 5436,
+ "''": 5437,
+ "ön": 5438,
+ "▁surface": 5439,
+ "▁International": 5440,
+ "▁charg": 5441,
+ "▁collection": 5442,
+ "▁users": 5443,
+ "▁obvious": 5444,
+ "▁century": 5445,
+ "icks": 5446,
+ "▁article": 5447,
+ "▁\"\\": 5448,
+ "dim": 5449,
+ "▁sin": 5450,
+ "enge": 5451,
+ "Control": 5452,
+ "▁commit": 5453,
+ "ensity": 5454,
+ "▁tra": 5455,
+ "criptor": 5456,
+ "▁NOT": 5457,
+ "well": 5458,
+ "▁Michael": 5459,
+ "▁nod": 5460,
+ "▁mort": 5461,
+ "ivo": 5462,
+ "isation": 5463,
+ "▁Po": 5464,
+ "▁Paris": 5465,
+ "▁administr": 5466,
+ "burg": 5467,
+ "cdot": 5468,
+ "▁military": 5469,
+ "▁Best": 5470,
+ "▁Ка": 5471,
+ "INE": 5472,
+ "▁throughout": 5473,
+ "Sl": 5474,
+ "▁impl": 5475,
+ "control": 5476,
+ "▁Ч": 5477,
+ "▁uit": 5478,
+ "▁unsigned": 5479,
+ "▁Mary": 5480,
+ "Char": 5481,
+ "мі": 5482,
+ "▁threat": 5483,
+ "▁court": 5484,
+ "ville": 5485,
+ "▁ш": 5486,
+ "▁Cam": 5487,
+ ".\r": 5488,
+ "▁currently": 5489,
+ "rot": 5490,
+ "▁Date": 5491,
+ "▁shit": 5492,
+ "▁${\\": 5493,
+ "unn": 5494,
+ "Us": 5495,
+ "▁buffer": 5496,
+ "▁sont": 5497,
+ "▁letter": 5498,
+ "inated": 5499,
+ "Change": 5500,
+ "▁href": 5501,
+ "▁lack": 5502,
+ "▁oil": 5503,
+ "▁Cons": 5504,
+ "▁Jer": 5505,
+ "BUG": 5506,
+ "iforn": 5507,
+ "▁properties": 5508,
+ "▁random": 5509,
+ "▁brother": 5510,
+ "▁piece": 5511,
+ "бу": 5512,
+ "istics": 5513,
+ "▁technology": 5514,
+ "global": 5515,
+ "▁transform": 5516,
+ "erd": 5517,
+ "▁Because": 5518,
+ "PECT": 5519,
+ "pret": 5520,
+ "▁году": 5521,
+ "▁Met": 5522,
+ "▁psy": 5523,
+ "▁од": 5524,
+ "▁god": 5525,
+ "▁Del": 5526,
+ "based": 5527,
+ "▁voor": 5528,
+ "▁Call": 5529,
+ "SA": 5530,
+ "▁filter": 5531,
+ "▁includes": 5532,
+ "olutions": 5533,
+ "fd": 5534,
+ "▁wind": 5535,
+ "▁бо": 5536,
+ "▁ability": 5537,
+ "card": 5538,
+ "▁numer": 5539,
+ "address": 5540,
+ "▁goal": 5541,
+ "ashington": 5542,
+ "▁slight": 5543,
+ "aba": 5544,
+ "▁Log": 5545,
+ "Settings": 5546,
+ "adow": 5547,
+ "▁pi": 5548,
+ "iring": 5549,
+ "FT": 5550,
+ "▁numbers": 5551,
+ "conf": 5552,
+ "task": 5553,
+ "▁în": 5554,
+ "ты": 5555,
+ "▁receive": 5556,
+ "▁root": 5557,
+ "▁India": 5558,
+ "patch": 5559,
+ "él": 5560,
+ "▁summer": 5561,
+ "▁methods": 5562,
+ "▁places": 5563,
+ "▁Ма": 5564,
+ "▁capital": 5565,
+ "▁evidence": 5566,
+ "▁German": 5567,
+ "\\,": 5568,
+ "DA": 5569,
+ "ecute": 5570,
+ "column": 5571,
+ "▁functions": 5572,
+ "▁counter": 5573,
+ "▁arms": 5574,
+ "▁feed": 5575,
+ "vey": 5576,
+ "hent": 5577,
+ "MAX": 5578,
+ "▁acqu": 5579,
+ "▁apply": 5580,
+ "▁husband": 5581,
+ "▁killed": 5582,
+ "▁Spec": 5583,
+ "entity": 5584,
+ "▁earlier": 5585,
+ "▁Miss": 5586,
+ "▁setting": 5587,
+ "itect": 5588,
+ "▁ded": 5589,
+ "Row": 5590,
+ "▁ran": 5591,
+ "▁Yes": 5592,
+ "▁financial": 5593,
+ "session": 5594,
+ "lear": 5595,
+ "ishing": 5596,
+ "▁nearly": 5597,
+ "▁dur": 5598,
+ "▁machine": 5599,
+ "xff": 5600,
+ "bro": 5601,
+ "▁symbol": 5602,
+ "lands": 5603,
+ "Acc": 5604,
+ "di": 5605,
+ "▁Robert": 5606,
+ "prop": 5607,
+ "urity": 5608,
+ "▁#####": 5609,
+ "▁walked": 5610,
+ "▁international": 5611,
+ "▁Е": 5612,
+ "Yes": 5613,
+ "▁release": 5614,
+ "▁starting": 5615,
+ "static": 5616,
+ "▁bei": 5617,
+ "allow": 5618,
+ "▁People": 5619,
+ "ez": 5620,
+ "▁parameter": 5621,
+ "Cache": 5622,
+ "▁$$": 5623,
+ "ampions": 5624,
+ "▁Mer": 5625,
+ "▁kom": 5626,
+ "leted": 5627,
+ "ois": 5628,
+ "▁Open": 5629,
+ "types": 5630,
+ "▁fue": 5631,
+ "acters": 5632,
+ "▁reference": 5633,
+ "Equals": 5634,
+ "▁aware": 5635,
+ "▁hol": 5636,
+ "▁demand": 5637,
+ "lor": 5638,
+ "▁veh": 5639,
+ "▁notice": 5640,
+ "▁component": 5641,
+ "fn": 5642,
+ "▁analysis": 5643,
+ "match": 5644,
+ "▁effective": 5645,
+ "product": 5646,
+ "ник": 5647,
+ "▁legal": 5648,
+ "ей": 5649,
+ "semb": 5650,
+ "▁located": 5651,
+ "▁су": 5652,
+ "QL": 5653,
+ "inct": 5654,
+ "eto": 5655,
+ "Draw": 5656,
+ "▁scale": 5657,
+ "ров": 5658,
+ "▁wants": 5659,
+ "How": 5660,
+ "▁wel": 5661,
+ "isions": 5662,
+ "▁deliver": 5663,
+ "under": 5664,
+ "▁deb": 5665,
+ "▁ju": 5666,
+ "values": 5667,
+ "▁sister": 5668,
+ "ков": 5669,
+ "▁Create": 5670,
+ "▁Inc": 5671,
+ "▁aux": 5672,
+ "▁White": 5673,
+ "Menu": 5674,
+ "aud": 5675,
+ "resource": 5676,
+ "▁cab": 5677,
+ "▁lif": 5678,
+ "▁culture": 5679,
+ "iche": 5680,
+ "▁whatever": 5681,
+ "▁designed": 5682,
+ "▁repe": 5683,
+ "▁Mont": 5684,
+ "▁charge": 5685,
+ "Names": 5686,
+ "▁insp": 5687,
+ "▁customers": 5688,
+ "osa": 5689,
+ "▁daughter": 5690,
+ "▁East": 5691,
+ "EQ": 5692,
+ "▁opin": 5693,
+ "▁Fre": 5694,
+ "▁seek": 5695,
+ "▁push": 5696,
+ "▁nav": 5697,
+ "▁burn": 5698,
+ "arden": 5699,
+ "hash": 5700,
+ "▁opportunity": 5701,
+ "▁Mat": 5702,
+ "oyal": 5703,
+ "▁pun": 5704,
+ "scale": 5705,
+ "ynamic": 5706,
+ "▁Type": 5707,
+ "iling": 5708,
+ "▁query": 5709,
+ "▁mist": 5710,
+ "ror": 5711,
+ "force": 5712,
+ "▁Once": 5713,
+ "▁medical": 5714,
+ "lie": 5715,
+ "▁student": 5716,
+ "ederal": 5717,
+ "▁lov": 5718,
+ "iform": 5719,
+ "▁altern": 5720,
+ "bin": 5721,
+ "oder": 5722,
+ "▁returns": 5723,
+ "register": 5724,
+ "uts": 5725,
+ "CI": 5726,
+ "▁Tor": 5727,
+ "CR": 5728,
+ "▁Los": 5729,
+ "amily": 5730,
+ "aire": 5731,
+ "++;": 5732,
+ "Controller": 5733,
+ "wide": 5734,
+ "xx": 5735,
+ "rowser": 5736,
+ "▁Book": 5737,
+ "Container": 5738,
+ "pload": 5739,
+ "▁Ev": 5740,
+ "▁tal": 5741,
+ "▁theory": 5742,
+ "eqnarray": 5743,
+ "бе": 5744,
+ "▁reported": 5745,
+ "▁meaning": 5746,
+ "▁sy": 5747,
+ "ribe": 5748,
+ "icate": 5749,
+ "hold": 5750,
+ "▁offers": 5751,
+ "▁templ": 5752,
+ "css": 5753,
+ "▁picture": 5754,
+ "▁async": 5755,
+ "▁stock": 5756,
+ "▁internal": 5757,
+ "ti": 5758,
+ "BO": 5759,
+ "Ver": 5760,
+ "спо": 5761,
+ "▁demon": 5762,
+ "▁laugh": 5763,
+ "▁End": 5764,
+ "▁kon": 5765,
+ "▁ideas": 5766,
+ "▁candid": 5767,
+ "Mem": 5768,
+ "izz": 5769,
+ "refix": 5770,
+ "▁AND": 5771,
+ "egen": 5772,
+ "El": 5773,
+ "▁campaign": 5774,
+ "Http": 5775,
+ "▁Rob": 5776,
+ "ді": 5777,
+ "▁bul": 5778,
+ "▁Ко": 5779,
+ "▁countries": 5780,
+ "».": 5781,
+ "▁expression": 5782,
+ "▁England": 5783,
+ "sf": 5784,
+ "▁certainly": 5785,
+ "agen": 5786,
+ "▁ча": 5787,
+ "▁ANY": 5788,
+ "▁connect": 5789,
+ "FE": 5790,
+ "▁android": 5791,
+ "▁Gold": 5792,
+ "▁oppos": 5793,
+ "overn": 5794,
+ "▁Commun": 5795,
+ ",_": 5796,
+ "asion": 5797,
+ "La": 5798,
+ "▁firm": 5799,
+ "▁Although": 5800,
+ "▁Good": 5801,
+ "▁Law": 5802,
+ "erve": 5803,
+ "▁brand": 5804,
+ "Min": 5805,
+ "fill": 5806,
+ "'],": 5807,
+ "▁Jew": 5808,
+ "iler": 5809,
+ "ingle": 5810,
+ "ithub": 5811,
+ "▁Div": 5812,
+ "▁cert": 5813,
+ "Height": 5814,
+ "rael": 5815,
+ "There": 5816,
+ "itute": 5817,
+ "▁amaz": 5818,
+ "look": 5819,
+ "▁SE": 5820,
+ "▁jo": 5821,
+ "▁pulled": 5822,
+ "▁resources": 5823,
+ "▁Max": 5824,
+ "▁agreed": 5825,
+ "asy": 5826,
+ "▁treatment": 5827,
+ "\">": 5828,
+ "ман": 5829,
+ "▁Err": 5830,
+ "orig": 5831,
+ "cos": 5832,
+ "▁Maybe": 5833,
+ "otal": 5834,
+ "▁train": 5835,
+ "▁Service": 5836,
+ "▁ih": 5837,
+ "▁spirit": 5838,
+ "Comp": 5839,
+ "sqrt": 5840,
+ "▁broad": 5841,
+ "}[": 5842,
+ "▁shape": 5843,
+ "▁doc": 5844,
+ "how": 5845,
+ "▁tag": 5846,
+ "atalog": 5847,
+ "sd": 5848,
+ "▁meas": 5849,
+ "▁Ро": 5850,
+ "▁exception": 5851,
+ "▁Tw": 5852,
+ "▁interesting": 5853,
+ "ATA": 5854,
+ "▁Rel": 5855,
+ "ár": 5856,
+ "▁useful": 5857,
+ "useum": 5858,
+ "▁bottom": 5859,
+ "▁otherwise": 5860,
+ "▁agree": 5861,
+ "cht": 5862,
+ "then": 5863,
+ "▁significant": 5864,
+ "}/": 5865,
+ "▁channel": 5866,
+ "icial": 5867,
+ "тив": 5868,
+ "vare": 5869,
+ "▁enter": 5870,
+ "Eng": 5871,
+ "uj": 5872,
+ "URE": 5873,
+ "queue": 5874,
+ "ono": 5875,
+ "▁contains": 5876,
+ "MI": 5877,
+ "▁nation": 5878,
+ "▁rules": 5879,
+ "fol": 5880,
+ "▁pa": 5881,
+ "arp": 5882,
+ "▁quiet": 5883,
+ "▁thus": 5884,
+ "ipped": 5885,
+ "annot": 5886,
+ "udes": 5887,
+ "():": 5888,
+ "names": 5889,
+ "▁compos": 5890,
+ "▁inj": 5891,
+ "una": 5892,
+ "bind": 5893,
+ "▁fully": 5894,
+ "ras": 5895,
+ "Utils": 5896,
+ "anges": 5897,
+ "dule": 5898,
+ "▁Christian": 5899,
+ "▁reve": 5900,
+ "änd": 5901,
+ "▁collect": 5902,
+ "▁celebr": 5903,
+ "anda": 5904,
+ "ín": 5905,
+ "join": 5906,
+ "▁paid": 5907,
+ "Core": 5908,
+ "Ge": 5909,
+ ".$": 5910,
+ "▁fif": 5911,
+ "▁uma": 5912,
+ "▁~": 5913,
+ "ervices": 5914,
+ "▁recently": 5915,
+ "desc": 5916,
+ "▁heavy": 5917,
+ "▁rule": 5918,
+ "▁Please": 5919,
+ "psi": 5920,
+ "▁console": 5921,
+ "▁fort": 5922,
+ ".\\": 5923,
+ "▁Washington": 5924,
+ "▁gar": 5925,
+ "▁Group": 5926,
+ "▁interview": 5927,
+ "anned": 5928,
+ "sql": 5929,
+ "▁anc": 5930,
+ "ја": 5931,
+ "Pack": 5932,
+ "▁Club": 5933,
+ "▁mask": 5934,
+ "▁concept": 5935,
+ "▁['": 5936,
+ "▁selected": 5937,
+ "▁Use": 5938,
+ "▁ele": 5939,
+ "ears": 5940,
+ "▁race": 5941,
+ "hy": 5942,
+ "Om": 5943,
+ "▁steps": 5944,
+ "ila": 5945,
+ "ests": 5946,
+ "eds": 5947,
+ "▁street": 5948,
+ "ners": 5949,
+ "▁birth": 5950,
+ "pop": 5951,
+ "▁ли": 5952,
+ "MB": 5953,
+ "кра": 5954,
+ "cir": 5955,
+ "epsilon": 5956,
+ "▁constant": 5957,
+ "ques": 5958,
+ "adas": 5959,
+ "▁knows": 5960,
+ "▁Py": 5961,
+ "cles": 5962,
+ "▁cit": 5963,
+ "▁pair": 5964,
+ "inese": 5965,
+ "▁Peter": 5966,
+ "▁finished": 5967,
+ "▁master": 5968,
+ "▁twenty": 5969,
+ "▁fell": 5970,
+ "▁central": 5971,
+ "▁mes": 5972,
+ "rev": 5973,
+ "STAT": 5974,
+ "stat": 5975,
+ "▁allows": 5976,
+ "▁gro": 5977,
+ "Click": 5978,
+ "▁stories": 5979,
+ "Fe": 5980,
+ "år": 5981,
+ "▁baby": 5982,
+ "encia": 5983,
+ "▁einer": 5984,
+ "Are": 5985,
+ "ebug": 5986,
+ "store": 5987,
+ "\",\"": 5988,
+ "lam": 5989,
+ "▁sv": 5990,
+ "ции": 5991,
+ "NULL": 5992,
+ "▁Leg": 5993,
+ "▁movie": 5994,
+ "▁hous": 5995,
+ "▁learned": 5996,
+ "bon": 5997,
+ "▁transfer": 5998,
+ "ifornia": 5999,
+ "psilon": 6000,
+ "▁Soft": 6001,
+ "▁commer": 6002,
+ "▁hadn": 6003,
+ "▁Ein": 6004,
+ "▁Two": 6005,
+ "craft": 6006,
+ "Process": 6007,
+ "▁под": 6008,
+ "argin": 6009,
+ "▁estim": 6010,
+ "▁Mem": 6011,
+ "ika": 6012,
+ "▁Tod": 6013,
+ "duc": 6014,
+ "▁danger": 6015,
+ "rive": 6016,
+ "Don": 6017,
+ "▁Que": 6018,
+ "hal": 6019,
+ "▁mm": 6020,
+ "▁Sur": 6021,
+ "Order": 6022,
+ "▁distribution": 6023,
+ "fa": 6024,
+ "▁Many": 6025,
+ "plicit": 6026,
+ "Empty": 6027,
+ "Handle": 6028,
+ "▁token": 6029,
+ "▁epis": 6030,
+ "▁assist": 6031,
+ "▁purpose": 6032,
+ "▁ц": 6033,
+ "NU": 6034,
+ "iders": 6035,
+ "rate": 6036,
+ "They": 6037,
+ "Parameter": 6038,
+ "Dec": 6039,
+ "▁strugg": 6040,
+ "▁shoot": 6041,
+ "IV": 6042,
+ "▁Great": 6043,
+ "▁Sil": 6044,
+ "▁loved": 6045,
+ "▁click": 6046,
+ "▁reserv": 6047,
+ "▁ве": 6048,
+ "▁spread": 6049,
+ "▁og": 6050,
+ "▁${": 6051,
+ "▁miles": 6052,
+ "▁successful": 6053,
+ "oj": 6054,
+ "▁Direct": 6055,
+ "▁ax": 6056,
+ "▁growth": 6057,
+ "Work": 6058,
+ "▁church": 6059,
+ "Inst": 6060,
+ "ICE": 6061,
+ "sten": 6062,
+ "род": 6063,
+ "▁Center": 6064,
+ "ses": 6065,
+ "got": 6066,
+ "delete": 6067,
+ "▁Ma": 6068,
+ "%%": 6069,
+ "▁crow": 6070,
+ "DF": 6071,
+ "front": 6072,
+ "▁blog": 6073,
+ "▁computer": 6074,
+ "ная": 6075,
+ "▁mir": 6076,
+ "▁Super": 6077,
+ "','": 6078,
+ "▁multi": 6079,
+ "▁gru": 6080,
+ "▁Jo": 6081,
+ "▁Canada": 6082,
+ "▁Thomas": 6083,
+ "▁larger": 6084,
+ "▁compar": 6085,
+ "Current": 6086,
+ "that": 6087,
+ "▁drop": 6088,
+ "ент": 6089,
+ "▁Republic": 6090,
+ "▁dise": 6091,
+ "▁effects": 6092,
+ "▁girls": 6093,
+ "encies": 6094,
+ "ellig": 6095,
+ "▁Note": 6096,
+ "▁Associ": 6097,
+ "▁uses": 6098,
+ "elled": 6099,
+ "▁warm": 6100,
+ "thread": 6101,
+ "font": 6102,
+ "▁zum": 6103,
+ "▁follows": 6104,
+ "▁whom": 6105,
+ "TA": 6106,
+ "▁wild": 6107,
+ "▁AR": 6108,
+ "iable": 6109,
+ "▁True": 6110,
+ "Position": 6111,
+ "▁sell": 6112,
+ "cher": 6113,
+ "▁Bus": 6114,
+ "▁lean": 6115,
+ "ACE": 6116,
+ "▁served": 6117,
+ "hw": 6118,
+ "▁Cur": 6119,
+ "▁north": 6120,
+ "Dat": 6121,
+ "▁>>": 6122,
+ "command": 6123,
+ "atz": 6124,
+ "▁mal": 6125,
+ "став": 6126,
+ "▁Press": 6127,
+ "▁characters": 6128,
+ "▁zero": 6129,
+ "AGE": 6130,
+ "rapper": 6131,
+ "▁kitchen": 6132,
+ "aming": 6133,
+ "▁restr": 6134,
+ "XX": 6135,
+ "▁College": 6136,
+ "▁Array": 6137,
+ "▁fresh": 6138,
+ "▁shift": 6139,
+ "▁specified": 6140,
+ "plete": 6141,
+ "ITE": 6142,
+ "▁Camp": 6143,
+ "rial": 6144,
+ "cb": 6145,
+ "▁TH": 6146,
+ "IB": 6147,
+ "osen": 6148,
+ "▁ú": 6149,
+ "▁params": 6150,
+ "ignment": 6151,
+ "adding": 6152,
+ "▁degree": 6153,
+ "Local": 6154,
+ "Oh": 6155,
+ "▁zur": 6156,
+ "▁levels": 6157,
+ "CS": 6158,
+ "finished": 6159,
+ "Case": 6160,
+ "riage": 6161,
+ "Vector": 6162,
+ "▁sea": 6163,
+ "antic": 6164,
+ "▁League": 6165,
+ "▁therefore": 6166,
+ "One": 6167,
+ "Return": 6168,
+ "Access": 6169,
+ "vas": 6170,
+ "▁ос": 6171,
+ "▁rat": 6172,
+ "Big": 6173,
+ "▁behavior": 6174,
+ "kr": 6175,
+ "▁undefined": 6176,
+ "▁Es": 6177,
+ "▁appeared": 6178,
+ "eles": 6179,
+ "▁WAR": 6180,
+ "Stat": 6181,
+ "▁Google": 6182,
+ "▁credit": 6183,
+ "▁File": 6184,
+ "anging": 6185,
+ "house": 6186,
+ "romise": 6187,
+ "gent": 6188,
+ "▁habit": 6189,
+ "▁society": 6190,
+ "▁encour": 6191,
+ "▁paint": 6192,
+ "pet": 6193,
+ "▁UK": 6194,
+ "aws": 6195,
+ "onom": 6196,
+ "Gl": 6197,
+ "}_{\\": 6198,
+ "eless": 6199,
+ "emy": 6200,
+ "▁Cong": 6201,
+ "▁developed": 6202,
+ "▁images": 6203,
+ "▁ö": 6204,
+ "▁font": 6205,
+ "clear": 6206,
+ "gin": 6207,
+ "▁Lord": 6208,
+ "▁transport": 6209,
+ "▁::": 6210,
+ "▁cup": 6211,
+ "ulate": 6212,
+ "▁During": 6213,
+ "priv": 6214,
+ "▁extrem": 6215,
+ "▁Di": 6216,
+ "▁doubt": 6217,
+ "Py": 6218,
+ "ifying": 6219,
+ "split": 6220,
+ "ego": 6221,
+ "github": 6222,
+ "▁),": 6223,
+ "ROM": 6224,
+ "▁chair": 6225,
+ "▁trade": 6226,
+ "▁nicht": 6227,
+ "Top": 6228,
+ "Store": 6229,
+ "▁parte": 6230,
+ "project": 6231,
+ "nia": 6232,
+ "▁від": 6233,
+ "war": 6234,
+ "▁Prof": 6235,
+ "▁caught": 6236,
+ "Thread": 6237,
+ "ства": 6238,
+ "author": 6239,
+ "▁doll": 6240,
+ "▁harm": 6241,
+ "▁Gen": 6242,
+ "tree": 6243,
+ "etime": 6244,
+ "cfg": 6245,
+ "▁guys": 6246,
+ "▁California": 6247,
+ "▁Green": 6248,
+ "▁movement": 6249,
+ "iej": 6250,
+ "▁statement": 6251,
+ "▁seeing": 6252,
+ "▁haven": 6253,
+ "vention": 6254,
+ "SL": 6255,
+ "chedul": 6256,
+ "iert": 6257,
+ "▁primary": 6258,
+ "▁civil": 6259,
+ "rian": 6260,
+ "▁button": 6261,
+ "▁lived": 6262,
+ "Pass": 6263,
+ "sor": 6264,
+ "▁watching": 6265,
+ "▁skills": 6266,
+ "tee": 6267,
+ "Level": 6268,
+ "▁scient": 6269,
+ "hs": 6270,
+ "▁agre": 6271,
+ "cat": 6272,
+ "▁tend": 6273,
+ "▁Mill": 6274,
+ "▁Cap": 6275,
+ "ORD": 6276,
+ "gle": 6277,
+ "▁сво": 6278,
+ "»,": 6279,
+ "▁ahead": 6280,
+ "vest": 6281,
+ "▁Jose": 6282,
+ "ischer": 6283,
+ "și": 6284,
+ "▁leaving": 6285,
+ "▁для": 6286,
+ "▁south": 6287,
+ "▁consum": 6288,
+ "Range": 6289,
+ "▁activities": 6290,
+ "Sec": 6291,
+ "▁sales": 6292,
+ "▁fix": 6293,
+ "▁jed": 6294,
+ "rum": 6295,
+ "vector": 6296,
+ "▁spot": 6297,
+ "▁manufact": 6298,
+ "кт": 6299,
+ "orrow": 6300,
+ "sign": 6301,
+ "▁college": 6302,
+ "▁driver": 6303,
+ "▁definitely": 6304,
+ "▁spend": 6305,
+ "mission": 6306,
+ "зу": 6307,
+ "atively": 6308,
+ "bi": 6309,
+ "Callback": 6310,
+ "▁particularly": 6311,
+ "▁hell": 6312,
+ "▁pool": 6313,
+ "PRE": 6314,
+ "▁clearly": 6315,
+ "PT": 6316,
+ "othes": 6317,
+ "▁Id": 6318,
+ "Location": 6319,
+ "▁Run": 6320,
+ "▁fixed": 6321,
+ "▁Hand": 6322,
+ "bal": 6323,
+ "double": 6324,
+ "Can": 6325,
+ "Omega": 6326,
+ "▁challeng": 6327,
+ "▁standing": 6328,
+ "iten": 6329,
+ "▁mechan": 6330,
+ "▁durch": 6331,
+ "▁dell": 6332,
+ "▁raised": 6333,
+ "▁weak": 6334,
+ "▁Du": 6335,
+ "grad": 6336,
+ "▁scene": 6337,
+ "poss": 6338,
+ "▁ton": 6339,
+ "▁earth": 6340,
+ "ulations": 6341,
+ "▁strength": 6342,
+ "aked": 6343,
+ "▁remain": 6344,
+ "▁Bi": 6345,
+ "▁customer": 6346,
+ "range": 6347,
+ "▁interested": 6348,
+ "ONE": 6349,
+ "▁coff": 6350,
+ "require": 6351,
+ "▁Only": 6352,
+ "▁Web": 6353,
+ "▁farm": 6354,
+ "▁activity": 6355,
+ "▁rout": 6356,
+ "bling": 6357,
+ "SY": 6358,
+ "▁Richard": 6359,
+ "▁Ref": 6360,
+ "▁кон": 6361,
+ "▁jun": 6362,
+ "born": 6363,
+ "ijn": 6364,
+ "Configuration": 6365,
+ "uman": 6366,
+ "EE": 6367,
+ "▁married": 6368,
+ "▁За": 6369,
+ "▁fat": 6370,
+ "▁kid": 6371,
+ "▁Tur": 6372,
+ "▁offered": 6373,
+ "nic": 6374,
+ "▁Big": 6375,
+ "Gamma": 6376,
+ "▁Health": 6377,
+ "▁TR": 6378,
+ "▁się": 6379,
+ "▁construction": 6380,
+ "▁Church": 6381,
+ "▁Bet": 6382,
+ "bus": 6383,
+ "▁earn": 6384,
+ "rict": 6385,
+ "▁пра": 6386,
+ "▁brain": 6387,
+ "▁fra": 6388,
+ "▁Op": 6389,
+ "FIG": 6390,
+ "ema": 6391,
+ "▁European": 6392,
+ "▁Saint": 6393,
+ "ARE": 6394,
+ "uri": 6395,
+ "▁River": 6396,
+ "{}": 6397,
+ "▁sitting": 6398,
+ "▁understanding": 6399,
+ "▁plans": 6400,
+ "ropri": 6401,
+ "▁older": 6402,
+ "▁pressure": 6403,
+ "Impl": 6404,
+ "▁peace": 6405,
+ "Connection": 6406,
+ "▁fi": 6407,
+ "rich": 6408,
+ "▁shut": 6409,
+ "apers": 6410,
+ "Port": 6411,
+ "▁Look": 6412,
+ "rim": 6413,
+ "auth": 6414,
+ "auto": 6415,
+ "▁highly": 6416,
+ "▁unless": 6417,
+ "▁Wal": 6418,
+ "▁ren": 6419,
+ "ws": 6420,
+ "▁core": 6421,
+ "(-": 6422,
+ "▁clim": 6423,
+ "ruit": 6424,
+ "▁callback": 6425,
+ "hest": 6426,
+ "▁Charles": 6427,
+ "▁Long": 6428,
+ "}=": 6429,
+ "ър": 6430,
+ "▁shared": 6431,
+ "ulated": 6432,
+ "gorithm": 6433,
+ "▁Home": 6434,
+ "▁village": 6435,
+ "ees": 6436,
+ "sv": 6437,
+ "▁restaur": 6438,
+ "rey": 6439,
+ "▁Cast": 6440,
+ "▁Person": 6441,
+ "кий": 6442,
+ "▁organiz": 6443,
+ "▁Rad": 6444,
+ "ponents": 6445,
+ "▁werden": 6446,
+ "▁bow": 6447,
+ "sen": 6448,
+ "ami": 6449,
+ "Interface": 6450,
+ "▁basis": 6451,
+ "▁Company": 6452,
+ "ernel": 6453,
+ "itu": 6454,
+ "Hash": 6455,
+ "▁aan": 6456,
+ "▁х": 6457,
+ "▁smile": 6458,
+ "xml": 6459,
+ "▁scen": 6460,
+ "amm": 6461,
+ "tool": 6462,
+ "aria": 6463,
+ "▁accur": 6464,
+ "settings": 6465,
+ "▁Jesus": 6466,
+ "acement": 6467,
+ "power": 6468,
+ "(!": 6469,
+ "▁calls": 6470,
+ "▁basic": 6471,
+ "▁settings": 6472,
+ "ript": 6473,
+ "pool": 6474,
+ "ctors": 6475,
+ "▁Foundation": 6476,
+ "▁weap": 6477,
+ "KEY": 6478,
+ "foot": 6479,
+ "▁radio": 6480,
+ "▁helped": 6481,
+ "mann": 6482,
+ "▁jump": 6483,
+ "▁tick": 6484,
+ "▁growing": 6485,
+ "aten": 6486,
+ "real": 6487,
+ "▁increasing": 6488,
+ "Device": 6489,
+ "varepsilon": 6490,
+ "▁sets": 6491,
+ "▁advant": 6492,
+ "Open": 6493,
+ "▁reasons": 6494,
+ "▁supposed": 6495,
+ "oes": 6496,
+ "ede": 6497,
+ "teen": 6498,
+ "ifdef": 6499,
+ "▁delete": 6500,
+ "▁&=": 6501,
+ "▁Bill": 6502,
+ "▁aim": 6503,
+ "▁Ok": 6504,
+ "▁Av": 6505,
+ "reci": 6506,
+ "acks": 6507,
+ "iste": 6508,
+ "Properties": 6509,
+ "▁tmp": 6510,
+ "▁dei": 6511,
+ "PER": 6512,
+ "DC": 6513,
+ "sta": 6514,
+ "нии": 6515,
+ "▁limited": 6516,
+ "▁greater": 6517,
+ "description": 6518,
+ "ori": 6519,
+ "aints": 6520,
+ "▁hy": 6521,
+ "▁Mel": 6522,
+ "▁CH": 6523,
+ "cons": 6524,
+ "▁surround": 6525,
+ "▁Who": 6526,
+ "arc": 6527,
+ "▁telev": 6528,
+ "itution": 6529,
+ "▁equal": 6530,
+ "кі": 6531,
+ "▁Israel": 6532,
+ "äh": 6533,
+ "▁Caption": 6534,
+ "▁exerc": 6535,
+ "empor": 6536,
+ "▁++": 6537,
+ "▁lib": 6538,
+ "make": 6539,
+ "▁MA": 6540,
+ "copy": 6541,
+ "friend": 6542,
+ "▁кото": 6543,
+ "▁damage": 6544,
+ "▁\\,": 6545,
+ "oded": 6546,
+ "▁none": 6547,
+ "▁evalu": 6548,
+ "ston": 6549,
+ ">,": 6550,
+ "FOR": 6551,
+ "▁norm": 6552,
+ "appe": 6553,
+ "Session": 6554,
+ "▁adult": 6555,
+ "▁hospital": 6556,
+ "▁recommend": 6557,
+ "property": 6558,
+ "stein": 6559,
+ "final": 6560,
+ "▁nu": 6561,
+ "second": 6562,
+ "▁aspect": 6563,
+ "\")]": 6564,
+ "жен": 6565,
+ "amento": 6566,
+ "▁rac": 6567,
+ "save": 6568,
+ "▁football": 6569,
+ "Ab": 6570,
+ "ungs": 6571,
+ "abil": 6572,
+ "▁Arch": 6573,
+ "system": 6574,
+ "hist": 6575,
+ "▁luck": 6576,
+ "render": 6577,
+ "▁sein": 6578,
+ "ioni": 6579,
+ "▁rot": 6580,
+ "▁corner": 6581,
+ "▁appropri": 6582,
+ "▁Software": 6583,
+ "▁tele": 6584,
+ "Delete": 6585,
+ "▁According": 6586,
+ "▁prison": 6587,
+ "▁lic": 6588,
+ "▁ми": 6589,
+ "term": 6590,
+ "sets": 6591,
+ "▁vel": 6592,
+ "▁rank": 6593,
+ "▁existing": 6594,
+ "▁Vir": 6595,
+ "▁trip": 6596,
+ "▁му": 6597,
+ "avax": 6598,
+ "▁ris": 6599,
+ "▁define": 6600,
+ "▁heat": 6601,
+ "car": 6602,
+ "▁convert": 6603,
+ "email": 6604,
+ "▁Under": 6605,
+ "▁Ш": 6606,
+ "▁Grand": 6607,
+ "▁exists": 6608,
+ "sys": 6609,
+ "eff": 6610,
+ "▁Top": 6611,
+ "▁č": 6612,
+ "▁tempor": 6613,
+ "▁arguments": 6614,
+ "▁supported": 6615,
+ "ensed": 6616,
+ "▁Francis": 6617,
+ "▁coord": 6618,
+ "▁achieve": 6619,
+ "▁Name": 6620,
+ "▁Jahr": 6621,
+ "▁Gi": 6622,
+ "she": 6623,
+ "▁Dev": 6624,
+ "▁alla": 6625,
+ "▁WIT": 6626,
+ "agment": 6627,
+ "custom": 6628,
+ "alls": 6629,
+ "&&": 6630,
+ "WE": 6631,
+ "▁holding": 6632,
+ "prototype": 6633,
+ "▁fing": 6634,
+ "▁bag": 6635,
+ "▁Party": 6636,
+ "stack": 6637,
+ "▁economic": 6638,
+ "▁Gal": 6639,
+ "idents": 6640,
+ "▁Jun": 6641,
+ "▁showed": 6642,
+ "osh": 6643,
+ "▁Bay": 6644,
+ "mail": 6645,
+ "▁SO": 6646,
+ "▁\"<": 6647,
+ "graphics": 6648,
+ "▁fu": 6649,
+ "click": 6650,
+ "▁battle": 6651,
+ "{{": 6652,
+ "▁Event": 6653,
+ "rior": 6654,
+ "chaft": 6655,
+ "▁favorite": 6656,
+ "usive": 6657,
+ "support": 6658,
+ "bm": 6659,
+ "Kind": 6660,
+ "▁safety": 6661,
+ "▁Ent": 6662,
+ "cup": 6663,
+ "▁Australia": 6664,
+ "▁destroy": 6665,
+ "▁organization": 6666,
+ "iden": 6667,
+ "################": 6668,
+ "dec": 6669,
+ "▁za": 6670,
+ "▁seven": 6671,
+ "arely": 6672,
+ "▁flag": 6673,
+ "Dir": 6674,
+ "▁Carl": 6675,
+ "▁doctor": 6676,
+ "▁variety": 6677,
+ "▁Lin": 6678,
+ "▁tom": 6679,
+ "^{(": 6680,
+ "Bo": 6681,
+ "antes": 6682,
+ "▁mine": 6683,
+ "▁Mit": 6684,
+ "▁describe": 6685,
+ "Args": 6686,
+ "LS": 6687,
+ "API": 6688,
+ "▁Luc": 6689,
+ "phone": 6690,
+ "▁science": 6691,
+ "▁Oper": 6692,
+ "Next": 6693,
+ "▁investig": 6694,
+ "▁demonstr": 6695,
+ "▁Govern": 6696,
+ "▁objects": 6697,
+ "▁Louis": 6698,
+ "▁Returns": 6699,
+ "▁han": 6700,
+ "nam": 6701,
+ "▁comme": 6702,
+ "▁presence": 6703,
+ "▁pel": 6704,
+ "▁detect": 6705,
+ ")=": 6706,
+ "▁Chinese": 6707,
+ "▁rich": 6708,
+ "▁classes": 6709,
+ "▁expand": 6710,
+ "▁Dom": 6711,
+ "▁Dec": 6712,
+ "sn": 6713,
+ "peed": 6714,
+ "▁Jim": 6715,
+ "should": 6716,
+ "▁Smith": 6717,
+ "▁pages": 6718,
+ "▁Jean": 6719,
+ "rics": 6720,
+ "▁Sund": 6721,
+ "ads": 6722,
+ "▁Their": 6723,
+ "unicip": 6724,
+ "ву": 6725,
+ "▁download": 6726,
+ "▁stress": 6727,
+ "▁Pet": 6728,
+ "menu": 6729,
+ "reme": 6730,
+ "▁compared": 6731,
+ "Ste": 6732,
+ "IND": 6733,
+ "container": 6734,
+ "▁Indian": 6735,
+ "oren": 6736,
+ "▁ses": 6737,
+ "▁Whe": 6738,
+ "▁roku": 6739,
+ "▁established": 6740,
+ "▁generally": 6741,
+ "▁fle": 6742,
+ "__(": 6743,
+ "=\"+": 6744,
+ "Var": 6745,
+ "▁Make": 6746,
+ "▁removed": 6747,
+ "zz": 6748,
+ "ün": 6749,
+ "▁mix": 6750,
+ "erk": 6751,
+ "iation": 6752,
+ "outer": 6753,
+ "SK": 6754,
+ "▁becomes": 6755,
+ "▁Hall": 6756,
+ "scious": 6757,
+ "▁watched": 6758,
+ "▁gather": 6759,
+ "▁Result": 6760,
+ "proof": 6761,
+ "pay": 6762,
+ "▁produced": 6763,
+ "▁|=": 6764,
+ "▁border": 6765,
+ "▁din": 6766,
+ "▁script": 6767,
+ "▁actions": 6768,
+ "▁mas": 6769,
+ "ща": 6770,
+ "ooth": 6771,
+ "▁Techn": 6772,
+ "Json": 6773,
+ "▁filled": 6774,
+ "ден": 6775,
+ "undle": 6776,
+ "сту": 6777,
+ "Tool": 6778,
+ "▁king": 6779,
+ "▁ven": 6780,
+ "stra": 6781,
+ "▁predict": 6782,
+ "▁lui": 6783,
+ "▁WARRAN": 6784,
+ "▁Fun": 6785,
+ "Script": 6786,
+ "▁powerful": 6787,
+ "▁lose": 6788,
+ "atically": 6789,
+ "▁daily": 6790,
+ "▁ring": 6791,
+ "▁arrived": 6792,
+ "Stack": 6793,
+ "scope": 6794,
+ "▁Back": 6795,
+ "elij": 6796,
+ "▁ze": 6797,
+ "keys": 6798,
+ "{\"": 6799,
+ "VID": 6800,
+ "▁license": 6801,
+ "what": 6802,
+ "▁proced": 6803,
+ "rant": 6804,
+ "estival": 6805,
+ "agram": 6806,
+ "▁LO": 6807,
+ "▁Henry": 6808,
+ "▁flags": 6809,
+ "Down": 6810,
+ "scription": 6811,
+ "▁families": 6812,
+ "isse": 6813,
+ "bour": 6814,
+ "▁Bur": 6815,
+ "—\"": 6816,
+ "▁brief": 6817,
+ "▁creating": 6818,
+ "▁clients": 6819,
+ "rangle": 6820,
+ "▁amazing": 6821,
+ "▁sind": 6822,
+ "▁covered": 6823,
+ "Well": 6824,
+ "сте": 6825,
+ "тор": 6826,
+ "▁Bas": 6827,
+ "total": 6828,
+ "▁Init": 6829,
+ "▁sand": 6830,
+ "Unit": 6831,
+ "▁murder": 6832,
+ "▁bright": 6833,
+ "▁trav": 6834,
+ "icans": 6835,
+ "▁attribute": 6836,
+ "fc": 6837,
+ "▁placed": 6838,
+ "EST": 6839,
+ "Vari": 6840,
+ "▁cos": 6841,
+ "▁attract": 6842,
+ "anel": 6843,
+ "}).": 6844,
+ "bytes": 6845,
+ "▁parse": 6846,
+ "▁belong": 6847,
+ "BN": 6848,
+ "▁Sol": 6849,
+ "Po": 6850,
+ "`,": 6851,
+ "▁calling": 6852,
+ "▁?>": 6853,
+ "▁iter": 6854,
+ "▁url": 6855,
+ "▁evening": 6856,
+ "reek": 6857,
+ "▁honest": 6858,
+ "▁director": 6859,
+ "RC": 6860,
+ "▁solid": 6861,
+ "▁phil": 6862,
+ "iene": 6863,
+ "FAULT": 6864,
+ "cope": 6865,
+ "▁History": 6866,
+ "▁Team": 6867,
+ "reedom": 6868,
+ "▁ru": 6869,
+ "UB": 6870,
+ "▁worse": 6871,
+ "imo": 6872,
+ "Mat": 6873,
+ "▁Mex": 6874,
+ "actor": 6875,
+ "▁vor": 6876,
+ "ться": 6877,
+ "▁experiment": 6878,
+ "▁Play": 6879,
+ "▁Another": 6880,
+ "▁happens": 6881,
+ "uan": 6882,
+ "▁patients": 6883,
+ "▁rend": 6884,
+ "▁Mo": 6885,
+ "▁Tex": 6886,
+ "▁wed": 6887,
+ "tn": 6888,
+ "insert": 6889,
+ "▁па": 6890,
+ "▁anti": 6891,
+ "Match": 6892,
+ "ampionship": 6893,
+ "▁forces": 6894,
+ "▁Hot": 6895,
+ "▁phase": 6896,
+ "▁template": 6897,
+ "stop": 6898,
+ "icated": 6899,
+ "▁managed": 6900,
+ "wait": 6901,
+ "▁*(": 6902,
+ "GB": 6903,
+ "▁appoint": 6904,
+ "ła": 6905,
+ "▁stick": 6906,
+ "▁FOR": 6907,
+ "▁Vis": 6908,
+ "tor": 6909,
+ "▁př": 6910,
+ "quest": 6911,
+ "uses": 6912,
+ "\");\r": 6913,
+ "▁suddenly": 6914,
+ "éc": 6915,
+ "ND": 6916,
+ "urop": 6917,
+ "ред": 6918,
+ "▁insurance": 6919,
+ "access": 6920,
+ "unfinished": 6921,
+ "▁tamb": 6922,
+ "▁sac": 6923,
+ "▁Court": 6924,
+ "▁missing": 6925,
+ "▁Where": 6926,
+ "▁Sum": 6927,
+ "}^{\\": 6928,
+ "▁sua": 6929,
+ "_,": 6930,
+ "▁thick": 6931,
+ "▁Trump": 6932,
+ "▁operations": 6933,
+ "FS": 6934,
+ "▁deux": 6935,
+ "dz": 6936,
+ "Template": 6937,
+ "▁\"/": 6938,
+ "▁odd": 6939,
+ "▁reality": 6940,
+ "▁teams": 6941,
+ "▁cer": 6942,
+ "oma": 6943,
+ "▁și": 6944,
+ "▁cloud": 6945,
+ "▁Department": 6946,
+ "Ne": 6947,
+ "▁requires": 6948,
+ "items": 6949,
+ "▁III": 6950,
+ "rightarrow": 6951,
+ ")->": 6952,
+ "▁writer": 6953,
+ "replace": 6954,
+ "▁thr": 6955,
+ "jen": 6956,
+ "▁ot": 6957,
+ "▁occup": 6958,
+ "▁eventually": 6959,
+ "▁Math": 6960,
+ "▁conserv": 6961,
+ "amer": 6962,
+ "▁Fort": 6963,
+ "▁dry": 6964,
+ "▁sexual": 6965,
+ "▁costs": 6966,
+ "▁forms": 6967,
+ "▁Vict": 6968,
+ "PAR": 6969,
+ "framework": 6970,
+ "▁ди": 6971,
+ "Operation": 6972,
+ "зна": 6973,
+ "which": 6974,
+ "▁tight": 6975,
+ "Invalid": 6976,
+ "▁partner": 6977,
+ "▁пред": 6978,
+ "▁thank": 6979,
+ "▁guard": 6980,
+ "hem": 6981,
+ "Body": 6982,
+ "▁emot": 6983,
+ "IX": 6984,
+ "fast": 6985,
+ "що": 6986,
+ "ño": 6987,
+ "night": 6988,
+ "▁Sci": 6989,
+ "ника": 6990,
+ "▁TO": 6991,
+ "▁individuals": 6992,
+ "сси": 6993,
+ "}),": 6994,
+ "False": 6995,
+ "(\"%": 6996,
+ "▁optim": 6997,
+ "▁-->": 6998,
+ "▁factor": 6999,
+ "▁smaller": 7000,
+ "▁contain": 7001,
+ "spect": 7002,
+ "Engine": 7003,
+ "▁announced": 7004,
+ "▁Democr": 7005,
+ "▁rob": 7006,
+ "▁flat": 7007,
+ "osoph": 7008,
+ "Search": 7009,
+ "ahl": 7010,
+ "▁Exception": 7011,
+ "▁Ol": 7012,
+ "equals": 7013,
+ "▁unter": 7014,
+ "shape": 7015,
+ "NS": 7016,
+ "Obj": 7017,
+ "▁species": 7018,
+ "weight": 7019,
+ "you": 7020,
+ "▁este": 7021,
+ "▁View": 7022,
+ "▁mission": 7023,
+ "▁journal": 7024,
+ "Values": 7025,
+ "▁einem": 7026,
+ "ismo": 7027,
+ "▁projects": 7028,
+ "▁Das": 7029,
+ "rible": 7030,
+ "▁serve": 7031,
+ "▁opening": 7032,
+ "▁hur": 7033,
+ "▁programs": 7034,
+ "▁USA": 7035,
+ "iliar": 7036,
+ "idos": 7037,
+ "Br": 7038,
+ "estamp": 7039,
+ "▁tools": 7040,
+ "anner": 7041,
+ "RT": 7042,
+ "▁Start": 7043,
+ "▁bath": 7044,
+ "▁coffee": 7045,
+ "orter": 7046,
+ "internal": 7047,
+ "files": 7048,
+ "INVAL": 7049,
+ "ako": 7050,
+ "dt": 7051,
+ "▁Second": 7052,
+ "▁alloc": 7053,
+ "▁ended": 7054,
+ "acional": 7055,
+ "▁manager": 7056,
+ "▁Sun": 7057,
+ "agg": 7058,
+ "▁leader": 7059,
+ "olved": 7060,
+ "▁что": 7061,
+ "▁traditional": 7062,
+ "shot": 7063,
+ "rup": 7064,
+ "CF": 7065,
+ "▁Each": 7066,
+ "wr": 7067,
+ "▁Som": 7068,
+ "▁materials": 7069,
+ "▁msg": 7070,
+ "▁syn": 7071,
+ "▁produce": 7072,
+ "▁storage": 7073,
+ "subsection": 7074,
+ "▁Sie": 7075,
+ "▁IP": 7076,
+ "CESS": 7077,
+ "▁wa": 7078,
+ "Record": 7079,
+ "▁marketing": 7080,
+ "plet": 7081,
+ "Dialog": 7082,
+ "▁mentioned": 7083,
+ "▁Na": 7084,
+ "▁Union": 7085,
+ "▁API": 7086,
+ "▁negative": 7087,
+ "txt": 7088,
+ "▁easier": 7089,
+ "legal": 7090,
+ "Dep": 7091,
+ "▁novel": 7092,
+ "eur": 7093,
+ "ació": 7094,
+ "▁Bud": 7095,
+ "▁carry": 7096,
+ "schaft": 7097,
+ "▁broken": 7098,
+ "▁trees": 7099,
+ ">();": 7100,
+ "▁emb": 7101,
+ "ieder": 7102,
+ "▁route": 7103,
+ "ikel": 7104,
+ "▁listen": 7105,
+ "ashion": 7106,
+ "▁Mrs": 7107,
+ "▁equipment": 7108,
+ "agger": 7109,
+ "▁Thus": 7110,
+ "▁matrix": 7111,
+ "alla": 7112,
+ "▁Tour": 7113,
+ "▁conversation": 7114,
+ "Mon": 7115,
+ "ournal": 7116,
+ "▁minute": 7117,
+ "Am": 7118,
+ "Api": 7119,
+ "▁forget": 7120,
+ "Me": 7121,
+ "levant": 7122,
+ "temp": 7123,
+ "▁telling": 7124,
+ "move": 7125,
+ "▁independent": 7126,
+ "toString": 7127,
+ "edit": 7128,
+ "▁Jac": 7129,
+ "azz": 7130,
+ "react": 7131,
+ "▁cin": 7132,
+ "▁Prov": 7133,
+ "isted": 7134,
+ "▁hash": 7135,
+ "onna": 7136,
+ "iki": 7137,
+ "▁generated": 7138,
+ "Render": 7139,
+ "▁psych": 7140,
+ "nav": 7141,
+ "▁entr": 7142,
+ "пра": 7143,
+ "rx": 7144,
+ "ATH": 7145,
+ "▁assume": 7146,
+ "Tree": 7147,
+ "sembly": 7148,
+ "▁Matt": 7149,
+ "caption": 7150,
+ "▁solutions": 7151,
+ "▁faith": 7152,
+ "▁digital": 7153,
+ "▁excell": 7154,
+ "▁Version": 7155,
+ "Debug": 7156,
+ "▁жи": 7157,
+ "▁carried": 7158,
+ "reset": 7159,
+ "▁slowly": 7160,
+ "ancing": 7161,
+ "▁owner": 7162,
+ "▁Ter": 7163,
+ "▁Did": 7164,
+ "▁gest": 7165,
+ "▁été": 7166,
+ "▁proof": 7167,
+ "Font": 7168,
+ "▁nob": 7169,
+ "Co": 7170,
+ "▁GNU": 7171,
+ "▁liber": 7172,
+ "itness": 7173,
+ "▁hij": 7174,
+ "▁vert": 7175,
+ "ша": 7176,
+ "FLAG": 7177,
+ "MENT": 7178,
+ "▁Son": 7179,
+ "Mult": 7180,
+ "▁district": 7181,
+ "connect": 7182,
+ "jection": 7183,
+ "lymp": 7184,
+ "▁realized": 7185,
+ "mos": 7186,
+ "ye": 7187,
+ "▁render": 7188,
+ "rio": 7189,
+ "▁interpret": 7190,
+ "▁slightly": 7191,
+ "fix": 7192,
+ "▁studies": 7193,
+ "▁rid": 7194,
+ "atre": 7195,
+ "▁benefits": 7196,
+ "▁Face": 7197,
+ "ivery": 7198,
+ "рия": 7199,
+ "document": 7200,
+ "▁asking": 7201,
+ "Last": 7202,
+ "arante": 7203,
+ "▁Martin": 7204,
+ "▁Ell": 7205,
+ "▁vector": 7206,
+ "▁forced": 7207,
+ "оло": 7208,
+ "PH": 7209,
+ "WR": 7210,
+ "▁Kl": 7211,
+ "▁sky": 7212,
+ "▁strategy": 7213,
+ "ocked": 7214,
+ "▁neck": 7215,
+ "ści": 7216,
+ "OUT": 7217,
+ ")),": 7218,
+ "Custom": 7219,
+ "▁wie": 7220,
+ "▁sweet": 7221,
+ "▁temp": 7222,
+ "▁foreign": 7223,
+ "▁hall": 7224,
+ "astr": 7225,
+ "Ass": 7226,
+ "MODE": 7227,
+ "▁maximum": 7228,
+ "annels": 7229,
+ "▁tip": 7230,
+ "▁seconds": 7231,
+ "▁stack": 7232,
+ "iga": 7233,
+ "▁raise": 7234,
+ "enable": 7235,
+ "oir": 7236,
+ "▁soul": 7237,
+ "Ke": 7238,
+ ")$.": 7239,
+ "▁Tim": 7240,
+ "ALSE": 7241,
+ "iser": 7242,
+ "contin": 7243,
+ "bel": 7244,
+ "▁mad": 7245,
+ "lichen": 7246,
+ "abe": 7247,
+ "safe": 7248,
+ "▁concent": 7249,
+ "bound": 7250,
+ "▁Requ": 7251,
+ "switch": 7252,
+ "▁stone": 7253,
+ "▁transl": 7254,
+ "▁vac": 7255,
+ "andon": 7256,
+ "▁Fore": 7257,
+ "▁sounds": 7258,
+ "▁Pop": 7259,
+ "▁HT": 7260,
+ "lia": 7261,
+ "enter": 7262,
+ "▁helps": 7263,
+ "edy": 7264,
+ "ствен": 7265,
+ "anted": 7266,
+ "▁Its": 7267,
+ "▁Step": 7268,
+ "Icon": 7269,
+ "▁EXPECT": 7270,
+ "ialized": 7271,
+ "Post": 7272,
+ "aze": 7273,
+ "▁Carol": 7274,
+ "▁req": 7275,
+ "▁critical": 7276,
+ "DS": 7277,
+ "▁seat": 7278,
+ "aped": 7279,
+ "▁upper": 7280,
+ "▁Sy": 7281,
+ "▁explain": 7282,
+ "▁'./": 7283,
+ "utils": 7284,
+ "possible": 7285,
+ "▁dont": 7286,
+ "Host": 7287,
+ "▁approxim": 7288,
+ "Async": 7289,
+ "▁grab": 7290,
+ "▁sources": 7291,
+ "▁Mos": 7292,
+ "▁Germany": 7293,
+ "▁rub": 7294,
+ "CHAN": 7295,
+ "▁rain": 7296,
+ "▁truly": 7297,
+ "▁joined": 7298,
+ "▁": 7299,
+ "▁Lo": 7300,
+ "Description": 7301,
+ "akt": 7302,
+ "▁Ann": 7303,
+ "^*": 7304,
+ "idae": 7305,
+ "(:": 7306,
+ "tw": 7307,
+ "Mar": 7308,
+ "produ": 7309,
+ "▁spoke": 7310,
+ "ют": 7311,
+ "▁walking": 7312,
+ "▁nodded": 7313,
+ "Props": 7314,
+ "Enabled": 7315,
+ "irk": 7316,
+ "FILE": 7317,
+ "equal": 7318,
+ "pping": 7319,
+ "oli": 7320,
+ "EV": 7321,
+ "enz": 7322,
+ "eting": 7323,
+ "▁sample": 7324,
+ "▁artist": 7325,
+ "[$": 7326,
+ "ità": 7327,
+ "йо": 7328,
+ "props": 7329,
+ "bu": 7330,
+ "ев": 7331,
+ "▁responsible": 7332,
+ "MT": 7333,
+ "▁caused": 7334,
+ "▁theme": 7335,
+ "▁Was": 7336,
+ "▁Before": 7337,
+ "acle": 7338,
+ "▁року": 7339,
+ "cu": 7340,
+ "DEV": 7341,
+ "▁hung": 7342,
+ "textbf": 7343,
+ "▁spin": 7344,
+ "▁latest": 7345,
+ "entially": 7346,
+ "▁Program": 7347,
+ "Metadata": 7348,
+ "password": 7349,
+ "▁hurt": 7350,
+ "кс": 7351,
+ "▁Aus": 7352,
+ "sey": 7353,
+ "allet": 7354,
+ "xF": 7355,
+ "▁Road": 7356,
+ "ется": 7357,
+ "▁rent": 7358,
+ "ция": 7359,
+ "▁Assert": 7360,
+ "іль": 7361,
+ "ück": 7362,
+ "▁sites": 7363,
+ "Document": 7364,
+ "▁obtained": 7365,
+ "▁ci": 7366,
+ "▁[\"": 7367,
+ "▁completed": 7368,
+ "aset": 7369,
+ "raid": 7370,
+ "▁sorry": 7371,
+ "▁fab": 7372,
+ "▁schools": 7373,
+ "ходи": 7374,
+ "▁scr": 7375,
+ "▁incor": 7376,
+ "▁'/": 7377,
+ "▁spr": 7378,
+ "▁Text": 7379,
+ "▁commercial": 7380,
+ "ingly": 7381,
+ "▁opinion": 7382,
+ "▁Star": 7383,
+ "Sign": 7384,
+ "▁javax": 7385,
+ "wi": 7386,
+ "lat": 7387,
+ "▁Key": 7388,
+ "varphi": 7389,
+ "ды": 7390,
+ "▁connected": 7391,
+ "▁adjust": 7392,
+ "▁Az": 7393,
+ "▁planning": 7394,
+ "---": 7395,
+ "Integer": 7396,
+ "auf": 7397,
+ "expected": 7398,
+ "▁fant": 7399,
+ "▁tou": 7400,
+ "Parent": 7401,
+ "▁Lat": 7402,
+ "▁thoughts": 7403,
+ "▁Jud": 7404,
+ "Parameters": 7405,
+ "Gr": 7406,
+ "ром": 7407,
+ "IA": 7408,
+ "▁Bob": 7409,
+ "lict": 7410,
+ "lan": 7411,
+ "omic": 7412,
+ "▁apart": 7413,
+ "▁trou": 7414,
+ "▁appreci": 7415,
+ "▁Christmas": 7416,
+ "irq": 7417,
+ "thon": 7418,
+ "▁Error": 7419,
+ "▁score": 7420,
+ "rome": 7421,
+ "▁neighbor": 7422,
+ "▁Mur": 7423,
+ "admin": 7424,
+ "▁Film": 7425,
+ "Rect": 7426,
+ "▁configuration": 7427,
+ "▁cs": 7428,
+ "gun": 7429,
+ "channel": 7430,
+ "▁Report": 7431,
+ "▁strateg": 7432,
+ "▁workers": 7433,
+ "fields": 7434,
+ "Schema": 7435,
+ "appa": 7436,
+ "olic": 7437,
+ "EO": 7438,
+ "▁Charl": 7439,
+ "▁Cup": 7440,
+ "png": 7441,
+ "▁Hill": 7442,
+ "owe": 7443,
+ "▁mostly": 7444,
+ "”.": 7445,
+ "▁finish": 7446,
+ "▁Со": 7447,
+ "▁stars": 7448,
+ "player": 7449,
+ "▁inner": 7450,
+ "component": 7451,
+ "tim": 7452,
+ "IE": 7453,
+ "▁ther": 7454,
+ "▁smart": 7455,
+ "▁sad": 7456,
+ "▁Council": 7457,
+ "area": 7458,
+ "lay": 7459,
+ "▁ба": 7460,
+ "▁gradu": 7461,
+ "▁chem": 7462,
+ "▁ho": 7463,
+ "Select": 7464,
+ "▁instr": 7465,
+ "▁kl": 7466,
+ "ifications": 7467,
+ "Long": 7468,
+ "▁sobre": 7469,
+ "▁Old": 7470,
+ "west": 7471,
+ "},\\": 7472,
+ "ingu": 7473,
+ "▁spring": 7474,
+ "▁nur": 7475,
+ "example": 7476,
+ "When": 7477,
+ "▁advice": 7478,
+ "▁ult": 7479,
+ "ennis": 7480,
+ "▁Love": 7481,
+ "▁\"\"": 7482,
+ "▁increased": 7483,
+ "▁finding": 7484,
+ "irty": 7485,
+ "istrict": 7486,
+ "▁layer": 7487,
+ "template": 7488,
+ "First": 7489,
+ "ным": 7490,
+ "igration": 7491,
+ "rency": 7492,
+ "owie": 7493,
+ "▁np": 7494,
+ "▁selection": 7495,
+ "▁Nach": 7496,
+ "▁PRO": 7497,
+ "▁polic": 7498,
+ "▁database": 7499,
+ "▁byte": 7500,
+ "▁providing": 7501,
+ "mac": 7502,
+ "▁metal": 7503,
+ "modules": 7504,
+ "▁Georg": 7505,
+ "▁Sa": 7506,
+ "▁establish": 7507,
+ "...\"": 7508,
+ "iu": 7509,
+ "kin": 7510,
+ "▁eth": 7511,
+ "▁Sand": 7512,
+ "▁Chapter": 7513,
+ "▁gal": 7514,
+ "▁ice": 7515,
+ "Red": 7516,
+ "▁dal": 7517,
+ "▁principal": 7518,
+ "Msg": 7519,
+ "▁remains": 7520,
+ "нг": 7521,
+ "Title": 7522,
+ "Rel": 7523,
+ "Display": 7524,
+ "Non": 7525,
+ "▁definition": 7526,
+ "▁attr": 7527,
+ "▁signal": 7528,
+ "hl": 7529,
+ "▁sel": 7530,
+ "▁volume": 7531,
+ "▁cache": 7532,
+ "hens": 7533,
+ "▁wird": 7534,
+ "[\\": 7535,
+ "NOT": 7536,
+ "▁election": 7537,
+ "utt": 7538,
+ "▁Window": 7539,
+ "ental": 7540,
+ "ifest": 7541,
+ "xf": 7542,
+ "▁Ра": 7543,
+ "▁overall": 7544,
+ "blic": 7545,
+ "▁editor": 7546,
+ "aden": 7547,
+ "▁cart": 7548,
+ "Left": 7549,
+ "uls": 7550,
+ "bing": 7551,
+ "Right": 7552,
+ "▁sé": 7553,
+ "Sim": 7554,
+ "▁camera": 7555,
+ "▁fav": 7556,
+ "Decl": 7557,
+ "spring": 7558,
+ "▁errors": 7559,
+ "Tab": 7560,
+ "println": 7561,
+ "▁Bern": 7562,
+ "nab": 7563,
+ "▁Base": 7564,
+ "▁auth": 7565,
+ "▁apparent": 7566,
+ "▁presented": 7567,
+ "▁remained": 7568,
+ "▁wet": 7569,
+ "Enc": 7570,
+ "INFO": 7571,
+ "▁Sing": 7572,
+ "package": 7573,
+ ")));": 7574,
+ "▁Social": 7575,
+ "▁Mass": 7576,
+ "▁despite": 7577,
+ "▁mobile": 7578,
+ "▁labor": 7579,
+ "Go": 7580,
+ "▁esp": 7581,
+ "▁Table": 7582,
+ "▁expert": 7583,
+ "▁flex": 7584,
+ "▁profession": 7585,
+ "▁pil": 7586,
+ "Collection": 7587,
+ "LOCK": 7588,
+ "▁applied": 7589,
+ "aller": 7590,
+ "orph": 7591,
+ "ENSE": 7592,
+ "▁был": 7593,
+ "▁db": 7594,
+ "overline": 7595,
+ "▁Code": 7596,
+ "▁bytes": 7597,
+ "▁trouble": 7598,
+ "▁насе": 7599,
+ "DD": 7600,
+ "▁Year": 7601,
+ "mbox": 7602,
+ "▁keeping": 7603,
+ "▁kick": 7604,
+ "äng": 7605,
+ "▁corresponding": 7606,
+ "▁library": 7607,
+ "▁*/\r": 7608,
+ "callback": 7609,
+ "ums": 7610,
+ "▁json": 7611,
+ "▁Mount": 7612,
+ "▁Stand": 7613,
+ "IGHT": 7614,
+ "▁News": 7615,
+ "▁comments": 7616,
+ "returns": 7617,
+ "Cal": 7618,
+ "▁award": 7619,
+ "▁bought": 7620,
+ "includegraphics": 7621,
+ "▁ле": 7622,
+ "dot": 7623,
+ "ronic": 7624,
+ "▁extremely": 7625,
+ "▁minor": 7626,
+ "ifer": 7627,
+ "java": 7628,
+ "endar": 7629,
+ "layout": 7630,
+ "plies": 7631,
+ "▁buf": 7632,
+ "▁Island": 7633,
+ "▁About": 7634,
+ "▁west": 7635,
+ "▁Scott": 7636,
+ "ACT": 7637,
+ "Why": 7638,
+ "▁largest": 7639,
+ "▁container": 7640,
+ "▁temperature": 7641,
+ "▁£": 7642,
+ "▁reduce": 7643,
+ "▁foi": 7644,
+ "han": 7645,
+ "▁bod": 7646,
+ "▁Van": 7647,
+ "▁nullptr": 7648,
+ "▁dating": 7649,
+ "▁chain": 7650,
+ "Flags": 7651,
+ "iento": 7652,
+ "sort": 7653,
+ "▁fan": 7654,
+ "▁determine": 7655,
+ "▁wear": 7656,
+ "BE": 7657,
+ "▁appropriate": 7658,
+ "лся": 7659,
+ "тов": 7660,
+ "▁goals": 7661,
+ "▁Map": 7662,
+ "▁Sar": 7663,
+ "▁Option": 7664,
+ "▁hate": 7665,
+ "▁zijn": 7666,
+ ",-": 7667,
+ "▁implied": 7668,
+ "bits": 7669,
+ "▁Men": 7670,
+ "skip": 7671,
+ "▁Mond": 7672,
+ "▁Hon": 7673,
+ "▁prove": 7674,
+ "van": 7675,
+ "▁traff": 7676,
+ "▁intr": 7677,
+ "pic": 7678,
+ "▁dropped": 7679,
+ "▁werd": 7680,
+ "▁separate": 7681,
+ "isa": 7682,
+ "▁tab": 7683,
+ "tml": 7684,
+ "▁\"$": 7685,
+ "mutex": 7686,
+ "▁Pan": 7687,
+ "serve": 7688,
+ "▁hotel": 7689,
+ "▁Last": 7690,
+ "step": 7691,
+ "▁vir": 7692,
+ "Rule": 7693,
+ "istan": 7694,
+ "oting": 7695,
+ "arks": 7696,
+ "(__": 7697,
+ "▁els": 7698,
+ "Player": 7699,
+ "]]": 7700,
+ "вич": 7701,
+ "ych": 7702,
+ "exception": 7703,
+ "=\"../": 7704,
+ "▁imagine": 7705,
+ "\"},": 7706,
+ "icago": 7707,
+ "eler": 7708,
+ "▁vs": 7709,
+ "▁Africa": 7710,
+ "▁Business": 7711,
+ "ocks": 7712,
+ "▁prz": 7713,
+ "▁fucking": 7714,
+ "▁picked": 7715,
+ "▁ві": 7716,
+ "▁\",": 7717,
+ "▁bott": 7718,
+ "▁failure": 7719,
+ "[:": 7720,
+ "▁Gar": 7721,
+ "apes": 7722,
+ "uple": 7723,
+ "▁fer": 7724,
+ "▁purchase": 7725,
+ "▁пер": 7726,
+ "▁bird": 7727,
+ "Widget": 7728,
+ "▁Sunday": 7729,
+ "▁Amaz": 7730,
+ "▁consult": 7731,
+ "utsch": 7732,
+ "anto": 7733,
+ "Storage": 7734,
+ "▁header": 7735,
+ "ühr": 7736,
+ "▁Ha": 7737,
+ "▁Association": 7738,
+ "▁sight": 7739,
+ "Cell": 7740,
+ "▁profile": 7741,
+ "▁female": 7742,
+ "ån": 7743,
+ "▁wid": 7744,
+ "zn": 7745,
+ "Direct": 7746,
+ "▁stret": 7747,
+ "aat": 7748,
+ "▁patient": 7749,
+ "here": 7750,
+ "▁Atl": 7751,
+ "inet": 7752,
+ "Definition": 7753,
+ "imary": 7754,
+ "Policy": 7755,
+ "▁dut": 7756,
+ "▁majority": 7757,
+ "сі": 7758,
+ "▁Project": 7759,
+ "ById": 7760,
+ "▁believed": 7761,
+ "▁Music": 7762,
+ "зы": 7763,
+ "anti": 7764,
+ "▁oder": 7765,
+ "Channel": 7766,
+ "▁sle": 7767,
+ "▁sequence": 7768,
+ "▁pieces": 7769,
+ "▁kne": 7770,
+ "▁absolutely": 7771,
+ "▁Philip": 7772,
+ "abilities": 7773,
+ "Que": 7774,
+ "▁Kar": 7775,
+ "Execut": 7776,
+ "▁Devel": 7777,
+ "▁electric": 7778,
+ "full": 7779,
+ "rolled": 7780,
+ "Dom": 7781,
+ "▁river": 7782,
+ "▁healthy": 7783,
+ "▁extern": 7784,
+ "fit": 7785,
+ "▁coach": 7786,
+ "▁Kr": 7787,
+ "asta": 7788,
+ "Compat": 7789,
+ "▁exit": 7790,
+ "▁Const": 7791,
+ "after": 7792,
+ "▁shoulder": 7793,
+ "▁jobs": 7794,
+ "zone": 7795,
+ "▁sale": 7796,
+ "ixel": 7797,
+ "▁determined": 7798,
+ "▁anyway": 7799,
+ "orf": 7800,
+ "▁Ger": 7801,
+ "allel": 7802,
+ "rees": 7803,
+ "asm": 7804,
+ "ims": 7805,
+ "▁records": 7806,
+ "▁corpor": 7807,
+ "▁intellig": 7808,
+ "▁Prem": 7809,
+ "▁driving": 7810,
+ "▁marriage": 7811,
+ "▁Thank": 7812,
+ "▁willing": 7813,
+ "MC": 7814,
+ "Fields": 7815,
+ "Items": 7816,
+ "▁micro": 7817,
+ "▁lift": 7818,
+ "irection": 7819,
+ "Account": 7820,
+ "▁architect": 7821,
+ "track": 7822,
+ "▁prin": 7823,
+ "PA": 7824,
+ "▁runs": 7825,
+ "▁Texas": 7826,
+ "isher": 7827,
+ "ensure": 7828,
+ "▁Both": 7829,
+ "ком": 7830,
+ "▁Color": 7831,
+ "Register": 7832,
+ "▁Joe": 7833,
+ "geq": 7834,
+ "lets": 7835,
+ "ading": 7836,
+ "▁army": 7837,
+ "▁Bank": 7838,
+ "otic": 7839,
+ "Product": 7840,
+ "import": 7841,
+ "▁Wed": 7842,
+ "▁cry": 7843,
+ "grade": 7844,
+ "dig": 7845,
+ "gal": 7846,
+ "кла": 7847,
+ "ested": 7848,
+ "ões": 7849,
+ "gers": 7850,
+ "ologie": 7851,
+ "том": 7852,
+ "razy": 7853,
+ "▁dinner": 7854,
+ "QU": 7855,
+ "▁fingers": 7856,
+ "ULE": 7857,
+ "claim": 7858,
+ "▁advantage": 7859,
+ "▁variable": 7860,
+ "▁medic": 7861,
+ "▁male": 7862,
+ "▁circum": 7863,
+ "▁мі": 7864,
+ "▁internet": 7865,
+ "WN": 7866,
+ "▁lab": 7867,
+ "azine": 7868,
+ "чно": 7869,
+ "▁loop": 7870,
+ "▁pred": 7871,
+ "▁consequ": 7872,
+ "▁balance": 7873,
+ "fortun": 7874,
+ "▁gift": 7875,
+ "▁drug": 7876,
+ "▁cash": 7877,
+ "ских": 7878,
+ "rg": 7879,
+ "istribut": 7880,
+ "▁highest": 7881,
+ "ême": 7882,
+ "emph": 7883,
+ "emon": 7884,
+ "▁performed": 7885,
+ "cut": 7886,
+ "▁closer": 7887,
+ "▁becoming": 7888,
+ "▁\"\",": 7889,
+ "star": 7890,
+ "pub": 7891,
+ "▁prepar": 7892,
+ "▁vote": 7893,
+ "ilde": 7894,
+ "▁impress": 7895,
+ "▁employees": 7896,
+ "▁einen": 7897,
+ "▁smooth": 7898,
+ "▁snow": 7899,
+ "▁purs": 7900,
+ "▁voc": 7901,
+ "▁Microsoft": 7902,
+ "PU": 7903,
+ "▁income": 7904,
+ "inos": 7905,
+ "▁operator": 7906,
+ "▁equival": 7907,
+ "▁password": 7908,
+ "ción": 7909,
+ "success": 7910,
+ "▁emp": 7911,
+ "HOUT": 7912,
+ "▁ca": 7913,
+ "flag": 7914,
+ "illy": 7915,
+ "crete": 7916,
+ "frak": 7917,
+ "▁hidden": 7918,
+ "▁\"%": 7919,
+ "ERN": 7920,
+ "рова": 7921,
+ "▁UN": 7922,
+ "roke": 7923,
+ "miss": 7924,
+ "▁split": 7925,
+ "Reference": 7926,
+ ")$,": 7927,
+ "eper": 7928,
+ "▁NO": 7929,
+ "▁square": 7930,
+ "sur": 7931,
+ "чен": 7932,
+ "ester": 7933,
+ "нь": 7934,
+ "}\"": 7935,
+ "rawn": 7936,
+ "rule": 7937,
+ "▁audience": 7938,
+ "este": 7939,
+ "ems": 7940,
+ "ICENSE": 7941,
+ "▁Ill": 7942,
+ "USE": 7943,
+ "▁bon": 7944,
+ "bur": 7945,
+ "▁sick": 7946,
+ "▁horse": 7947,
+ "▁Educ": 7948,
+ "▁benefit": 7949,
+ "▁cro": 7950,
+ "Application": 7951,
+ "▁corre": 7952,
+ "▁guarante": 7953,
+ "DATA": 7954,
+ "▁explained": 7955,
+ "TX": 7956,
+ "▁ont": 7957,
+ "▁Flor": 7958,
+ "▁reports": 7959,
+ "▁Real": 7960,
+ "uded": 7961,
+ "lean": 7962,
+ "▁citiz": 7963,
+ "▁decide": 7964,
+ "WS": 7965,
+ "▁domain": 7966,
+ "▁reflect": 7967,
+ "▁minimum": 7968,
+ "▁legs": 7969,
+ "▁smiled": 7970,
+ "fi": 7971,
+ "▁pure": 7972,
+ "▁Custom": 7973,
+ "▁essential": 7974,
+ "▁observed": 7975,
+ "Bytes": 7976,
+ "▁ctx": 7977,
+ "▁rates": 7978,
+ "mbre": 7979,
+ "▁worry": 7980,
+ ")^": 7981,
+ "▁Research": 7982,
+ "Root": 7983,
+ "Windows": 7984,
+ "ulture": 7985,
+ "▁relative": 7986,
+ "▁seu": 7987,
+ "▁nie": 7988,
+ "▁shook": 7989,
+ "iously": 7990,
+ "▁advert": 7991,
+ "See": 7992,
+ "▁Central": 7993,
+ "▁batter": 7994,
+ "▁signed": 7995,
+ "TS": 7996,
+ "oni": 7997,
+ "▁prepared": 7998,
+ "gate": 7999,
+ "▁Care": 8000,
+ "care": 8001,
+ "▁supply": 8002,
+ "Exp": 8003,
+ "bolds": 8004,
+ "▁trail": 8005,
+ "▁fish": 8006,
+ "▁units": 8007,
+ "venue": 8008,
+ "хи": 8009,
+ "▁Wood": 8010,
+ "▁category": 8011,
+ "▁ble": 8012,
+ "▁override": 8013,
+ "foo": 8014,
+ "▁influence": 8015,
+ "enth": 8016,
+ "rij": 8017,
+ "▁adapt": 8018,
+ "icians": 8019,
+ "deleted": 8020,
+ "▁vision": 8021,
+ "ctrl": 8022,
+ "Lambda": 8023,
+ "tp": 8024,
+ "mond": 8025,
+ "aturday": 8026,
+ "normal": 8027,
+ "▁thousand": 8028,
+ "▁Profess": 8029,
+ "▁disease": 8030,
+ "clip": 8031,
+ "▁гра": 8032,
+ "boldsymbol": 8033,
+ "OB": 8034,
+ "▁challenge": 8035,
+ "▁motion": 8036,
+ "▁whis": 8037,
+ "▁leaders": 8038,
+ "▁colon": 8039,
+ "▁suit": 8040,
+ "mid": 8041,
+ "ampion": 8042,
+ "ág": 8043,
+ "▁views": 8044,
+ "▁appears": 8045,
+ "ancel": 8046,
+ "▁zwe": 8047,
+ "IST": 8048,
+ "▁leaves": 8049,
+ "▁enh": 8050,
+ "Active": 8051,
+ "▁dit": 8052,
+ "ificate": 8053,
+ "matrix": 8054,
+ "Expression": 8055,
+ "Reader": 8056,
+ "▁mental": 8057,
+ "embre": 8058,
+ "▁decor": 8059,
+ "arts": 8060,
+ "▁vent": 8061,
+ "nel": 8062,
+ "lines": 8063,
+ "upid": 8064,
+ "erved": 8065,
+ "▁boys": 8066,
+ "аль": 8067,
+ "MOD": 8068,
+ "isl": 8069,
+ "▁[[": 8070,
+ "phy": 8071,
+ "▁..": 8072,
+ "▁agent": 8073,
+ "▁Services": 8074,
+ "▁iron": 8075,
+ "▁components": 8076,
+ "▁fre": 8077,
+ "ictionary": 8078,
+ "▁tests": 8079,
+ ".~\\": 8080,
+ "obs": 8081,
+ "▁Ми": 8082,
+ "▁обла": 8083,
+ "▁assess": 8084,
+ "▁Friday": 8085,
+ "▁weather": 8086,
+ "kg": 8087,
+ "стра": 8088,
+ ".}": 8089,
+ "endant": 8090,
+ "anna": 8091,
+ "▁Japanese": 8092,
+ "cmp": 8093,
+ "▁Army": 8094,
+ "onym": 8095,
+ "▁relax": 8096,
+ "dates": 8097,
+ "▁Russian": 8098,
+ "▁excellent": 8099,
+ "'))": 8100,
+ "ILITY": 8101,
+ "▁showing": 8102,
+ "▁Daniel": 8103,
+ "мя": 8104,
+ "▁Main": 8105,
+ "Phi": 8106,
+ "▁Rock": 8107,
+ "▁grew": 8108,
+ "▁yield": 8109,
+ "ière": 8110,
+ "seg": 8111,
+ "}}$": 8112,
+ "▁strict": 8113,
+ "▁vehicle": 8114,
+ "UD": 8115,
+ "AF": 8116,
+ "Sw": 8117,
+ "▁chest": 8118,
+ "▁officer": 8119,
+ "▁ear": 8120,
+ "HER": 8121,
+ "noon": 8122,
+ "▁journey": 8123,
+ "NT": 8124,
+ "▁divers": 8125,
+ "▁Finally": 8126,
+ "Found": 8127,
+ "▁AS": 8128,
+ "rik": 8129,
+ "▁constr": 8130,
+ "▁sust": 8131,
+ "account": 8132,
+ "▁walls": 8133,
+ "▁entirely": 8134,
+ "Iter": 8135,
+ "cha": 8136,
+ "ishes": 8137,
+ "IVE": 8138,
+ "▁prime": 8139,
+ "▁…": 8140,
+ "xe": 8141,
+ "uten": 8142,
+ "arse": 8143,
+ "▁Pa": 8144,
+ "pute": 8145,
+ "äl": 8146,
+ "▁protection": 8147,
+ "▁keys": 8148,
+ "May": 8149,
+ "Byte": 8150,
+ "Const": 8151,
+ "BL": 8152,
+ "▁пе": 8153,
+ "▁spl": 8154,
+ "▁clothes": 8155,
+ "ashed": 8156,
+ "Mark": 8157,
+ "ème": 8158,
+ "▁fait": 8159,
+ "▁introduced": 8160,
+ "unlock": 8161,
+ "▁Instead": 8162,
+ "ansion": 8163,
+ "region": 8164,
+ "▁Americans": 8165,
+ "▁indeed": 8166,
+ "widget": 8167,
+ "▁realize": 8168,
+ "▁fro": 8169,
+ "BIT": 8170,
+ "▁React": 8171,
+ "READ": 8172,
+ "asket": 8173,
+ "never": 8174,
+ "▁poll": 8175,
+ "icol": 8176,
+ "▁prev": 8177,
+ "▁hyp": 8178,
+ "▁Fur": 8179,
+ "cloud": 8180,
+ "▁Lee": 8181,
+ "pling": 8182,
+ "▁Child": 8183,
+ "▁ideal": 8184,
+ "Selector": 8185,
+ "STATUS": 8186,
+ "ucture": 8187,
+ "▁wine": 8188,
+ "▁possibly": 8189,
+ "▁putting": 8190,
+ "▁riv": 8191,
+ "▁wearing": 8192,
+ "▁Source": 8193,
+ "▁Cas": 8194,
+ "Changed": 8195,
+ "▁thanks": 8196,
+ "TIME": 8197,
+ "▁sport": 8198,
+ "▁Award": 8199,
+ "▁glad": 8200,
+ "▁Pass": 8201,
+ "▁Pos": 8202,
+ "sche": 8203,
+ "▁CD": 8204,
+ "▁afford": 8205,
+ "▁Women": 8206,
+ "▁District": 8207,
+ "▁identity": 8208,
+ "▁parties": 8209,
+ ":%": 8210,
+ "▁drag": 8211,
+ "▁mai": 8212,
+ "!(": 8213,
+ "langle": 8214,
+ "▁knowing": 8215,
+ "Project": 8216,
+ "▁regarding": 8217,
+ "▁Joseph": 8218,
+ "ге": 8219,
+ "▁Dar": 8220,
+ "▁Hor": 8221,
+ "▁animals": 8222,
+ "▁extension": 8223,
+ "ская": 8224,
+ "▁Han": 8225,
+ "btn": 8226,
+ "aciones": 8227,
+ "▁familiar": 8228,
+ "holder": 8229,
+ ":\r": 8230,
+ "stood": 8231,
+ "▁liked": 8232,
+ "CODE": 8233,
+ "▁enable": 8234,
+ "▁ped": 8235,
+ "iti": 8236,
+ "hab": 8237,
+ "DIR": 8238,
+ "▁beat": 8239,
+ "ті": 8240,
+ "▁Minister": 8241,
+ "▁py": 8242,
+ "Pat": 8243,
+ "▁exhib": 8244,
+ "▁Build": 8245,
+ "▁Field": 8246,
+ "ician": 8247,
+ "▁collabor": 8248,
+ "▁quarter": 8249,
+ "▁False": 8250,
+ "km": 8251,
+ "▁virtual": 8252,
+ "owa": 8253,
+ "▁Jon": 8254,
+ "amin": 8255,
+ "uen": 8256,
+ "▁ин": 8257,
+ "imation": 8258,
+ "oving": 8259,
+ "▁testing": 8260,
+ "sect": 8261,
+ "ITION": 8262,
+ "!\\": 8263,
+ "apy": 8264,
+ "▁transition": 8265,
+ "ository": 8266,
+ "ODO": 8267,
+ "PD": 8268,
+ "né": 8269,
+ "▁generate": 8270,
+ "▁native": 8271,
+ "▁('": 8272,
+ "▁elle": 8273,
+ "RR": 8274,
+ "▁hun": 8275,
+ "_->": 8276,
+ "agnost": 8277,
+ "▁proposed": 8278,
+ "▁Game": 8279,
+ "▁efforts": 8280,
+ "вя": 8281,
+ "tc": 8282,
+ "ск": 8283,
+ "▁intent": 8284,
+ "▁Bre": 8285,
+ "isc": 8286,
+ "▁protest": 8287,
+ "▁holds": 8288,
+ "ometry": 8289,
+ "▁Have": 8290,
+ "▁detail": 8291,
+ "▁WITHOUT": 8292,
+ "yer": 8293,
+ "▁Kon": 8294,
+ "▁noticed": 8295,
+ "▁requirements": 8296,
+ "DEBUG": 8297,
+ "kins": 8298,
+ "▁Span": 8299,
+ "▁cars": 8300,
+ "meta": 8301,
+ "▁kil": 8302,
+ "▁Bron": 8303,
+ "▁experienced": 8304,
+ "▁remind": 8305,
+ "ourse": 8306,
+ "▁Western": 8307,
+ "tered": 8308,
+ "▁devices": 8309,
+ "▁pictures": 8310,
+ "▁tut": 8311,
+ "\"`": 8312,
+ "▁impossible": 8313,
+ "▁rail": 8314,
+ "▁feels": 8315,
+ "icas": 8316,
+ "illing": 8317,
+ "▁accident": 8318,
+ "▁'@": 8319,
+ "________": 8320,
+ "▁notes": 8321,
+ "oman": 8322,
+ "Parser": 8323,
+ "▁discovered": 8324,
+ "▁Roman": 8325,
+ "▁budget": 8326,
+ "▁guide": 8327,
+ "king": 8328,
+ "▁incred": 8329,
+ "olar": 8330,
+ "enden": 8331,
+ "Desc": 8332,
+ "▁wave": 8333,
+ "бли": 8334,
+ "igt": 8335,
+ "▁restrict": 8336,
+ "▁Ret": 8337,
+ "▁mac": 8338,
+ "ур": 8339,
+ "BS": 8340,
+ "ís": 8341,
+ "▁generation": 8342,
+ "dem": 8343,
+ "alo": 8344,
+ "бра": 8345,
+ "▁ordered": 8346,
+ "drop": 8347,
+ "▁pp": 8348,
+ "▁Review": 8349,
+ "▁literally": 8350,
+ "▁Sir": 8351,
+ "▁Yeah": 8352,
+ "▁density": 8353,
+ "riz": 8354,
+ "inde": 8355,
+ "▁gain": 8356,
+ "▁panel": 8357,
+ "jet": 8358,
+ "▁Times": 8359,
+ "▁nella": 8360,
+ "▁previously": 8361,
+ "points": 8362,
+ "Send": 8363,
+ "▁Brown": 8364,
+ "each": 8365,
+ "▁trigger": 8366,
+ "ometimes": 8367,
+ "icos": 8368,
+ "GR": 8369,
+ "Panel": 8370,
+ "ogen": 8371,
+ "▁cm": 8372,
+ "ructions": 8373,
+ "▁kiss": 8374,
+ "▁solo": 8375,
+ "▁famous": 8376,
+ "ran": 8377,
+ "про": 8378,
+ "▁thro": 8379,
+ "Graph": 8380,
+ "imit": 8381,
+ "▁Value": 8382,
+ "▁starts": 8383,
+ "ipeline": 8384,
+ "hd": 8385,
+ "TC": 8386,
+ "▁discussion": 8387,
+ "▁truck": 8388,
+ "aka": 8389,
+ "Only": 8390,
+ "▁Equ": 8391,
+ "▁kö": 8392,
+ "▁Bes": 8393,
+ "▁critic": 8394,
+ "▁propos": 8395,
+ "▁batt": 8396,
+ "▁Section": 8397,
+ "Show": 8398,
+ "gp": 8399,
+ "STATE": 8400,
+ "POST": 8401,
+ "▁Nord": 8402,
+ "▁innov": 8403,
+ "▁crim": 8404,
+ "axis": 8405,
+ "▁Turn": 8406,
+ "conn": 8407,
+ "Runtime": 8408,
+ "▁remaining": 8409,
+ "oston": 8410,
+ "▁Э": 8411,
+ "▁windows": 8412,
+ "▁Royal": 8413,
+ "▁vide": 8414,
+ "PP": 8415,
+ "chron": 8416,
+ "▁san": 8417,
+ "▁rise": 8418,
+ "▁delle": 8419,
+ "▁Dur": 8420,
+ "▁rapid": 8421,
+ "cert": 8422,
+ "LA": 8423,
+ "edge": 8424,
+ "▁\\]": 8425,
+ "▁entered": 8426,
+ "▁laws": 8427,
+ "▁photo": 8428,
+ "▁applications": 8429,
+ "▁Berlin": 8430,
+ "▁arrest": 8431,
+ "▁federal": 8432,
+ "▁Russia": 8433,
+ "▁usual": 8434,
+ "▁raw": 8435,
+ "▁più": 8436,
+ "être": 8437,
+ "JSON": 8438,
+ "SION": 8439,
+ "xture": 8440,
+ "istent": 8441,
+ "▁Power": 8442,
+ "Bit": 8443,
+ "▁capacity": 8444,
+ "▁cards": 8445,
+ "UID": 8446,
+ "iments": 8447,
+ "▁dar": 8448,
+ "▁Chicago": 8449,
+ "▁comfortable": 8450,
+ "tip": 8451,
+ "bas": 8452,
+ "▁mu": 8453,
+ "▁enemy": 8454,
+ "yan": 8455,
+ "▁фи": 8456,
+ "▁updated": 8457,
+ "ango": 8458,
+ "Ev": 8459,
+ "Effect": 8460,
+ "osing": 8461,
+ "rence": 8462,
+ "▁Congress": 8463,
+ "▁defe": 8464,
+ "▁ip": 8465,
+ "▁tout": 8466,
+ "▁freedom": 8467,
+ "▁ao": 8468,
+ "▁Therefore": 8469,
+ "Edit": 8470,
+ "▁Virgin": 8471,
+ "REE": 8472,
+ "argo": 8473,
+ "▁Dam": 8474,
+ "▁traffic": 8475,
+ "ños": 8476,
+ "▁alle": 8477,
+ "▁depth": 8478,
+ "Now": 8479,
+ "▁sides": 8480,
+ "▁годи": 8481,
+ "Descriptor": 8482,
+ "▁artikel": 8483,
+ "▁narrow": 8484,
+ "___": 8485,
+ "kw": 8486,
+ "uto": 8487,
+ "▁Facebook": 8488,
+ "tegr": 8489,
+ "boolean": 8490,
+ "nik": 8491,
+ "bd": 8492,
+ "Track": 8493,
+ "▁gran": 8494,
+ "reshold": 8495,
+ "вет": 8496,
+ "wrap": 8497,
+ "▁noise": 8498,
+ "igu": 8499,
+ "▁Bon": 8500,
+ "▁wy": 8501,
+ "linux": 8502,
+ "cks": 8503,
+ "▁fans": 8504,
+ "▁mach": 8505,
+ "▁prices": 8506,
+ "év": 8507,
+ "outs": 8508,
+ "standing": 8509,
+ "▁categ": 8510,
+ ";\\": 8511,
+ "▁decre": 8512,
+ "▁Saturday": 8513,
+ "▁menu": 8514,
+ "▁Nov": 8515,
+ "▁Yet": 8516,
+ "▁так": 8517,
+ "liche": 8518,
+ "▁Academ": 8519,
+ "▁communication": 8520,
+ "using": 8521,
+ "▁Society": 8522,
+ "▁nuc": 8523,
+ "pective": 8524,
+ "orial": 8525,
+ "▁afraid": 8526,
+ "▁animal": 8527,
+ "▁turning": 8528,
+ "dst": 8529,
+ "mathfrak": 8530,
+ "lers": 8531,
+ "▁lots": 8532,
+ "▁á": 8533,
+ "▁Tra": 8534,
+ "np": 8535,
+ "▁rose": 8536,
+ "▁GL": 8537,
+ "▁helping": 8538,
+ "▁winter": 8539,
+ "▁ком": 8540,
+ "Mock": 8541,
+ "▁investment": 8542,
+ "Use": 8543,
+ "▁Canad": 8544,
+ "нд": 8545,
+ "Copy": 8546,
+ "▁fly": 8547,
+ "SER": 8548,
+ "▁Far": 8549,
+ "▁Ros": 8550,
+ "amil": 8551,
+ "▁fighting": 8552,
+ "▁religious": 8553,
+ "super": 8554,
+ "screen": 8555,
+ "▁furn": 8556,
+ "▁surprised": 8557,
+ "▁replied": 8558,
+ "Activity": 8559,
+ "▁Down": 8560,
+ "▁insert": 8561,
+ "▁Olymp": 8562,
+ "▁pointed": 8563,
+ "▁Card": 8564,
+ "driver": 8565,
+ "▁Da": 8566,
+ "!--": 8567,
+ "roud": 8568,
+ "undo": 8569,
+ "▁messages": 8570,
+ "▁Point": 8571,
+ "VM": 8572,
+ "▁plane": 8573,
+ "xc": 8574,
+ "▁television": 8575,
+ "ён": 8576,
+ "▁thousands": 8577,
+ "▁cris": 8578,
+ "▁delay": 8579,
+ "▁Next": 8580,
+ "▁nombre": 8581,
+ "▁tu": 8582,
+ "▁skip": 8583,
+ "road": 8584,
+ "istration": 8585,
+ "▁tur": 8586,
+ "▁Develop": 8587,
+ "▁Па": 8588,
+ "▁дру": 8589,
+ "▁wonderful": 8590,
+ ">&": 8591,
+ "▁Liber": 8592,
+ "▁scope": 8593,
+ "▁manage": 8594,
+ "▁dass": 8595,
+ "▁recall": 8596,
+ "PM": 8597,
+ "▁relevant": 8598,
+ "▁Earth": 8599,
+ "▁как": 8600,
+ "▁apr": 8601,
+ "▁ASS": 8602,
+ "ién": 8603,
+ "▁SH": 8604,
+ "oom": 8605,
+ "itet": 8606,
+ "none": 8607,
+ "asi": 8608,
+ "▁motor": 8609,
+ "▁Show": 8610,
+ "nb": 8611,
+ "▁factors": 8612,
+ "▁forest": 8613,
+ "▁вре": 8614,
+ "thm": 8615,
+ "▁municip": 8616,
+ "▁turns": 8617,
+ "▁Division": 8618,
+ "EC": 8619,
+ "▁disappe": 8620,
+ "structor": 8621,
+ "▁somewhere": 8622,
+ "▁African": 8623,
+ "▁Institute": 8624,
+ "Grid": 8625,
+ "▁teacher": 8626,
+ "uries": 8627,
+ "▁respectively": 8628,
+ "▁SD": 8629,
+ "▁alive": 8630,
+ "▁pou": 8631,
+ "▁Water": 8632,
+ "фе": 8633,
+ "▁changing": 8634,
+ "▁afternoon": 8635,
+ "▁orders": 8636,
+ "Ret": 8637,
+ "Pointer": 8638,
+ "▁sav": 8639,
+ "erg": 8640,
+ "oked": 8641,
+ "essions": 8642,
+ "▁Fire": 8643,
+ "aret": 8644,
+ "imm": 8645,
+ "▁desire": 8646,
+ "▁що": 8647,
+ "▁Design": 8648,
+ "uture": 8649,
+ "▁Office": 8650,
+ "▁cmd": 8651,
+ "▁eating": 8652,
+ "Network": 8653,
+ "▁rough": 8654,
+ "operator": 8655,
+ "IGN": 8656,
+ "▁sports": 8657,
+ "▁weren": 8658,
+ "▁noted": 8659,
+ "▁twice": 8660,
+ "III": 8661,
+ "▁anx": 8662,
+ "▁elim": 8663,
+ "▁ав": 8664,
+ "▁io": 8665,
+ "▁speech": 8666,
+ "▁condu": 8667,
+ "elles": 8668,
+ "idade": 8669,
+ "▁advance": 8670,
+ "RI": 8671,
+ "oca": 8672,
+ "/\\": 8673,
+ "apshot": 8674,
+ "▁tail": 8675,
+ "models": 8676,
+ "ogy": 8677,
+ "▁Jeff": 8678,
+ "iration": 8679,
+ "▁Kore": 8680,
+ "▁leads": 8681,
+ "bat": 8682,
+ "Adapter": 8683,
+ "category": 8684,
+ "angular": 8685,
+ "▁saved": 8686,
+ "▁uniform": 8687,
+ "▁né": 8688,
+ "▁businesses": 8689,
+ "Hist": 8690,
+ "▁ар": 8691,
+ "domain": 8692,
+ "▁Si": 8693,
+ "raise": 8694,
+ "▁warn": 8695,
+ "hetic": 8696,
+ "▁Gro": 8697,
+ ")).": 8698,
+ "}>": 8699,
+ "зе": 8700,
+ "▁Amazon": 8701,
+ "▁Organ": 8702,
+ "▁Lake": 8703,
+ "▁agreement": 8704,
+ "xa": 8705,
+ "▁perman": 8706,
+ "▁containing": 8707,
+ "▁strange": 8708,
+ "сті": 8709,
+ "▁stupid": 8710,
+ "▁speaking": 8711,
+ "▁Internet": 8712,
+ "prefix": 8713,
+ "esc": 8714,
+ "Assert": 8715,
+ "prote": 8716,
+ "▁manner": 8717,
+ "▁Sz": 8718,
+ "unte": 8719,
+ "iot": 8720,
+ "Profile": 8721,
+ "oven": 8722,
+ "▁formed": 8723,
+ "▁lit": 8724,
+ "▁economy": 8725,
+ "▁cz": 8726,
+ "wid": 8727,
+ "REQ": 8728,
+ "▁chosen": 8729,
+ "▁Produ": 8730,
+ "oster": 8731,
+ "stances": 8732,
+ "awa": 8733,
+ "▁Ren": 8734,
+ "▁confirm": 8735,
+ "▁Бо": 8736,
+ "▁billion": 8737,
+ "▁déc": 8738,
+ "ých": 8739,
+ "▁illustr": 8740,
+ "TIES": 8741,
+ "▁Pub": 8742,
+ "▁ban": 8743,
+ "aded": 8744,
+ "ahn": 8745,
+ "▁Cath": 8746,
+ "nonumber": 8747,
+ "▁worst": 8748,
+ "▁Ме": 8749,
+ "▁suggested": 8750,
+ "stats": 8751,
+ "▁cant": 8752,
+ "▁align": 8753,
+ "kappa": 8754,
+ "▁hen": 8755,
+ "▁initi": 8756,
+ "'])": 8757,
+ "BI": 8758,
+ "▁garden": 8759,
+ "▁secure": 8760,
+ "▁\\[": 8761,
+ "handler": 8762,
+ "elli": 8763,
+ "ldots": 8764,
+ "secut": 8765,
+ "▁extended": 8766,
+ "}-": 8767,
+ "anie": 8768,
+ "▁Find": 8769,
+ "▁Museum": 8770,
+ "▁Conne": 8771,
+ "yy": 8772,
+ "▁passion": 8773,
+ "akers": 8774,
+ "ahr": 8775,
+ "ologies": 8776,
+ "▁equation": 8777,
+ "▁occasion": 8778,
+ "Let": 8779,
+ "']['": 8780,
+ "Print": 8781,
+ "anes": 8782,
+ "iente": 8783,
+ "▁Today": 8784,
+ "LECT": 8785,
+ "▁Af": 8786,
+ ",,": 8787,
+ "▁Та": 8788,
+ "▁```": 8789,
+ "even": 8790,
+ "sin": 8791,
+ "urer": 8792,
+ "▁°": 8793,
+ "otimes": 8794,
+ "▁IO": 8795,
+ "▁poet": 8796,
+ "()));": 8797,
+ "▁−": 8798,
+ "▁adopt": 8799,
+ "phere": 8800,
+ "#[": 8801,
+ "▁centre": 8802,
+ "oves": 8803,
+ "▁ans": 8804,
+ "dp": 8805,
+ "▁Kir": 8806,
+ "▁applicable": 8807,
+ "fp": 8808,
+ "▁visual": 8809,
+ "▁okay": 8810,
+ "oro": 8811,
+ "▁opportunities": 8812,
+ "Repository": 8813,
+ "▁ll": 8814,
+ "▁Rod": 8815,
+ "▁shel": 8816,
+ "▁launch": 8817,
+ "▁conven": 8818,
+ "▁Spe": 8819,
+ "Amer": 8820,
+ "▁cette": 8821,
+ "Cond": 8822,
+ "dep": 8823,
+ "Own": 8824,
+ "▁hook": 8825,
+ "▁dict": 8826,
+ "▁Those": 8827,
+ "▁fellow": 8828,
+ "▁philosoph": 8829,
+ "vin": 8830,
+ "ferences": 8831,
+ "hav": 8832,
+ "▁adding": 8833,
+ "iverse": 8834,
+ "game": 8835,
+ "▁Blue": 8836,
+ "▁clin": 8837,
+ "note": 8838,
+ "▁Ram": 8839,
+ "мер": 8840,
+ "covery": 8841,
+ "ña": 8842,
+ "▁би": 8843,
+ "▁fashion": 8844,
+ "▁broke": 8845,
+ "▁'\\": 8846,
+ "▁reader": 8847,
+ "ное": 8848,
+ "ности": 8849,
+ "▁payment": 8850,
+ "▁Lic": 8851,
+ "▁lips": 8852,
+ "▁academ": 8853,
+ "▁Mot": 8854,
+ "ells": 8855,
+ "CHECK": 8856,
+ "▁ру": 8857,
+ "▁MS": 8858,
+ "Editor": 8859,
+ "▁zone": 8860,
+ "iture": 8861,
+ "▁IT": 8862,
+ "runtime": 8863,
+ "▁proceed": 8864,
+ "лов": 8865,
+ "▁Maria": 8866,
+ "olver": 8867,
+ "▁Thanks": 8868,
+ "▁shouldn": 8869,
+ "▁Joh": 8870,
+ "▁Model": 8871,
+ "▁Sov": 8872,
+ "!'": 8873,
+ "Di": 8874,
+ "▁cancer": 8875,
+ "Ident": 8876,
+ "▁exchange": 8877,
+ "iller": 8878,
+ "inf": 8879,
+ "LEN": 8880,
+ "(){": 8881,
+ "aga": 8882,
+ "\"],": 8883,
+ "uh": 8884,
+ "▁Ken": 8885,
+ "▁photos": 8886,
+ "▁tiny": 8887,
+ "▁gent": 8888,
+ "ül": 8889,
+ "▁Take": 8890,
+ "idel": 8891,
+ "outing": 8892,
+ "Internal": 8893,
+ "▁cells": 8894,
+ "ним": 8895,
+ "hard": 8896,
+ "▁Town": 8897,
+ "obe": 8898,
+ "plex": 8899,
+ "тер": 8900,
+ "tons": 8901,
+ "▁concentr": 8902,
+ "mock": 8903,
+ "vc": 8904,
+ "áz": 8905,
+ "▁Championship": 8906,
+ "▁бе": 8907,
+ "??": 8908,
+ "éri": 8909,
+ "aly": 8910,
+ "▁Ц": 8911,
+ "ierte": 8912,
+ "▁totally": 8913,
+ "▁Auf": 8914,
+ "▁ourselves": 8915,
+ "▁Self": 8916,
+ "Forms": 8917,
+ "ighter": 8918,
+ "▁island": 8919,
+ "fmt": 8920,
+ "▁rc": 8921,
+ "▁tells": 8922,
+ "BB": 8923,
+ "dit": 8924,
+ "▁variables": 8925,
+ "▁intended": 8926,
+ "izont": 8927,
+ "▁plays": 8928,
+ "dam": 8929,
+ "seq": 8930,
+ "▁Sup": 8931,
+ "▁cultural": 8932,
+ "▁scream": 8933,
+ "__,": 8934,
+ "cipl": 8935,
+ "Timeout": 8936,
+ "▁ж": 8937,
+ "orte": 8938,
+ "▁replaced": 8939,
+ "EM": 8940,
+ "▁abandon": 8941,
+ "▁Special": 8942,
+ "ellen": 8943,
+ "▁Bru": 8944,
+ "irmed": 8945,
+ "Te": 8946,
+ "olt": 8947,
+ "ju": 8948,
+ "Argument": 8949,
+ "▁neut": 8950,
+ "scape": 8951,
+ "▁Ray": 8952,
+ "▁Polit": 8953,
+ "▁crowd": 8954,
+ "▁Windows": 8955,
+ "iego": 8956,
+ "▁escape": 8957,
+ "▁Apache": 8958,
+ "sync": 8959,
+ "eben": 8960,
+ "ifies": 8961,
+ "ether": 8962,
+ "Meta": 8963,
+ "▁biggest": 8964,
+ "Game": 8965,
+ "▁transaction": 8966,
+ "Env": 8967,
+ "▁Мо": 8968,
+ "▁plenty": 8969,
+ "▁mel": 8970,
+ "пре": 8971,
+ "▁motiv": 8972,
+ "▁ор": 8973,
+ "organ": 8974,
+ "▁mock": 8975,
+ "▁$_": 8976,
+ "ене": 8977,
+ "▁Number": 8978,
+ "cknow": 8979,
+ "▁Update": 8980,
+ "zero": 8981,
+ "▁surprise": 8982,
+ "cean": 8983,
+ "pdf": 8984,
+ "Global": 8985,
+ "▁attend": 8986,
+ "▁fond": 8987,
+ "▁understood": 8988,
+ "Nav": 8989,
+ "▁Mic": 8990,
+ "=$": 8991,
+ "oking": 8992,
+ "▁Stadium": 8993,
+ "Close": 8994,
+ "▁competition": 8995,
+ "▁soldiers": 8996,
+ "▁OP": 8997,
+ "agne": 8998,
+ "▁Anton": 8999,
+ "Main": 9000,
+ "ák": 9001,
+ "▁#[": 9002,
+ "▁Commit": 9003,
+ "pyx": 9004,
+ "▁east": 9005,
+ "▁Order": 9006,
+ "Float": 9007,
+ "▁accepted": 9008,
+ "▁monitor": 9009,
+ "▁pad": 9010,
+ "onic": 9011,
+ "▁pushed": 9012,
+ "▁replace": 9013,
+ "CRE": 9014,
+ "▁ride": 9015,
+ "found": 9016,
+ "=%": 9017,
+ "вой": 9018,
+ "▁matches": 9019,
+ "▁Lie": 9020,
+ "▁experiences": 9021,
+ "Pool": 9022,
+ "ups": 9023,
+ "AV": 9024,
+ "▁existence": 9025,
+ "▁thin": 9026,
+ "▁magn": 9027,
+ "COMP": 9028,
+ "home": 9029,
+ "▁ni": 9030,
+ "▁wurden": 9031,
+ "лав": 9032,
+ "▁teeth": 9033,
+ "▁Stan": 9034,
+ "appro": 9035,
+ "anny": 9036,
+ "ifts": 9037,
+ "▁unknown": 9038,
+ "▁homes": 9039,
+ "▁entity": 9040,
+ "cie": 9041,
+ "ление": 9042,
+ "iar": 9043,
+ "▁compliance": 9044,
+ "▁focused": 9045,
+ "uzz": 9046,
+ "=\\\"": 9047,
+ "components": 9048,
+ "Attr": 9049,
+ "allery": 9050,
+ "▁identify": 9051,
+ "Ok": 9052,
+ "pie": 9053,
+ "▁Still": 9054,
+ "▁offering": 9055,
+ "▁busy": 9056,
+ "ctl": 9057,
+ "itors": 9058,
+ "▁concerned": 9059,
+ "▁brown": 9060,
+ "clk": 9061,
+ "Selected": 9062,
+ "▁Block": 9063,
+ "▁egy": 9064,
+ "icing": 9065,
+ "▁URL": 9066,
+ "▁topic": 9067,
+ "▁Product": 9068,
+ "▁чи": 9069,
+ "▁trial": 9070,
+ "▁weekend": 9071,
+ "lu": 9072,
+ "▁IV": 9073,
+ "▁Egy": 9074,
+ "xC": 9075,
+ "▁nove": 9076,
+ "▁lett": 9077,
+ "enne": 9078,
+ "()).": 9079,
+ ".**": 9080,
+ "▁promise": 9081,
+ "election": 9082,
+ "Auth": 9083,
+ "rv": 9084,
+ "ril": 9085,
+ "▁conduct": 9086,
+ "▁maintain": 9087,
+ "▁boat": 9088,
+ "▁opposite": 9089,
+ "spin": 9090,
+ "webpack": 9091,
+ "anta": 9092,
+ "▁orient": 9093,
+ "▁suc": 9094,
+ "▁exercise": 9095,
+ "▁efficient": 9096,
+ "▁tradition": 9097,
+ "▁zw": 9098,
+ "▁Sud": 9099,
+ "going": 9100,
+ "▁Pier": 9101,
+ "inv": 9102,
+ "ipes": 9103,
+ "ensuremath": 9104,
+ "▁conver": 9105,
+ "creen": 9106,
+ "▁terror": 9107,
+ "▁Dou": 9108,
+ "▁invalid": 9109,
+ "ceived": 9110,
+ "▁Arab": 9111,
+ "▁wire": 9112,
+ "application": 9113,
+ "shift": 9114,
+ "Generic": 9115,
+ "▁Plan": 9116,
+ "▁Wall": 9117,
+ "▁directory": 9118,
+ "▁egg": 9119,
+ "▁wealth": 9120,
+ "random": 9121,
+ "attribute": 9122,
+ "▁hide": 9123,
+ "Serial": 9124,
+ "cam": 9125,
+ "▁ital": 9126,
+ "▁Line": 9127,
+ "▁CHECK": 9128,
+ "ployment": 9129,
+ "▁massive": 9130,
+ "▁extract": 9131,
+ "chain": 9132,
+ "Rest": 9133,
+ "▁Las": 9134,
+ "▁bear": 9135,
+ "▁links": 9136,
+ "▁newsp": 9137,
+ "▁FC": 9138,
+ "Card": 9139,
+ "aks": 9140,
+ "▁visible": 9141,
+ "▁Marc": 9142,
+ "▁Boston": 9143,
+ "▁reserved": 9144,
+ "▁roof": 9145,
+ "licenses": 9146,
+ "dc": 9147,
+ "▁Information": 9148,
+ "▁witness": 9149,
+ "Sk": 9150,
+ "*),": 9151,
+ "Scope": 9152,
+ "'];": 9153,
+ "▁Mir": 9154,
+ "uding": 9155,
+ "▁trend": 9156,
+ "rep": 9157,
+ "▁musical": 9158,
+ "▁neither": 9159,
+ "▁Creat": 9160,
+ "▁positions": 9161,
+ "LC": 9162,
+ "ridge": 9163,
+ "▁officers": 9164,
+ "▁violence": 9165,
+ "▁Tem": 9166,
+ "▁Sus": 9167,
+ "▁Way": 9168,
+ "After": 9169,
+ "acket": 9170,
+ "▁Sou": 9171,
+ "acer": 9172,
+ "||": 9173,
+ "▁remark": 9174,
+ "water": 9175,
+ "ně": 9176,
+ "▁Са": 9177,
+ "▁sed": 9178,
+ "Each": 9179,
+ "▁photograph": 9180,
+ "▁letters": 9181,
+ "▁invent": 9182,
+ "▁Mas": 9183,
+ "▁songs": 9184,
+ "ól": 9185,
+ "kind": 9186,
+ "▁Non": 9187,
+ "▁dust": 9188,
+ "**:": 9189,
+ "nabla": 9190,
+ ".\",": 9191,
+ "Lock": 9192,
+ "▁До": 9193,
+ "▁cluster": 9194,
+ "loss": 9195,
+ "▁ASSERT": 9196,
+ "fall": 9197,
+ "▁reject": 9198,
+ "▁Spring": 9199,
+ "▁wedding": 9200,
+ "▁grav": 9201,
+ "ression": 9202,
+ "limit": 9203,
+ "RES": 9204,
+ "]}": 9205,
+ "▁listed": 9206,
+ "▁Tele": 9207,
+ "hline": 9208,
+ "▁chief": 9209,
+ "MEM": 9210,
+ "дар": 9211,
+ "▁expensive": 9212,
+ "trace": 9213,
+ "▁Rog": 9214,
+ "▁Coll": 9215,
+ "▁Author": 9216,
+ "▁Board": 9217,
+ "▁Capt": 9218,
+ "TEXT": 9219,
+ "▁recon": 9220,
+ "esta": 9221,
+ "▁properly": 9222,
+ "▁&\\": 9223,
+ "leton": 9224,
+ "iker": 9225,
+ "Gu": 9226,
+ "▁Kom": 9227,
+ "oco": 9228,
+ "▁anymore": 9229,
+ "▁taste": 9230,
+ "▁Santa": 9231,
+ "gex": 9232,
+ "▁Secret": 9233,
+ "▁talent": 9234,
+ "▁moments": 9235,
+ "▁Ba": 9236,
+ "▁extr": 9237,
+ "▁Commission": 9238,
+ "▁modify": 9239,
+ "▁Figure": 9240,
+ "▁domin": 9241,
+ "▁plot": 9242,
+ "enger": 9243,
+ "utch": 9244,
+ "▁cities": 9245,
+ "▁nut": 9246,
+ "profile": 9247,
+ "▁Stat": 9248,
+ "▁nodes": 9249,
+ "▁ns": 9250,
+ "essages": 9251,
+ "impl": 9252,
+ "icker": 9253,
+ "▁examples": 9254,
+ "abeth": 9255,
+ "▁stated": 9256,
+ "fire": 9257,
+ "bul": 9258,
+ "▁dangerous": 9259,
+ "▁Pay": 9260,
+ "▁Gre": 9261,
+ "▁Monday": 9262,
+ "esome": 9263,
+ "igan": 9264,
+ "rund": 9265,
+ "prise": 9266,
+ "fail": 9267,
+ "▁Never": 9268,
+ "Av": 9269,
+ "▁linear": 9270,
+ "▁ul": 9271,
+ "WAR": 9272,
+ "рен": 9273,
+ "▁AT": 9274,
+ "▁dop": 9275,
+ "▁nou": 9276,
+ "Dest": 9277,
+ "▁claims": 9278,
+ "enda": 9279,
+ "▁crazy": 9280,
+ "gel": 9281,
+ "oggle": 9282,
+ "▁representation": 9283,
+ "inen": 9284,
+ "▁alternative": 9285,
+ "DM": 9286,
+ "ABILITY": 9287,
+ "faces": 9288,
+ "▁doors": 9289,
+ "ativ": 9290,
+ "Look": 9291,
+ "▁JSON": 9292,
+ "▁appearance": 9293,
+ "бря": 9294,
+ "SQL": 9295,
+ "▁silence": 9296,
+ "udo": 9297,
+ "▁Director": 9298,
+ "Statement": 9299,
+ "selected": 9300,
+ "high": 9301,
+ "prime": 9302,
+ "▁ignore": 9303,
+ "▁colors": 9304,
+ "ushing": 9305,
+ "▁virt": 9306,
+ "manager": 9307,
+ "▁remote": 9308,
+ "ło": 9309,
+ "small": 9310,
+ "▁crime": 9311,
+ "rb": 9312,
+ "▁creation": 9313,
+ "▁flight": 9314,
+ "▁Sign": 9315,
+ "ILE": 9316,
+ "▁DO": 9317,
+ "comment": 9318,
+ "▁Cost": 9319,
+ ".__": 9320,
+ "▁Cop": 9321,
+ "▁vom": 9322,
+ "▁Science": 9323,
+ "ления": 9324,
+ "oop": 9325,
+ "interface": 9326,
+ "▁WARRANTIES": 9327,
+ "▁Page": 9328,
+ "******": 9329,
+ "ском": 9330,
+ "TRUE": 9331,
+ "▁repeated": 9332,
+ "▁его": 9333,
+ "шо": 9334,
+ "▁roz": 9335,
+ "Pe": 9336,
+ "▁ISBN": 9337,
+ "irts": 9338,
+ "poses": 9339,
+ "})$": 9340,
+ "▁І": 9341,
+ "children": 9342,
+ "bles": 9343,
+ "ECT": 9344,
+ "▁iz": 9345,
+ "▁builder": 9346,
+ "▁Media": 9347,
+ "iat": 9348,
+ "▁contrast": 9349,
+ "”,": 9350,
+ "▁Link": 9351,
+ "▁Education": 9352,
+ "▁joint": 9353,
+ "▁external": 9354,
+ "▁роз": 9355,
+ "▁bits": 9356,
+ "FORM": 9357,
+ "erman": 9358,
+ "wp": 9359,
+ "▁Mike": 9360,
+ "▁Master": 9361,
+ "▁senior": 9362,
+ "▁Nav": 9363,
+ "▁recorded": 9364,
+ "eling": 9365,
+ "esh": 9366,
+ "fx": 9367,
+ "кан": 9368,
+ "▁tall": 9369,
+ "▁Johnson": 9370,
+ "▁sono": 9371,
+ "▁anche": 9372,
+ "icken": 9373,
+ "loop": 9374,
+ "iciency": 9375,
+ "emporary": 9376,
+ "▁Does": 9377,
+ "▁relation": 9378,
+ "мы": 9379,
+ "was": 9380,
+ "low": 9381,
+ "ichte": 9382,
+ "▁Jones": 9383,
+ "▁bedroom": 9384,
+ "DIS": 9385,
+ "▁magnet": 9386,
+ "▁Engine": 9387,
+ "▁feelings": 9388,
+ "GC": 9389,
+ "▁torn": 9390,
+ "▁relationships": 9391,
+ "▁Ре": 9392,
+ "▁proud": 9393,
+ "▁twe": 9394,
+ "oval": 9395,
+ "▁waste": 9396,
+ "▁reduced": 9397,
+ "ilton": 9398,
+ "BP": 9399,
+ "▁forgot": 9400,
+ "▁bodies": 9401,
+ "▁Haw": 9402,
+ "lag": 9403,
+ "▁www": 9404,
+ "door": 9405,
+ "▁sufficient": 9406,
+ "▁dollars": 9407,
+ "Len": 9408,
+ "▁talked": 9409,
+ "▁bond": 9410,
+ "▁Bor": 9411,
+ "}}{": 9412,
+ "rod": 9413,
+ "Password": 9414,
+ "quare": 9415,
+ "▁lights": 9416,
+ "eren": 9417,
+ "▁thirty": 9418,
+ "NC": 9419,
+ "▁TODO": 9420,
+ "▁respond": 9421,
+ "ких": 9422,
+ "direct": 9423,
+ "ação": 9424,
+ "▁heav": 9425,
+ "Media": 9426,
+ "exit": 9427,
+ "License": 9428,
+ "`.": 9429,
+ "▁mixed": 9430,
+ "▁desk": 9431,
+ "▁teaching": 9432,
+ "▁maj": 9433,
+ "▁nerv": 9434,
+ "inations": 9435,
+ "typeof": 9436,
+ "▁coast": 9437,
+ "▁же": 9438,
+ "▁beside": 9439,
+ "ummy": 9440,
+ "Doc": 9441,
+ "▁schedule": 9442,
+ "▁recover": 9443,
+ "▁Further": 9444,
+ "▁steel": 9445,
+ "boot": 9446,
+ "▁Perhaps": 9447,
+ "▁съ": 9448,
+ "▁Os": 9449,
+ "rick": 9450,
+ "▁Ви": 9451,
+ "Support": 9452,
+ "▁(_": 9453,
+ "nil": 9454,
+ "pis": 9455,
+ "xpected": 9456,
+ "▁processing": 9457,
+ "Build": 9458,
+ "arian": 9459,
+ "▁icon": 9460,
+ "▁CA": 9461,
+ "wick": 9462,
+ "=(": 9463,
+ "▁algorithm": 9464,
+ "▁Young": 9465,
+ "▁Management": 9466,
+ "▁ancient": 9467,
+ "ность": 9468,
+ "oti": 9469,
+ "▁combination": 9470,
+ "world": 9471,
+ "nn": 9472,
+ "▁dram": 9473,
+ "enabled": 9474,
+ "Ac": 9475,
+ "CCESS": 9476,
+ "aration": 9477,
+ "▁blocks": 9478,
+ "▁Angeles": 9479,
+ "▁Qual": 9480,
+ "▁succeed": 9481,
+ "network": 9482,
+ "▁oblig": 9483,
+ "springframework": 9484,
+ "▁Tre": 9485,
+ "okes": 9486,
+ "mun": 9487,
+ "▁Network": 9488,
+ "Del": 9489,
+ "▁estate": 9490,
+ "▁liqu": 9491,
+ "▁pob": 9492,
+ "▁dad": 9493,
+ "▁distinct": 9494,
+ "▁Tit": 9495,
+ "▁Lear": 9496,
+ "ferred": 9497,
+ "android": 9498,
+ "▁subsequ": 9499,
+ "▁Florida": 9500,
+ "subset": 9501,
+ "▁whisper": 9502,
+ "Vol": 9503,
+ "ulous": 9504,
+ "▁crew": 9505,
+ "▁lug": 9506,
+ "pid": 9507,
+ "ocity": 9508,
+ "skb": 9509,
+ "▁tea": 9510,
+ "ун": 9511,
+ "▁honor": 9512,
+ "▁Ins": 9513,
+ "▁gew": 9514,
+ "Details": 9515,
+ "eneath": 9516,
+ "atar": 9517,
+ "▁_{": 9518,
+ "amen": 9519,
+ "▁setup": 9520,
+ "Transaction": 9521,
+ "▁blank": 9522,
+ "Failed": 9523,
+ "job": 9524,
+ "▁pret": 9525,
+ "ße": 9526,
+ "loor": 9527,
+ "ří": 9528,
+ "ncia": 9529,
+ "▁anywhere": 9530,
+ "▁Light": 9531,
+ "▁Ak": 9532,
+ "BD": 9533,
+ "▁excited": 9534,
+ "agers": 9535,
+ "▁warning": 9536,
+ "▁processes": 9537,
+ "hu": 9538,
+ "▁youth": 9539,
+ "▁dogs": 9540,
+ "▁oct": 9541,
+ "▁nine": 9542,
+ "Writer": 9543,
+ "grid": 9544,
+ "▁importance": 9545,
+ "estic": 9546,
+ "▁carefully": 9547,
+ "master": 9548,
+ "▁decisions": 9549,
+ "▁pin": 9550,
+ "▁crack": 9551,
+ "TEST": 9552,
+ "▁Local": 9553,
+ "▁Right": 9554,
+ "▁vast": 9555,
+ "▁faster": 9556,
+ "▁institut": 9557,
+ "▁annual": 9558,
+ "LAN": 9559,
+ "▁episode": 9560,
+ "▁XV": 9561,
+ "▁delivery": 9562,
+ "tl": 9563,
+ "FP": 9564,
+ "circ": 9565,
+ "▁typically": 9566,
+ "igo": 9567,
+ "▁intel": 9568,
+ "nat": 9569,
+ "xb": 9570,
+ "стро": 9571,
+ ")-": 9572,
+ "▁Bal": 9573,
+ "▁Jos": 9574,
+ "▁gonna": 9575,
+ "▁Rest": 9576,
+ "jor": 9577,
+ "onia": 9578,
+ "orship": 9579,
+ "overy": 9580,
+ "LINE": 9581,
+ "]:": 9582,
+ "Queue": 9583,
+ "▁compare": 9584,
+ "▁apartment": 9585,
+ "▁rul": 9586,
+ "Dr": 9587,
+ "gency": 9588,
+ "▁obviously": 9589,
+ "zie": 9590,
+ "ycl": 9591,
+ "fortunately": 9592,
+ "▁stepped": 9593,
+ "▁Seg": 9594,
+ "▁Which": 9595,
+ "▁PC": 9596,
+ "▁ast": 9597,
+ "endor": 9598,
+ "▁permission": 9599,
+ "COL": 9600,
+ "▁TEST": 9601,
+ "Pay": 9602,
+ "ères": 9603,
+ "▁studied": 9604,
+ "▁accompl": 9605,
+ "role": 9606,
+ "Where": 9607,
+ "protobuf": 9608,
+ "metadata": 9609,
+ "Job": 9610,
+ "▁Four": 9611,
+ "plements": 9612,
+ "disable": 9613,
+ "▁loud": 9614,
+ "▁happening": 9615,
+ "▁Using": 9616,
+ "rog": 9617,
+ "▁depends": 9618,
+ "ím": 9619,
+ "'\\": 9620,
+ "▁taught": 9621,
+ "shared": 9622,
+ "▁attributes": 9623,
+ "▁Action": 9624,
+ "▁dess": 9625,
+ "▁houses": 9626,
+ "▁reset": 9627,
+ "▁bien": 9628,
+ "▁explicit": 9629,
+ "LOW": 9630,
+ "->_": 9631,
+ "▁PM": 9632,
+ "Category": 9633,
+ "oice": 9634,
+ "into": 9635,
+ "▁mail": 9636,
+ "▁authority": 9637,
+ "▁unable": 9638,
+ "filename": 9639,
+ "ék": 9640,
+ "лей": 9641,
+ "▁sector": 9642,
+ "appoint": 9643,
+ "▁hang": 9644,
+ "▁cel": 9645,
+ "related": 9646,
+ "itate": 9647,
+ "▁'<": 9648,
+ "amber": 9649,
+ "▁cheap": 9650,
+ "▁enabled": 9651,
+ "▁division": 9652,
+ "Any": 9653,
+ "▁hier": 9654,
+ "▁Head": 9655,
+ "ntax": 9656,
+ "uda": 9657,
+ "▁limitations": 9658,
+ "▁studio": 9659,
+ "media": 9660,
+ "▁circle": 9661,
+ "нова": 9662,
+ "▁laug": 9663,
+ "acts": 9664,
+ "▁Во": 9665,
+ "ód": 9666,
+ "pled": 9667,
+ "LOC": 9668,
+ "Expr": 9669,
+ ">:": 9670,
+ "▁prés": 9671,
+ "▁laughed": 9672,
+ "▁Three": 9673,
+ "лы": 9674,
+ "▁ends": 9675,
+ "▁fundament": 9676,
+ "▁inher": 9677,
+ "▁liv": 9678,
+ "bid": 9679,
+ "▁responsibility": 9680,
+ "▁checked": 9681,
+ "▁Pac": 9682,
+ "▁fault": 9683,
+ "▁yellow": 9684,
+ "▁salt": 9685,
+ "▁Francisco": 9686,
+ "▁^": 9687,
+ "▁ON": 9688,
+ "▁beauty": 9689,
+ "yg": 9690,
+ "▁Aff": 9691,
+ "▁Eq": 9692,
+ "▁magic": 9693,
+ "▁handler": 9694,
+ "xE": 9695,
+ "▁numerous": 9696,
+ "▁hole": 9697,
+ "▁rooms": 9698,
+ "cción": 9699,
+ "▁Arm": 9700,
+ "person": 9701,
+ "▁buildings": 9702,
+ "▁plate": 9703,
+ "bled": 9704,
+ "errors": 9705,
+ "▁Again": 9706,
+ "▁Default": 9707,
+ "▁Hard": 9708,
+ "tó": 9709,
+ "hus": 9710,
+ "▁dimension": 9711,
+ "iale": 9712,
+ "▁Mult": 9713,
+ "▁Government": 9714,
+ "Func": 9715,
+ "▁blow": 9716,
+ "▁rect": 9717,
+ "erra": 9718,
+ "connection": 9719,
+ "▁passing": 9720,
+ "ßen": 9721,
+ "phas": 9722,
+ "ensional": 9723,
+ "record": 9724,
+ "cohol": 9725,
+ "▁Harry": 9726,
+ "izontal": 9727,
+ "▁finger": 9728,
+ "▁younger": 9729,
+ "▁SC": 9730,
+ "operation": 9731,
+ "BY": 9732,
+ "heim": 9733,
+ "▁Bad": 9734,
+ "▁storm": 9735,
+ "▁Nat": 9736,
+ "▁buying": 9737,
+ "▁Sometimes": 9738,
+ "▁Ста": 9739,
+ "essed": 9740,
+ "▁damn": 9741,
+ "▁meg": 9742,
+ "umes": 9743,
+ "ünd": 9744,
+ "тра": 9745,
+ "▁silver": 9746,
+ "wd": 9747,
+ "hidden": 9748,
+ "ardo": 9749,
+ "▁communities": 9750,
+ "▁diet": 9751,
+ "otted": 9752,
+ "▁bat": 9753,
+ "ancer": 9754,
+ "▁fmt": 9755,
+ "▁Pen": 9756,
+ "▁til": 9757,
+ "Enum": 9758,
+ "PATH": 9759,
+ "▁matters": 9760,
+ "timeout": 9761,
+ "------------": 9762,
+ "kan": 9763,
+ "▁Corpor": 9764,
+ "=\"../../": 9765,
+ "▁Ale": 9766,
+ "hentication": 9767,
+ "▁complic": 9768,
+ "▁Security": 9769,
+ "OFF": 9770,
+ "Rad": 9771,
+ "apse": 9772,
+ "▁dance": 9773,
+ "▁permissions": 9774,
+ "▁warrant": 9775,
+ "▁lad": 9776,
+ "▁isol": 9777,
+ "dl": 9778,
+ "▁Au": 9779,
+ "yes": 9780,
+ "▁tv": 9781,
+ "▁provider": 9782,
+ "▁terrible": 9783,
+ "▁department": 9784,
+ "eral": 9785,
+ "▁implementation": 9786,
+ "SR": 9787,
+ "▁hearing": 9788,
+ "▁Kn": 9789,
+ "FR": 9790,
+ "tv": 9791,
+ "▁diss": 9792,
+ "FUN": 9793,
+ "▁durante": 9794,
+ "osis": 9795,
+ "▁tasks": 9796,
+ "▁Blo": 9797,
+ "вод": 9798,
+ "▁branch": 9799,
+ "▁politics": 9800,
+ "▁Elle": 9801,
+ "▁leadership": 9802,
+ "expr": 9803,
+ "▁techniques": 9804,
+ "prec": 9805,
+ "Sigma": 9806,
+ "imately": 9807,
+ "tk": 9808,
+ "achment": 9809,
+ "▁Enter": 9810,
+ "▁creative": 9811,
+ "▁зна": 9812,
+ "appy": 9813,
+ "unched": 9814,
+ "▁'',": 9815,
+ "onder": 9816,
+ "{-": 9817,
+ "NUM": 9818,
+ "▁narr": 9819,
+ "Memory": 9820,
+ "▁winning": 9821,
+ "▁Follow": 9822,
+ "*/\r": 9823,
+ "vision": 9824,
+ "resents": 9825,
+ "zione": 9826,
+ "▁latter": 9827,
+ "▁requests": 9828,
+ "▁margin": 9829,
+ "▁{\"": 9830,
+ "video": 9831,
+ "cn": 9832,
+ "▁Image": 9833,
+ "Tim": 9834,
+ "CONFIG": 9835,
+ "▁allowing": 9836,
+ "▁combined": 9837,
+ "PUT": 9838,
+ "▁instanceof": 9839,
+ "igin": 9840,
+ "▁pero": 9841,
+ "▁''": 9842,
+ "▁confidence": 9843,
+ "▁equivalent": 9844,
+ "pad": 9845,
+ "effect": 9846,
+ "RX": 9847,
+ "▁lang": 9848,
+ "strong": 9849,
+ "▁bridge": 9850,
+ "aya": 9851,
+ "▁treated": 9852,
+ "▁forth": 9853,
+ "SW": 9854,
+ "▁accounts": 9855,
+ "▁PO": 9856,
+ "▁listening": 9857,
+ "Route": 9858,
+ "()))": 9859,
+ "cpy": 9860,
+ "▁reform": 9861,
+ "▁gate": 9862,
+ "▁Walk": 9863,
+ "▁somehow": 9864,
+ "tf": 9865,
+ "▁layout": 9866,
+ "umin": 9867,
+ "▁considering": 9868,
+ "▁premi": 9869,
+ "▁Mom": 9870,
+ "athan": 9871,
+ "Gen": 9872,
+ "▁planet": 9873,
+ "amples": 9874,
+ "▁MO": 9875,
+ "shop": 9876,
+ "▁premier": 9877,
+ "▁simpl": 9878,
+ "▁segu": 9879,
+ "LY": 9880,
+ "Sum": 9881,
+ "▁tables": 9882,
+ "ska": 9883,
+ "▁ž": 9884,
+ "pd": 9885,
+ "▁sous": 9886,
+ "▁conference": 9887,
+ "▁Dat": 9888,
+ "Scroll": 9889,
+ "▁standards": 9890,
+ "▁гру": 9891,
+ "esse": 9892,
+ "▁citizens": 9893,
+ "▁occurred": 9894,
+ "▁democr": 9895,
+ "▁elev": 9896,
+ "▁Sem": 9897,
+ "ensus": 9898,
+ "headers": 9899,
+ "▁Chris": 9900,
+ "imento": 9901,
+ "kom": 9902,
+ "Cor": 9903,
+ "MIN": 9904,
+ "usher": 9905,
+ "Database": 9906,
+ "▁formal": 9907,
+ "igne": 9908,
+ "▁organizations": 9909,
+ "▁Ire": 9910,
+ "Xml": 9911,
+ "из": 9912,
+ "▁pray": 9913,
+ "▁bomb": 9914,
+ "▁mand": 9915,
+ "erts": 9916,
+ "▁clock": 9917,
+ "▁buck": 9918,
+ "вали": 9919,
+ "ensch": 9920,
+ "▁volt": 9921,
+ "▁films": 9922,
+ "▁plants": 9923,
+ "inode": 9924,
+ "Boolean": 9925,
+ "▁restaurant": 9926,
+ "ían": 9927,
+ "▁debut": 9928,
+ "pages": 9929,
+ "▁wordt": 9930,
+ "▁Ба": 9931,
+ "▁greatest": 9932,
+ "(\"/": 9933,
+ "▁copyright": 9934,
+ "▁rit": 9935,
+ "sizeof": 9936,
+ "Trace": 9937,
+ "uent": 9938,
+ "тур": 9939,
+ "▁ko": 9940,
+ ":\\": 9941,
+ "▁bigger": 9942,
+ "▁perfectly": 9943,
+ "tenance": 9944,
+ "MASK": 9945,
+ "ré": 9946,
+ "▁ett": 9947,
+ "▁nose": 9948,
+ "▁craft": 9949,
+ "iteral": 9950,
+ "▁discussed": 9951,
+ "▁Jewish": 9952,
+ "Cap": 9953,
+ "▁Unless": 9954,
+ "▁Jackson": 9955,
+ "Attributes": 9956,
+ "▁lunch": 9957,
+ "öl": 9958,
+ "atr": 9959,
+ "▁paying": 9960,
+ "Parse": 9961,
+ "()\r": 9962,
+ "lad": 9963,
+ "▁rare": 9964,
+ "▁[];": 9965,
+ "stone": 9966,
+ "▁unc": 9967,
+ "▁defense": 9968,
+ "}+": 9969,
+ "▁Global": 9970,
+ "▁Soviet": 9971,
+ "▁Australian": 9972,
+ "▁gli": 9973,
+ "variant": 9974,
+ "▁Ron": 9975,
+ "▁loan": 9976,
+ "Step": 9977,
+ "member": 9978,
+ "Sch": 9979,
+ "▁Committee": 9980,
+ "▁spending": 9981,
+ "▁Tri": 9982,
+ "▁Journal": 9983,
+ "▁sugar": 9984,
+ "elly": 9985,
+ "HTML": 9986,
+ "▁advent": 9987,
+ "wing": 9988,
+ "▁Whether": 9989,
+ "oration": 9990,
+ "▁NE": 9991,
+ "iveness": 9992,
+ "▁hav": 9993,
+ "▁conscious": 9994,
+ "een": 9995,
+ "Symbol": 9996,
+ "▁ку": 9997,
+ "Logger": 9998,
+ "▁Little": 9999,
+ "widet": 10000,
+ "ocation": 10001,
+ "pin": 10002,
+ "▁symmet": 10003,
+ "▁AD": 10004,
+ "▁posts": 10005,
+ "shal": 10006,
+ "▁Conf": 10007,
+ "▁chose": 10008,
+ "mal": 10009,
+ "ulo": 10010,
+ "▁Method": 10011,
+ "▁missed": 10012,
+ "Remove": 10013,
+ "Auto": 10014,
+ "VALUE": 10015,
+ "thlet": 10016,
+ "▁Force": 10017,
+ "pf": 10018,
+ "▁Я": 10019,
+ "late": 10020,
+ "▁pul": 10021,
+ "Pop": 10022,
+ "▁advanced": 10023,
+ "aires": 10024,
+ "ressed": 10025,
+ "AME": 10026,
+ "bell": 10027,
+ "aching": 10028,
+ "ić": 10029,
+ "echo": 10030,
+ "HS": 10031,
+ "▁funny": 10032,
+ "рии": 10033,
+ "▁eer": 10034,
+ "▁veget": 10035,
+ "▁fourth": 10036,
+ "cf": 10037,
+ "transform": 10038,
+ "▁grown": 10039,
+ "▁McC": 10040,
+ "site": 10041,
+ "▁beneath": 10042,
+ "▁shell": 10043,
+ "xd": 10044,
+ "Play": 10045,
+ "short": 10046,
+ "Role": 10047,
+ "▁religion": 10048,
+ "inator": 10049,
+ "}": 10050,
+ "▁Eliz": 10051,
+ "Microsoft": 10052,
+ "▁vez": 10053,
+ "▁рабо": 10054,
+ "reich": 10055,
+ "vet": 10056,
+ "enum": 10057,
+ "▁welcome": 10058,
+ "nament": 10059,
+ "▁jan": 10060,
+ "▁cycle": 10061,
+ "▁acknow": 10062,
+ "▁wound": 10063,
+ "idi": 10064,
+ "▁possibility": 10065,
+ "annotation": 10066,
+ "▁technical": 10067,
+ "▁fold": 10068,
+ "eh": 10069,
+ "istence": 10070,
+ "▁reply": 10071,
+ "etes": 10072,
+ "▁decades": 10073,
+ "wan": 10074,
+ "▁кра": 10075,
+ "▁Lab": 10076,
+ "▁unf": 10077,
+ "▁imper": 10078,
+ "▁bug": 10079,
+ "▁Though": 10080,
+ "throws": 10081,
+ "Visible": 10082,
+ "prev": 10083,
+ "▁Ty": 10084,
+ "▁depending": 10085,
+ "▁policies": 10086,
+ "andy": 10087,
+ "▁Italian": 10088,
+ "uma": 10089,
+ "▁signs": 10090,
+ "▁Through": 10091,
+ "бы": 10092,
+ "bot": 10093,
+ "▁publish": 10094,
+ ")**": 10095,
+ "ATTR": 10096,
+ "iral": 10097,
+ "VT": 10098,
+ "▁recognized": 10099,
+ "▁Lind": 10100,
+ "ection": 10101,
+ "▁relatively": 10102,
+ "▁Ah": 10103,
+ "▁Dig": 10104,
+ "ць": 10105,
+ "icket": 10106,
+ "▁specifically": 10107,
+ "nost": 10108,
+ "▁grass": 10109,
+ "▁causes": 10110,
+ "тво": 10111,
+ "utter": 10112,
+ "▁Festival": 10113,
+ "greg": 10114,
+ "▁weapons": 10115,
+ "▁sir": 10116,
+ "▁Virginia": 10117,
+ "login": 10118,
+ "▁schedul": 10119,
+ "ського": 10120,
+ "▁losing": 10121,
+ "▁Europ": 10122,
+ "\"><": 10123,
+ "asp": 10124,
+ "ajo": 10125,
+ "exports": 10126,
+ "▁Node": 10127,
+ "▁jako": 10128,
+ "▁ya": 10129,
+ "▁successfully": 10130,
+ "▁friendly": 10131,
+ "buff": 10132,
+ "DEFAULT": 10133,
+ "▁pregn": 10134,
+ "Required": 10135,
+ "▁binary": 10136,
+ "isting": 10137,
+ "▁stared": 10138,
+ "▁circumstances": 10139,
+ "▁хо": 10140,
+ "rei": 10141,
+ "▁Го": 10142,
+ "Transform": 10143,
+ "cnt": 10144,
+ "▁Ext": 10145,
+ "report": 10146,
+ "VERSION": 10147,
+ "▁analy": 10148,
+ "▁Marg": 10149,
+ "▁alleg": 10150,
+ "builder": 10151,
+ "ToString": 10152,
+ "Layer": 10153,
+ "íst": 10154,
+ "Prop": 10155,
+ "▁Emp": 10156,
+ "}]": 10157,
+ "▁selling": 10158,
+ "▁queue": 10159,
+ "▁seriously": 10160,
+ "▁Lead": 10161,
+ "textit": 10162,
+ "testing": 10163,
+ "▁Пре": 10164,
+ "security": 10165,
+ "iał": 10166,
+ "ún": 10167,
+ "chip": 10168,
+ "▁candidate": 10169,
+ "▁minister": 10170,
+ "eria": 10171,
+ "▁Het": 10172,
+ "дин": 10173,
+ "▁Britain": 10174,
+ "▁barely": 10175,
+ "▁sty": 10176,
+ "▁Spanish": 10177,
+ "▁Ven": 10178,
+ "timer": 10179,
+ "ків": 10180,
+ "▁documents": 10181,
+ "('.": 10182,
+ "▁debug": 10183,
+ "▁contro": 10184,
+ "стоя": 10185,
+ "▁joy": 10186,
+ "Sn": 10187,
+ "Inv": 10188,
+ "▁protocol": 10189,
+ "▁faces": 10190,
+ "▁Despite": 10191,
+ "sed": 10192,
+ "Conf": 10193,
+ "ARG": 10194,
+ "▁evolution": 10195,
+ "▁tod": 10196,
+ "▁Promise": 10197,
+ "▁posted": 10198,
+ "Perm": 10199,
+ "bet": 10200,
+ "Ang": 10201,
+ "Just": 10202,
+ "▁rum": 10203,
+ "layer": 10204,
+ "▁behavi": 10205,
+ "ipping": 10206,
+ "▁dynam": 10207,
+ "▁scheme": 10208,
+ "▁proto": 10209,
+ ")/": 10210,
+ "Collections": 10211,
+ "riev": 10212,
+ "▁Click": 10213,
+ "▁uns": 10214,
+ "widetilde": 10215,
+ "▁remembered": 10216,
+ "гі": 10217,
+ "inates": 10218,
+ "▁incorpor": 10219,
+ "▁Description": 10220,
+ "▁prepare": 10221,
+ "▁Final": 10222,
+ "uation": 10223,
+ "▁Queen": 10224,
+ ">;": 10225,
+ "▁automatically": 10226,
+ "▁sharp": 10227,
+ "▁meat": 10228,
+ "ateur": 10229,
+ "astern": 10230,
+ "▁stuck": 10231,
+ "ASSERT": 10232,
+ "▁planned": 10233,
+ "dots": 10234,
+ "ookie": 10235,
+ "▁Histor": 10236,
+ "▁reviews": 10237,
+ "IMP": 10238,
+ "▁answered": 10239,
+ "Total": 10240,
+ "▁sau": 10241,
+ "▁Mexico": 10242,
+ "continue": 10243,
+ "▁Apple": 10244,
+ "likely": 10245,
+ "зва": 10246,
+ "users": 10247,
+ "▁identified": 10248,
+ "▁Lev": 10249,
+ "▁mol": 10250,
+ "▁Islam": 10251,
+ "▁committed": 10252,
+ "writ": 10253,
+ "бер": 10254,
+ "rift": 10255,
+ "▁interrupt": 10256,
+ "▁readonly": 10257,
+ "schema": 10258,
+ "Sm": 10259,
+ "Double": 10260,
+ "aza": 10261,
+ "▁Hal": 10262,
+ "Move": 10263,
+ "▁Series": 10264,
+ "inline": 10265,
+ "▁которы": 10266,
+ "soc": 10267,
+ "▁tent": 10268,
+ "▁amer": 10269,
+ "aki": 10270,
+ "▁lady": 10271,
+ "▁tired": 10272,
+ "ifi": 10273,
+ "▁même": 10274,
+ "ouver": 10275,
+ "▁aside": 10276,
+ "Did": 10277,
+ "',\r": 10278,
+ "▁bringing": 10279,
+ "Drawing": 10280,
+ "aro": 10281,
+ "▁Rh": 10282,
+ "▁Naz": 10283,
+ "esso": 10284,
+ "▁reaction": 10285,
+ "mitted": 10286,
+ "▁absolute": 10287,
+ "haust": 10288,
+ "(()": 10289,
+ "▁Task": 10290,
+ "ERS": 10291,
+ "▁^{": 10292,
+ "VD": 10293,
+ "▁tone": 10294,
+ "dist": 10295,
+ "vs": 10296,
+ "▁wheel": 10297,
+ "▁administration": 10298,
+ "▁interests": 10299,
+ "▁pointer": 10300,
+ "▁encounter": 10301,
+ "aver": 10302,
+ "▁nord": 10303,
+ "ket": 10304,
+ "▁beach": 10305,
+ "▁enjoyed": 10306,
+ "contains": 10307,
+ "▁append": 10308,
+ "Wait": 10309,
+ "▁squad": 10310,
+ "zel": 10311,
+ "▁medium": 10312,
+ "▁sending": 10313,
+ "▁Lady": 10314,
+ "ções": 10315,
+ "▁destination": 10316,
+ "nych": 10317,
+ "▁conflict": 10318,
+ "▁Ly": 10319,
+ "▁vul": 10320,
+ "▁basically": 10321,
+ "reated": 10322,
+ "black": 10323,
+ "ugins": 10324,
+ "▁calm": 10325,
+ "érie": 10326,
+ "har": 10327,
+ "лан": 10328,
+ "▁Се": 10329,
+ "watch": 10330,
+ "▁Put": 10331,
+ "▁dump": 10332,
+ "acher": 10333,
+ "scroll": 10334,
+ "▁claimed": 10335,
+ "▁Control": 10336,
+ "▁blind": 10337,
+ "enti": 10338,
+ "▁Keep": 10339,
+ "▁Development": 10340,
+ "images": 10341,
+ "▁tough": 10342,
+ "gebra": 10343,
+ "▁sept": 10344,
+ "hew": 10345,
+ "▁skill": 10346,
+ "▁Tay": 10347,
+ "▁któ": 10348,
+ "owner": 10349,
+ "pare": 10350,
+ "▁fee": 10351,
+ "▁continues": 10352,
+ "▁kan": 10353,
+ "bes": 10354,
+ "▁cha": 10355,
+ "ovo": 10356,
+ "▁Night": 10357,
+ "icture": 10358,
+ "shire": 10359,
+ "▁essay": 10360,
+ "▁suppose": 10361,
+ "etic": 10362,
+ "Art": 10363,
+ "acon": 10364,
+ "lla": 10365,
+ "words": 10366,
+ "▁comparison": 10367,
+ "▁BE": 10368,
+ "▁challenges": 10369,
+ "▁ol": 10370,
+ "citep": 10371,
+ "▁Foot": 10372,
+ "▁Such": 10373,
+ "▁papers": 10374,
+ "activ": 10375,
+ "quer": 10376,
+ "тя": 10377,
+ "▁То": 10378,
+ "ський": 10379,
+ "thur": 10380,
+ "done": 10381,
+ "▁shock": 10382,
+ "▁dedicated": 10383,
+ "▁correspond": 10384,
+ "Second": 10385,
+ "▁bull": 10386,
+ "life": 10387,
+ "indent": 10388,
+ "▁figures": 10389,
+ "▁Andrew": 10390,
+ "isp": 10391,
+ "▁favour": 10392,
+ "зда": 10393,
+ "▁Elect": 10394,
+ "Full": 10395,
+ "▁nearby": 10396,
+ "▁Register": 10397,
+ "Scale": 10398,
+ "ications": 10399,
+ "ин": 10400,
+ "▁AM": 10401,
+ "pair": 10402,
+ "▁perspective": 10403,
+ "▁nos": 10404,
+ "apa": 10405,
+ "ostał": 10406,
+ "▁Pers": 10407,
+ "icer": 10408,
+ "▁plastic": 10409,
+ "дов": 10410,
+ "ciples": 10411,
+ "zą": 10412,
+ "clos": 10413,
+ "▁уча": 10414,
+ "▁Á": 10415,
+ "plugin": 10416,
+ "▁angle": 10417,
+ "▁commission": 10418,
+ "▁funds": 10419,
+ "▁indu": 10420,
+ "▁drawn": 10421,
+ "ám": 10422,
+ "▁developing": 10423,
+ "▁segment": 10424,
+ "isme": 10425,
+ "scr": 10426,
+ "▁lies": 10427,
+ "▁IL": 10428,
+ "▁api": 10429,
+ "Extension": 10430,
+ "▁scal": 10431,
+ "install": 10432,
+ "▁Week": 10433,
+ "▁gentle": 10434,
+ "▁Canadian": 10435,
+ "▁dialog": 10436,
+ "▁articles": 10437,
+ "Theme": 10438,
+ "SM": 10439,
+ "▁Bul": 10440,
+ "▁leur": 10441,
+ "▁stom": 10442,
+ "Plugin": 10443,
+ "▁после": 10444,
+ "▁stead": 10445,
+ "▁ś": 10446,
+ "ipher": 10447,
+ "▁prze": 10448,
+ "▁draft": 10449,
+ "bottom": 10450,
+ "▁{};": 10451,
+ "▁stayed": 10452,
+ "feature": 10453,
+ "▁vot": 10454,
+ "▁fabric": 10455,
+ "ça": 10456,
+ "('#": 10457,
+ "rea": 10458,
+ "▁reput": 10459,
+ "▁Cir": 10460,
+ "▁AL": 10461,
+ "▁assertEquals": 10462,
+ "results": 10463,
+ "▁Cross": 10464,
+ "ursday": 10465,
+ "▁audio": 10466,
+ "▁gap": 10467,
+ "▁streets": 10468,
+ "▁scientific": 10469,
+ "platform": 10470,
+ "▁auss": 10471,
+ "▁Cro": 10472,
+ "▁partial": 10473,
+ "unc": 10474,
+ "▁choices": 10475,
+ "▁или": 10476,
+ "pred": 10477,
+ "▁heads": 10478,
+ "terday": 10479,
+ "▁Nick": 10480,
+ "▁weird": 10481,
+ "asant": 10482,
+ "▁represented": 10483,
+ "▁пи": 10484,
+ "DP": 10485,
+ "orders": 10486,
+ "clock": 10487,
+ "▁Ho": 10488,
+ "arters": 10489,
+ "Cmd": 10490,
+ "oga": 10491,
+ "Keys": 10492,
+ "Report": 10493,
+ "▁Vill": 10494,
+ "▁Mu": 10495,
+ "▁owned": 10496,
+ "SUCCESS": 10497,
+ "▁typeof": 10498,
+ "hdr": 10499,
+ "uable": 10500,
+ "▁neighborhood": 10501,
+ "▁AP": 10502,
+ "▁resulting": 10503,
+ "▁shadow": 10504,
+ "STRING": 10505,
+ "▁videos": 10506,
+ "лення": 10507,
+ "expect": 10508,
+ "▁Valley": 10509,
+ "▁goto": 10510,
+ "▁Sher": 10511,
+ "frastr": 10512,
+ "▁operating": 10513,
+ "▁это": 10514,
+ "▁Licensed": 10515,
+ "Variable": 10516,
+ "▁PR": 10517,
+ "▁Hans": 10518,
+ "clone": 10519,
+ "▁Gesch": 10520,
+ "▁Band": 10521,
+ "........": 10522,
+ "uing": 10523,
+ "▁hundreds": 10524,
+ "▁ок": 10525,
+ "▁emotional": 10526,
+ "▁Indust": 10527,
+ ")+": 10528,
+ "▁Egypt": 10529,
+ "▁franç": 10530,
+ "▁š": 10531,
+ "▁fasc": 10532,
+ "onto": 10533,
+ "▁Adam": 10534,
+ "▁laid": 10535,
+ "▁rig": 10536,
+ "▁detailed": 10537,
+ "▁implements": 10538,
+ "▁university": 10539,
+ "▁Hy": 10540,
+ "▁grid": 10541,
+ "▁regions": 10542,
+ "Stop": 10543,
+ "▁slot": 10544,
+ "▁angry": 10545,
+ "▁-=": 10546,
+ "▁waited": 10547,
+ "Vert": 10548,
+ "\":\"": 10549,
+ "▁elem": 10550,
+ "▁rég": 10551,
+ "owed": 10552,
+ "Member": 10553,
+ "▁ratio": 10554,
+ "isen": 10555,
+ "▁Lem": 10556,
+ "gery": 10557,
+ "▁cream": 10558,
+ "▁était": 10559,
+ "▁geb": 10560,
+ "unique": 10561,
+ "▁Deb": 10562,
+ "▁factory": 10563,
+ "że": 10564,
+ "dialog": 10565,
+ "▁Config": 10566,
+ "Sync": 10567,
+ "angers": 10568,
+ "▁governing": 10569,
+ "▁Hun": 10570,
+ "Space": 10571,
+ "▁jest": 10572,
+ "icious": 10573,
+ "▁emphas": 10574,
+ "umps": 10575,
+ "▁Esp": 10576,
+ "▁sul": 10577,
+ "▁historical": 10578,
+ "ija": 10579,
+ "▁lying": 10580,
+ "▁Steve": 10581,
+ "▁measures": 10582,
+ "osto": 10583,
+ "?”": 10584,
+ "▁pocket": 10585,
+ "▁Sat": 10586,
+ "▁pitch": 10587,
+ "▁natur": 10588,
+ "▁humans": 10589,
+ "▁Simon": 10590,
+ "adores": 10591,
+ "(\"\\": 10592,
+ "inking": 10593,
+ "▁expos": 10594,
+ "material": 10595,
+ "▁apparently": 10596,
+ "▁Camb": 10597,
+ "▁Box": 10598,
+ "▁spaces": 10599,
+ "exists": 10600,
+ "▁acting": 10601,
+ "ORY": 10602,
+ "зова": 10603,
+ "Good": 10604,
+ "ienne": 10605,
+ "▁Williams": 10606,
+ "▁fruit": 10607,
+ "iera": 10608,
+ "▁Lim": 10609,
+ "▁trait": 10610,
+ "▁artists": 10611,
+ "▁absor": 10612,
+ "rait": 10613,
+ "LOAD": 10614,
+ "▁movies": 10615,
+ "▁dynamic": 10616,
+ "asts": 10617,
+ "▁Integer": 10618,
+ "▁smoke": 10619,
+ "пі": 10620,
+ "angel": 10621,
+ ">(\"": 10622,
+ "▁instrument": 10623,
+ "▁fuel": 10624,
+ "ної": 10625,
+ "atalogue": 10626,
+ "▁serial": 10627,
+ "Files": 10628,
+ "▁bathroom": 10629,
+ "ilo": 10630,
+ "esto": 10631,
+ "▁pm": 10632,
+ "entials": 10633,
+ "▁Online": 10634,
+ "white": 10635,
+ "▁tips": 10636,
+ "▁capable": 10637,
+ "Fig": 10638,
+ "TV": 10639,
+ "▁он": 10640,
+ "ké": 10641,
+ "bitr": 10642,
+ "Mapping": 10643,
+ "▁tak": 10644,
+ "ющи": 10645,
+ "вля": 10646,
+ ")\",": 10647,
+ "▁Karl": 10648,
+ "▁Human": 10649,
+ "▁Pot": 10650,
+ "▁represents": 10651,
+ "▁consistent": 10652,
+ "_(": 10653,
+ "wen": 10654,
+ "▁Rose": 10655,
+ "law": 10656,
+ "▁FROM": 10657,
+ "▁begins": 10658,
+ "▁edit": 10659,
+ "▁mountain": 10660,
+ "▁chapter": 10661,
+ "▁wondered": 10662,
+ "▁industrial": 10663,
+ "▁Major": 10664,
+ "▁ges": 10665,
+ "▁directed": 10666,
+ "eros": 10667,
+ "▁Wild": 10668,
+ "liament": 10669,
+ "Book": 10670,
+ "username": 10671,
+ "hot": 10672,
+ "▁nam": 10673,
+ "▁league": 10674,
+ "bra": 10675,
+ "кон": 10676,
+ "▁Tal": 10677,
+ "▁Ва": 10678,
+ "▁exports": 10679,
+ "(@": 10680,
+ "▁sharing": 10681,
+ "▁Tro": 10682,
+ "ść": 10683,
+ "uesday": 10684,
+ "ylv": 10685,
+ "▁guitar": 10686,
+ "elen": 10687,
+ "Selection": 10688,
+ "▁confident": 10689,
+ "rypto": 10690,
+ "▁hors": 10691,
+ "editor": 10692,
+ "▁shoulders": 10693,
+ "getName": 10694,
+ "encing": 10695,
+ "SELECT": 10696,
+ "вши": 10697,
+ "▁kinds": 10698,
+ "▁Wel": 10699,
+ "▁purposes": 10700,
+ "Matrix": 10701,
+ "invalid": 10702,
+ "▁owners": 10703,
+ "▁Records": 10704,
+ "▁Process": 10705,
+ "▁chat": 10706,
+ "▁Dor": 10707,
+ "▁bin": 10708,
+ "redit": 10709,
+ "oire": 10710,
+ "▁Total": 10711,
+ "▁Family": 10712,
+ "ARY": 10713,
+ "▁bread": 10714,
+ "▁compre": 10715,
+ "▁shoes": 10716,
+ "▁raz": 10717,
+ "▁trace": 10718,
+ "nej": 10719,
+ "orted": 10720,
+ "hn": 10721,
+ "▁procedure": 10722,
+ "properties": 10723,
+ "plier": 10724,
+ "▁hero": 10725,
+ "panel": 10726,
+ "▁marked": 10727,
+ "▁worried": 10728,
+ "\\|": 10729,
+ "pts": 10730,
+ "▁Support": 10731,
+ "▁serving": 10732,
+ "Fail": 10733,
+ "▁disappoint": 10734,
+ "▁Scot": 10735,
+ "▁pleasure": 10736,
+ "▁judge": 10737,
+ "zeich": 10738,
+ "▁forever": 10739,
+ "▁Zeit": 10740,
+ "uous": 10741,
+ "inent": 10742,
+ "▁dw": 10743,
+ "▁waren": 10744,
+ "▁flash": 10745,
+ "▁troops": 10746,
+ "▁drugs": 10747,
+ "▁diam": 10748,
+ ".~": 10749,
+ "imp": 10750,
+ "inned": 10751,
+ "▁EV": 10752,
+ "Struct": 10753,
+ "▁justice": 10754,
+ "▁officials": 10755,
+ "ffff": 10756,
+ "▁Common": 10757,
+ "▁Cat": 10758,
+ "▁tomorrow": 10759,
+ "▁él": 10760,
+ "Texture": 10761,
+ "qpoint": 10762,
+ "▁Fried": 10763,
+ "▁Term": 10764,
+ "pgfqpoint": 10765,
+ "▁nem": 10766,
+ "norm": 10767,
+ "▁hardly": 10768,
+ "oda": 10769,
+ "zeta": 10770,
+ "emic": 10771,
+ "▁полу": 10772,
+ "▁loaded": 10773,
+ "kes": 10774,
+ "ció": 10775,
+ "▁fool": 10776,
+ "▁trick": 10777,
+ "▁dst": 10778,
+ "Find": 10779,
+ "▁все": 10780,
+ "}},": 10781,
+ "▁framework": 10782,
+ "▁merely": 10783,
+ "▁union": 10784,
+ "▁Edward": 10785,
+ "rif": 10786,
+ "Flag": 10787,
+ "▁crisis": 10788,
+ "▁finite": 10789,
+ "▁lol": 10790,
+ "▁Kim": 10791,
+ "ната": 10792,
+ "since": 10793,
+ "▁compat": 10794,
+ "▁pert": 10795,
+ "ibilities": 10796,
+ "▁también": 10797,
+ "ibli": 10798,
+ "▁teen": 10799,
+ "▁sympt": 10800,
+ "oral": 10801,
+ "ders": 10802,
+ "otte": 10803,
+ "при": 10804,
+ "▁Jane": 10805,
+ "▁originally": 10806,
+ "▁throat": 10807,
+ "mag": 10808,
+ "sup": 10809,
+ "uni": 10810,
+ "$$": 10811,
+ "▁Library": 10812,
+ "▁attacks": 10813,
+ "ingen": 10814,
+ "('/": 10815,
+ "▁hes": 10816,
+ "coin": 10817,
+ "ounce": 10818,
+ "▁Academy": 10819,
+ "MODULE": 10820,
+ "isms": 10821,
+ "▁Adv": 10822,
+ "▁Bol": 10823,
+ "▁incident": 10824,
+ ")^{": 10825,
+ "▁bij": 10826,
+ "▁Rome": 10827,
+ "▁Italy": 10828,
+ "events": 10829,
+ "▁Fern": 10830,
+ "▁ber": 10831,
+ "▁silent": 10832,
+ "▁pier": 10833,
+ "▁YO": 10834,
+ "▁plain": 10835,
+ "Bas": 10836,
+ "▁pill": 10837,
+ "rase": 10838,
+ "▁carrying": 10839,
+ "▁resp": 10840,
+ "ную": 10841,
+ "▁typical": 10842,
+ "Wrapper": 10843,
+ "▁gau": 10844,
+ "▁chemical": 10845,
+ "▁hal": 10846,
+ "throw": 10847,
+ "Cluster": 10848,
+ "▁Gab": 10849,
+ "▁Girl": 10850,
+ "quir": 10851,
+ "▁Arg": 10852,
+ "▁relief": 10853,
+ "▁Ве": 10854,
+ "dm": 10855,
+ "▁frustr": 10856,
+ "\\%": 10857,
+ "▁stores": 10858,
+ "▁bottle": 10859,
+ "▁Lew": 10860,
+ "two": 10861,
+ "stad": 10862,
+ "▁cheek": 10863,
+ "▁concerns": 10864,
+ "▁helpful": 10865,
+ "▁coverage": 10866,
+ "isi": 10867,
+ "ADD": 10868,
+ "async": 10869,
+ "▁approximately": 10870,
+ "iffer": 10871,
+ "hook": 10872,
+ "▁enum": 10873,
+ "ová": 10874,
+ "▁evil": 10875,
+ "▁constantly": 10876,
+ "apply": 10877,
+ "▁siè": 10878,
+ "▁practices": 10879,
+ "▁teachers": 10880,
+ "▁Sn": 10881,
+ "▁Awards": 10882,
+ "▁substant": 10883,
+ "▁$.": 10884,
+ "dk": 10885,
+ "▁mob": 10886,
+ "▁ingred": 10887,
+ "vere": 10888,
+ "Multi": 10889,
+ "пер": 10890,
+ "stal": 10891,
+ "yard": 10892,
+ "required": 10893,
+ "vement": 10894,
+ "▁intelligence": 10895,
+ "▁thinks": 10896,
+ "▁personally": 10897,
+ "▁trained": 10898,
+ "orney": 10899,
+ ")": 10900,
+ "gged": 10901,
+ "EINVAL": 10902,
+ "arna": 10903,
+ "▁Hamilton": 10904,
+ "merce": 10905,
+ "ekt": 10906,
+ "OF": 10907,
+ ")[": 10908,
+ "rug": 10909,
+ "ición": 10910,
+ "▁survey": 10911,
+ "nesday": 10912,
+ "▁pag": 10913,
+ "▁boundary": 10914,
+ "▁quantum": 10915,
+ "▁drawing": 10916,
+ "▁volunte": 10917,
+ "▁Word": 10918,
+ "sky": 10919,
+ "▁Greg": 10920,
+ "coll": 10921,
+ "hide": 10922,
+ "▁swim": 10923,
+ "▁revealed": 10924,
+ "adv": 10925,
+ "дя": 10926,
+ ".\");": 10927,
+ "▁explan": 10928,
+ "▁Current": 10929,
+ "▁gotten": 10930,
+ "▁falling": 10931,
+ "▁contained": 10932,
+ "UND": 10933,
+ "▁Should": 10934,
+ "▁killing": 10935,
+ "▁aspects": 10936,
+ "icted": 10937,
+ "▁Param": 10938,
+ "\",\r": 10939,
+ "TION": 10940,
+ "));\r": 10941,
+ "▁Iran": 10942,
+ "beit": 10943,
+ "▁Bu": 10944,
+ "▁[],": 10945,
+ "SSION": 10946,
+ "▁Mah": 10947,
+ "▁resolution": 10948,
+ "▁boss": 10949,
+ "lg": 10950,
+ "chor": 10951,
+ "▁Unter": 10952,
+ "▁debt": 10953,
+ "▁vid": 10954,
+ "gie": 10955,
+ "▁uno": 10956,
+ "CB": 10957,
+ "plom": 10958,
+ "LICENSE": 10959,
+ "▁Kenn": 10960,
+ "▁finns": 10961,
+ "ONG": 10962,
+ "▁somewhat": 10963,
+ "▁actor": 10964,
+ "▁Status": 10965,
+ "▁probability": 10966,
+ "fb": 10967,
+ "▁chart": 10968,
+ "▁stands": 10969,
+ "policy": 10970,
+ "▁onder": 10971,
+ "tabular": 10972,
+ "▁Ash": 10973,
+ "▁boost": 10974,
+ "▁desper": 10975,
+ "month": 10976,
+ "▁alert": 10977,
+ "▁suite": 10978,
+ "▁gén": 10979,
+ "▁vacc": 10980,
+ "▁Has": 10981,
+ "Mask": 10982,
+ "▁Thursday": 10983,
+ "▁proved": 10984,
+ "▁Nel": 10985,
+ "▁moral": 10986,
+ "▁ja": 10987,
+ "auer": 10988,
+ "codec": 10989,
+ "▁instant": 10990,
+ "amps": 10991,
+ "▁milk": 10992,
+ "WORD": 10993,
+ "▁Ö": 10994,
+ "Email": 10995,
+ "Elements": 10996,
+ "▁forma": 10997,
+ "Free": 10998,
+ "MAP": 10999,
+ "▁Ж": 11000,
+ "sym": 11001,
+ "▁ти": 11002,
+ "▁Econom": 11003,
+ "▁Vi": 11004,
+ "▁Columb": 11005,
+ "▁_,": 11006,
+ "oret": 11007,
+ "Sequ": 11008,
+ "plan": 11009,
+ "▁frequency": 11010,
+ "irement": 11011,
+ "▁assumed": 11012,
+ "▁Ca": 11013,
+ "▁Bit": 11014,
+ "▁коман": 11015,
+ "▁smell": 11016,
+ "Security": 11017,
+ "▁aqu": 11018,
+ "oor": 11019,
+ "price": 11020,
+ "inity": 11021,
+ "▁axis": 11022,
+ "release": 11023,
+ "▁resolve": 11024,
+ "▁tears": 11025,
+ "▁bother": 11026,
+ "▁Community": 11027,
+ "▁registered": 11028,
+ "▁revolution": 11029,
+ "?.": 11030,
+ "▁versions": 11031,
+ "%%%%": 11032,
+ "ydro": 11033,
+ "Success": 11034,
+ "▁Win": 11035,
+ "▁Boy": 11036,
+ "▁Dub": 11037,
+ "▁kw": 11038,
+ "▁noch": 11039,
+ "▁charges": 11040,
+ "arios": 11041,
+ "uar": 11042,
+ ";&": 11043,
+ "▁había": 11044,
+ "(`": 11045,
+ "▁tx": 11046,
+ "elve": 11047,
+ "▁años": 11048,
+ "▁math": 11049,
+ "▁Alf": 11050,
+ "▁Fund": 11051,
+ "▁manifest": 11052,
+ "▁attached": 11053,
+ "▁spiritual": 11054,
+ "▁Alexander": 11055,
+ "unes": 11056,
+ "▁seed": 11057,
+ "▁Но": 11058,
+ "▁magazine": 11059,
+ "▁eigen": 11060,
+ "▁обра": 11061,
+ "ea": 11062,
+ "▁PH": 11063,
+ "swing": 11064,
+ "▁Asia": 11065,
+ "ју": 11066,
+ "▁KIND": 11067,
+ "Identifier": 11068,
+ "once": 11069,
+ "▁alcohol": 11070,
+ "ції": 11071,
+ "styles": 11072,
+ "assertEqual": 11073,
+ "▁Ra": 11074,
+ "графи": 11075,
+ "▁millions": 11076,
+ "▁chunk": 11077,
+ "дер": 11078,
+ "Package": 11079,
+ "UST": 11080,
+ "▁Nothing": 11081,
+ "(\"#": 11082,
+ "▁Mid": 11083,
+ "▁нача": 11084,
+ "ły": 11085,
+ "AAAA": 11086,
+ "▁launched": 11087,
+ "▁wake": 11088,
+ "▁guests": 11089,
+ "▁differences": 11090,
+ "udi": 11091,
+ "▁aid": 11092,
+ "▁Sport": 11093,
+ "ulator": 11094,
+ "execute": 11095,
+ "plot": 11096,
+ "ching": 11097,
+ "▁Norm": 11098,
+ "tm": 11099,
+ "\\+": 11100,
+ "ARD": 11101,
+ "▁beer": 11102,
+ "▁під": 11103,
+ "IAL": 11104,
+ "storage": 11105,
+ "▁Anna": 11106,
+ "▁yards": 11107,
+ "▁technique": 11108,
+ "▁où": 11109,
+ "atten": 11110,
+ "UNT": 11111,
+ "don": 11112,
+ "фор": 11113,
+ "▁hoping": 11114,
+ "▁victory": 11115,
+ "itat": 11116,
+ "▁significantly": 11117,
+ "▁practical": 11118,
+ "ije": 11119,
+ "▁expansion": 11120,
+ "JS": 11121,
+ "ixels": 11122,
+ "USER": 11123,
+ "Shape": 11124,
+ "▁extent": 11125,
+ "lio": 11126,
+ "▁pued": 11127,
+ "olid": 11128,
+ "▁gam": 11129,
+ "▁sevent": 11130,
+ "▁Ga": 11131,
+ "anguages": 11132,
+ "(((": 11133,
+ "ъл": 11134,
+ "▁Exper": 11135,
+ "asty": 11136,
+ "rieg": 11137,
+ "gio": 11138,
+ "odo": 11139,
+ "▁colle": 11140,
+ "▁stored": 11141,
+ "▁Sche": 11142,
+ "istant": 11143,
+ "▁lip": 11144,
+ "BR": 11145,
+ "▁aug": 11146,
+ "▁Search": 11147,
+ ")=\\": 11148,
+ "▁Ur": 11149,
+ "▁sole": 11150,
+ "illo": 11151,
+ "▁mehr": 11152,
+ "kit": 11153,
+ "▁interior": 11154,
+ "LIST": 11155,
+ "adel": 11156,
+ "▁shopping": 11157,
+ "▁slä": 11158,
+ "Your": 11159,
+ "DITION": 11160,
+ "▁Http": 11161,
+ "raham": 11162,
+ "три": 11163,
+ "▁brings": 11164,
+ "Rev": 11165,
+ "▁propag": 11166,
+ "ityEngine": 11167,
+ "()),": 11168,
+ "▁ingår": 11169,
+ "▁Ireland": 11170,
+ "▁\"./": 11171,
+ "▁Harr": 11172,
+ "▁admin": 11173,
+ "eno": 11174,
+ "▁kr": 11175,
+ "▁está": 11176,
+ "▁props": 11177,
+ "tok": 11178,
+ "omorph": 11179,
+ "▁affected": 11180,
+ "Phone": 11181,
+ "▁degrees": 11182,
+ "some": 11183,
+ "▁nin": 11184,
+ "EVENT": 11185,
+ "▁interaction": 11186,
+ "▁Tuesday": 11187,
+ "iterator": 11188,
+ "▁Nob": 11189,
+ "▁scatter": 11190,
+ "ucket": 11191,
+ "complete": 11192,
+ "▁duty": 11193,
+ "▁answers": 11194,
+ "Progress": 11195,
+ "eed": 11196,
+ "рон": 11197,
+ "▁vie": 11198,
+ "▁depos": 11199,
+ "▁packet": 11200,
+ "▁tow": 11201,
+ "▁deleg": 11202,
+ "audio": 11203,
+ "▁vary": 11204,
+ "▁migr": 11205,
+ "фі": 11206,
+ "esa": 11207,
+ "Events": 11208,
+ "haus": 11209,
+ "▁Sav": 11210,
+ "▁Portug": 11211,
+ "▁сто": 11212,
+ "ilation": 11213,
+ "▁metadata": 11214,
+ "las": 11215,
+ "▁ai": 11216,
+ "▁anger": 11217,
+ "▁ham": 11218,
+ "▁Anal": 11219,
+ "▁frequently": 11220,
+ "▁FALSE": 11221,
+ "oche": 11222,
+ "rez": 11223,
+ "▁Viet": 11224,
+ "quis": 11225,
+ "▁charged": 11226,
+ "äs": 11227,
+ "▁Path": 11228,
+ "▁accurate": 11229,
+ "▁Plus": 11230,
+ "keit": 11231,
+ "▁Input": 11232,
+ "when": 11233,
+ "eras": 11234,
+ "▁воз": 11235,
+ "▁derived": 11236,
+ "aje": 11237,
+ "▁Had": 11238,
+ "uren": 11239,
+ "ór": 11240,
+ "}=\\": 11241,
+ "ureau": 11242,
+ "aland": 11243,
+ "Execution": 11244,
+ "eden": 11245,
+ "▁seeking": 11246,
+ "changed": 11247,
+ "▁trem": 11248,
+ "ску": 11249,
+ "▁Geme": 11250,
+ "inating": 11251,
+ "▁columns": 11252,
+ "EP": 11253,
+ "▁injury": 11254,
+ "endent": 11255,
+ "▁headed": 11256,
+ "ASE": 11257,
+ "▁Muslim": 11258,
+ "▁climate": 11259,
+ "▁fake": 11260,
+ "CMD": 11261,
+ "ји": 11262,
+ "▁Arts": 11263,
+ "fection": 11264,
+ "▁pit": 11265,
+ ">\\": 11266,
+ "anal": 11267,
+ "Section": 11268,
+ "plus": 11269,
+ "üt": 11270,
+ "▁embed": 11271,
+ "▁strings": 11272,
+ "Before": 11273,
+ "proc": 11274,
+ "▁спо": 11275,
+ "trl": 11276,
+ "vr": 11277,
+ "Background": 11278,
+ "logger": 11279,
+ "agraph": 11280,
+ "iest": 11281,
+ "▁goods": 11282,
+ "batch": 11283,
+ "▁optional": 11284,
+ "▁Taylor": 11285,
+ "▁recognize": 11286,
+ "walk": 11287,
+ "▁Hit": 11288,
+ "▁Elizabeth": 11289,
+ "}:": 11290,
+ "▁careful": 11291,
+ "краї": 11292,
+ "▁locations": 11293,
+ "▁structures": 11294,
+ "▁disk": 11295,
+ "▁ships": 11296,
+ "▁suo": 11297,
+ "▁sowie": 11298,
+ "▁Ess": 11299,
+ "▁Hash": 11300,
+ "▁reasonable": 11301,
+ "▁Moreover": 11302,
+ "▁formula": 11303,
+ "▁Centre": 11304,
+ "▁residents": 11305,
+ "RS": 11306,
+ "Ids": 11307,
+ "▁Know": 11308,
+ "▁trib": 11309,
+ "▁rés": 11310,
+ "▁stable": 11311,
+ "▁Would": 11312,
+ "▁breaking": 11313,
+ "▁meal": 11314,
+ "▁phen": 11315,
+ "▁fel": 11316,
+ "▁Fred": 11317,
+ "Author": 11318,
+ "▁capture": 11319,
+ "opts": 11320,
+ "▁everywhere": 11321,
+ "▁sque": 11322,
+ "▁moder": 11323,
+ "setup": 11324,
+ "▁Supp": 11325,
+ "▁whenever": 11326,
+ "{(": 11327,
+ "wart": 11328,
+ "▁toe": 11329,
+ "Prefix": 11330,
+ "hou": 11331,
+ "gage": 11332,
+ ">\"": 11333,
+ "▁frag": 11334,
+ "▁Theorem": 11335,
+ "memory": 11336,
+ "▁contents": 11337,
+ "docs": 11338,
+ "}'": 11339,
+ "▁Irish": 11340,
+ "Then": 11341,
+ "aats": 11342,
+ "Save": 11343,
+ "▁agency": 11344,
+ "▁име": 11345,
+ "дова": 11346,
+ "▁Function": 11347,
+ "NN": 11348,
+ "destroy": 11349,
+ "▁Message": 11350,
+ "▁cancel": 11351,
+ "▁superior": 11352,
+ "▁ec": 11353,
+ "▁literature": 11354,
+ "▁PART": 11355,
+ "Il": 11356,
+ "▁Cab": 11357,
+ "engine": 11358,
+ "▁basket": 11359,
+ "worth": 11360,
+ "▁Sel": 11361,
+ "fetch": 11362,
+ "▁Stadt": 11363,
+ "▁Ки": 11364,
+ "▁conj": 11365,
+ "▁seiner": 11366,
+ "▁confirmed": 11367,
+ "▁Argent": 11368,
+ "amar": 11369,
+ "pgfpath": 11370,
+ "▁struggle": 11371,
+ "Pattern": 11372,
+ "▁Middle": 11373,
+ "itan": 11374,
+ "▁moon": 11375,
+ "orough": 11376,
+ "▁Catholic": 11377,
+ "▁struck": 11378,
+ "]->": 11379,
+ "▁weapon": 11380,
+ "▁subst": 11381,
+ "▁instructions": 11382,
+ "▁occas": 11383,
+ "protected": 11384,
+ "▁Less": 11385,
+ "▁batch": 11386,
+ "▁contra": 11387,
+ "▁deck": 11388,
+ "▁ignored": 11389,
+ "▁refused": 11390,
+ "trigger": 11391,
+ "▁criminal": 11392,
+ "GA": 11393,
+ "olly": 11394,
+ "▁Bell": 11395,
+ "▁Ю": 11396,
+ "forward": 11397,
+ "▁prefix": 11398,
+ "▁immediate": 11399,
+ "▁assigned": 11400,
+ "▁elected": 11401,
+ "▁tonight": 11402,
+ "▁Dies": 11403,
+ "▁Beach": 11404,
+ "▁preced": 11405,
+ "ował": 11406,
+ "▁galax": 11407,
+ "▁logic": 11408,
+ "enza": 11409,
+ "▁Captain": 11410,
+ "▁Hay": 11411,
+ "▁facts": 11412,
+ "▁ни": 11413,
+ "té": 11414,
+ "▁sb": 11415,
+ "oped": 11416,
+ "▁combat": 11417,
+ "▁explore": 11418,
+ "▁(-": 11419,
+ "Loader": 11420,
+ "▁Wilson": 11421,
+ "▁locked": 11422,
+ ":": 11423,
+ "▁Od": 11424,
+ "▁Prote": 11425,
+ "▁disabled": 11426,
+ "▁hatte": 11427,
+ "▁shout": 11428,
+ "▁constructor": 11429,
+ "бі": 11430,
+ "▁tras": 11431,
+ "▁Father": 11432,
+ "▁adj": 11433,
+ "▁Carolina": 11434,
+ "▁Food": 11435,
+ "bad": 11436,
+ "atore": 11437,
+ "parameters": 11438,
+ "▁Full": 11439,
+ "[-": 11440,
+ "▁\"#": 11441,
+ "▁Try": 11442,
+ "ської": 11443,
+ "▁exhaust": 11444,
+ "▁scroll": 11445,
+ "_;": 11446,
+ "Who": 11447,
+ "▁delivered": 11448,
+ "▁referred": 11449,
+ "▁prospect": 11450,
+ "scan": 11451,
+ "▁modified": 11452,
+ "Generator": 11453,
+ "▁excess": 11454,
+ "▁kg": 11455,
+ "zet": 11456,
+ "icz": 11457,
+ "clipse": 11458,
+ "▁tank": 11459,
+ "▁guns": 11460,
+ "▁Ges": 11461,
+ "inton": 11462,
+ "▁Wednesday": 11463,
+ "▁mainly": 11464,
+ "parser": 11465,
+ "▁effectively": 11466,
+ "▁Ку": 11467,
+ "▁resident": 11468,
+ "▁Li": 11469,
+ "▁flying": 11470,
+ "▁mayor": 11471,
+ "üh": 11472,
+ "uta": 11473,
+ "▁colour": 11474,
+ "▁aircraft": 11475,
+ "terior": 11476,
+ "nr": 11477,
+ "▁keeps": 11478,
+ "fan": 11479,
+ "▁shirt": 11480,
+ "Compar": 11481,
+ "▁Eth": 11482,
+ "Mac": 11483,
+ "clean": 11484,
+ "slice": 11485,
+ "czy": 11486,
+ "▁gender": 11487,
+ "▁butter": 11488,
+ "AUT": 11489,
+ "▁Element": 11490,
+ "Fin": 11491,
+ "dma": 11492,
+ "sample": 11493,
+ "Registry": 11494,
+ "▁classic": 11495,
+ "▁drove": 11496,
+ "pb": 11497,
+ "defined": 11498,
+ "▁reward": 11499,
+ "yal": 11500,
+ "]),": 11501,
+ "▁BAS": 11502,
+ "▁hyper": 11503,
+ "▁Ни": 11504,
+ "▁).": 11505,
+ "Psi": 11506,
+ "▁entries": 11507,
+ "▁Kingdom": 11508,
+ "▁Song": 11509,
+ "▁prompt": 11510,
+ "centering": 11511,
+ "▁Holly": 11512,
+ "eman": 11513,
+ "▁painting": 11514,
+ "▁formation": 11515,
+ "▁Request": 11516,
+ "controller": 11517,
+ "Region": 11518,
+ "PY": 11519,
+ "idades": 11520,
+ "TL": 11521,
+ "▁disable": 11522,
+ "▁rein": 11523,
+ "rical": 11524,
+ "\"\r": 11525,
+ "%)": 11526,
+ "▁Sab": 11527,
+ "▁Without": 11528,
+ "Serv": 11529,
+ "▁Short": 11530,
+ "▁ю": 11531,
+ "▁resc": 11532,
+ "▁patterns": 11533,
+ "▁ArrayList": 11534,
+ "symbol": 11535,
+ "aco": 11536,
+ "▁Hom": 11537,
+ "help": 11538,
+ "▁hasta": 11539,
+ "▁installed": 11540,
+ "atie": 11541,
+ "▁visited": 11542,
+ "▁Бе": 11543,
+ "){\\": 11544,
+ "▁desde": 11545,
+ "JECT": 11546,
+ "▁drew": 11547,
+ "▁Stock": 11548,
+ "▁Cru": 11549,
+ "DEF": 11550,
+ "obby": 11551,
+ "izable": 11552,
+ "ogether": 11553,
+ "▁aber": 11554,
+ "▁dan": 11555,
+ "alis": 11556,
+ "tail": 11557,
+ "▁expressed": 11558,
+ "▁Access": 11559,
+ "Seg": 11560,
+ "▁Lib": 11561,
+ "▁supports": 11562,
+ "background": 11563,
+ "▁commune": 11564,
+ "called": 11565,
+ "▁printf": 11566,
+ "▁Prince": 11567,
+ "ните": 11568,
+ "depend": 11569,
+ "▁dels": 11570,
+ "neur": 11571,
+ "▁recommended": 11572,
+ "▁founded": 11573,
+ "▁markets": 11574,
+ "▁destroyed": 11575,
+ "▁abstract": 11576,
+ "▁serie": 11577,
+ "▁Dun": 11578,
+ "Term": 11579,
+ "▁portion": 11580,
+ "adapter": 11581,
+ "isset": 11582,
+ "чески": 11583,
+ "▁integer": 11584,
+ "▁returning": 11585,
+ "enties": 11586,
+ "▁Fair": 11587,
+ "▁USB": 11588,
+ "▁Price": 11589,
+ "igate": 11590,
+ "▁settled": 11591,
+ "({\\": 11592,
+ "nek": 11593,
+ "▁therm": 11594,
+ "▁cig": 11595,
+ "ány": 11596,
+ "▁investigation": 11597,
+ "ometer": 11598,
+ "SUP": 11599,
+ "Some": 11600,
+ "sing": 11601,
+ "Constant": 11602,
+ "▁retail": 11603,
+ "ży": 11604,
+ "▁drinking": 11605,
+ "▁Invest": 11606,
+ "SV": 11607,
+ "iginal": 11608,
+ "▁Bow": 11609,
+ "{{\\": 11610,
+ "▁assistance": 11611,
+ "▁intellect": 11612,
+ "INIT": 11613,
+ "aug": 11614,
+ "▁Leon": 11615,
+ "Sur": 11616,
+ "▁admit": 11617,
+ "▁Command": 11618,
+ "illes": 11619,
+ "rov": 11620,
+ "▁oh": 11621,
+ "▁não": 11622,
+ "▁matching": 11623,
+ "▁genu": 11624,
+ "▁Ox": 11625,
+ "тся": 11626,
+ "notation": 11627,
+ "GO": 11628,
+ "▁Nap": 11629,
+ "▁verify": 11630,
+ "▁aussi": 11631,
+ "DateTime": 11632,
+ "▁suitable": 11633,
+ "▁indicate": 11634,
+ "▁Live": 11635,
+ "Feature": 11636,
+ "▁tracks": 11637,
+ "▁hasn": 11638,
+ "▁Java": 11639,
+ "▁closely": 11640,
+ "▁Dad": 11641,
+ "ceive": 11642,
+ "▁Market": 11643,
+ "agy": 11644,
+ "▁\"-": 11645,
+ "awn": 11646,
+ "stell": 11647,
+ "pton": 11648,
+ "zeit": 11649,
+ "▁Vector": 11650,
+ "▁MAX": 11651,
+ "▁Federal": 11652,
+ "wall": 11653,
+ "▁Jen": 11654,
+ "delay": 11655,
+ "▁limits": 11656,
+ "▁Quest": 11657,
+ "Cam": 11658,
+ "▁Fel": 11659,
+ "writer": 11660,
+ "LP": 11661,
+ "▁moves": 11662,
+ "▁Execut": 11663,
+ "▁DB": 11664,
+ "oker": 11665,
+ "scribe": 11666,
+ "elijk": 11667,
+ "Constants": 11668,
+ "Addr": 11669,
+ "▁}}": 11670,
+ "▁channels": 11671,
+ "iy": 11672,
+ "riority": 11673,
+ "▁trading": 11674,
+ "▁facilities": 11675,
+ "▁Pack": 11676,
+ "▁sys": 11677,
+ "▁meta": 11678,
+ "▁estimate": 11679,
+ "▁Later": 11680,
+ "issue": 11681,
+ "▁Having": 11682,
+ "▁guest": 11683,
+ "▁nobody": 11684,
+ "depth": 11685,
+ "▁został": 11686,
+ "пера": 11687,
+ ")}\\": 11688,
+ "bg": 11689,
+ "▁Twitter": 11690,
+ "▁darkness": 11691,
+ "jpg": 11692,
+ "contr": 11693,
+ "kernel": 11694,
+ "]\\": 11695,
+ "▁extend": 11696,
+ "roc": 11697,
+ "NET": 11698,
+ "MSG": 11699,
+ "▁burst": 11700,
+ "▁repair": 11701,
+ "▁fetch": 11702,
+ "ieg": 11703,
+ "ús": 11704,
+ "Screen": 11705,
+ "blem": 11706,
+ "AppCompat": 11707,
+ "▁chap": 11708,
+ "ELD": 11709,
+ "▁Penn": 11710,
+ "▁promote": 11711,
+ "▁Ukr": 11712,
+ "arest": 11713,
+ "▁samples": 11714,
+ "▁Greek": 11715,
+ "▁constru": 11716,
+ "▁universe": 11717,
+ "elijke": 11718,
+ "▁preferred": 11719,
+ "▁Де": 11720,
+ "▁Ira": 11721,
+ "▁dow": 11722,
+ "agues": 11723,
+ "HERE": 11724,
+ "▁experts": 11725,
+ "Protocol": 11726,
+ "PIO": 11727,
+ "▁naz": 11728,
+ "▁Kh": 11729,
+ "hör": 11730,
+ "▁distingu": 11731,
+ "▁BY": 11732,
+ "▁seine": 11733,
+ "eping": 11734,
+ "▁fairly": 11735,
+ "▁Mean": 11736,
+ "ixer": 11737,
+ "insi": 11738,
+ "▁authors": 11739,
+ "**.": 11740,
+ "AI": 11741,
+ "▁edges": 11742,
+ "▁shooting": 11743,
+ "Admin": 11744,
+ "▁maps": 11745,
+ "chant": 11746,
+ "▁COVID": 11747,
+ "▁linked": 11748,
+ "▁ske": 11749,
+ "▁powers": 11750,
+ "ád": 11751,
+ "▁stomach": 11752,
+ "▁usage": 11753,
+ "▁defend": 11754,
+ "▁sustain": 11755,
+ "▁updates": 11756,
+ "▁assign": 11757,
+ "HL": 11758,
+ "▁Sea": 11759,
+ "▁discipl": 11760,
+ "Video": 11761,
+ "▁Chief": 11762,
+ "▁bunch": 11763,
+ "▁Obama": 11764,
+ "nis": 11765,
+ "vor": 11766,
+ "▁agents": 11767,
+ "cas": 11768,
+ "chter": 11769,
+ "▁glanced": 11770,
+ "supported": 11771,
+ "▁Consider": 11772,
+ "▁Everyone": 11773,
+ "▁lect": 11774,
+ "▁Stone": 11775,
+ "▁Jam": 11776,
+ "ogram": 11777,
+ "formance": 11778,
+ "▁\\\"": 11779,
+ "▁patch": 11780,
+ "▁vit": 11781,
+ "Power": 11782,
+ "▁harder": 11783,
+ "Anal": 11784,
+ "▁desired": 11785,
+ "▁jug": 11786,
+ "▁supporting": 11787,
+ "DU": 11788,
+ "]],": 11789,
+ "▁Administr": 11790,
+ "ucky": 11791,
+ "▁controller": 11792,
+ "▁issued": 11793,
+ "▁Sin": 11794,
+ "▁affili": 11795,
+ "▁partners": 11796,
+ "cdots": 11797,
+ "ctic": 11798,
+ "Car": 11799,
+ "▁NY": 11800,
+ "▁priority": 11801,
+ "original": 11802,
+ "Sql": 11803,
+ "▁declared": 11804,
+ "▁Hotel": 11805,
+ "▁browser": 11806,
+ "▁grande": 11807,
+ "}^\\": 11808,
+ "bow": 11809,
+ "▁accommod": 11810,
+ "Directory": 11811,
+ "▁suffering": 11812,
+ "▁logger": 11813,
+ "▁breakfast": 11814,
+ "uli": 11815,
+ "▁boot": 11816,
+ "▁contribution": 11817,
+ "NESS": 11818,
+ "▁Ten": 11819,
+ "semble": 11820,
+ "▁housing": 11821,
+ "Raw": 11822,
+ "ANCE": 11823,
+ "▁При": 11824,
+ "▁brit": 11825,
+ "essa": 11826,
+ "inson": 11827,
+ "▁Ball": 11828,
+ "entes": 11829,
+ "▁Bra": 11830,
+ "score": 11831,
+ "GER": 11832,
+ "route": 11833,
+ "apsed": 11834,
+ "рой": 11835,
+ "diff": 11836,
+ "▁broadcast": 11837,
+ "▁tar": 11838,
+ "▁delight": 11839,
+ ")?": 11840,
+ "chester": 11841,
+ "Platform": 11842,
+ "▁emergency": 11843,
+ "▁ces": 11844,
+ "nership": 11845,
+ "▁situations": 11846,
+ "▁familjen": 11847,
+ "▁Geb": 11848,
+ "enta": 11849,
+ "úblic": 11850,
+ "▁Place": 11851,
+ "ILL": 11852,
+ "▁march": 11853,
+ "▁fundamental": 11854,
+ "attributes": 11855,
+ "кти": 11856,
+ "▁Fu": 11857,
+ "FD": 11858,
+ "▁рас": 11859,
+ "▁academic": 11860,
+ "pres": 11861,
+ "▁rising": 11862,
+ "▁Braz": 11863,
+ "▁receiving": 11864,
+ "WARN": 11865,
+ "▁judg": 11866,
+ "▁necessarily": 11867,
+ "]=": 11868,
+ "▁deeply": 11869,
+ "▁gray": 11870,
+ "Headers": 11871,
+ "▁coal": 11872,
+ "\\{": 11873,
+ "Mut": 11874,
+ "bach": 11875,
+ "▁profit": 11876,
+ "вого": 11877,
+ "igs": 11878,
+ "ograp": 11879,
+ "\";\r": 11880,
+ "▁advoc": 11881,
+ "Generated": 11882,
+ "мери": 11883,
+ "▁Cond": 11884,
+ "▁agric": 11885,
+ "BASE": 11886,
+ "▁arrang": 11887,
+ "▁flowers": 11888,
+ "iw": 11889,
+ "▁];": 11890,
+ "▁вой": 11891,
+ "umerate": 11892,
+ "▁ihr": 11893,
+ "▁пар": 11894,
+ "▁mont": 11895,
+ "widehat": 11896,
+ "mg": 11897,
+ "▁btn": 11898,
+ "▁besk": 11899,
+ "▁acts": 11900,
+ "ós": 11901,
+ "~~~~": 11902,
+ "▁curve": 11903,
+ "language": 11904,
+ "▁TRUE": 11905,
+ "▁cleaning": 11906,
+ "Math": 11907,
+ "▁regional": 11908,
+ "▁estimated": 11909,
+ "arity": 11910,
+ "ierung": 11911,
+ "/{": 11912,
+ "jango": 11913,
+ "$_": 11914,
+ "▁threw": 11915,
+ "rq": 11916,
+ "cop": 11917,
+ "nergy": 11918,
+ "▁Account": 11919,
+ "pal": 11920,
+ "▁Nic": 11921,
+ "]))": 11922,
+ "▁awesome": 11923,
+ "▁Load": 11924,
+ "unnel": 11925,
+ "▁rows": 11926,
+ "▁foreach": 11927,
+ "▁Pod": 11928,
+ "▁EN": 11929,
+ "▁.=": 11930,
+ "uate": 11931,
+ "frastructure": 11932,
+ "▁Watch": 11933,
+ "Stand": 11934,
+ "▁routine": 11935,
+ "▁pic": 11936,
+ "helper": 11937,
+ "▁horses": 11938,
+ "▁requested": 11939,
+ "▁---": 11940,
+ "border": 11941,
+ "▁lifted": 11942,
+ "▁Ped": 11943,
+ "Import": 11944,
+ "ље": 11945,
+ "▁Ли": 11946,
+ "▁myst": 11947,
+ "THER": 11948,
+ "▁AC": 11949,
+ "Proxy": 11950,
+ "prov": 11951,
+ "▁Nik": 11952,
+ "hemat": 11953,
+ "ональ": 11954,
+ "▁\".": 11955,
+ "ului": 11956,
+ "▁improved": 11957,
+ "ieren": 11958,
+ "ocolate": 11959,
+ "Sche": 11960,
+ "unic": 11961,
+ "▁Professor": 11962,
+ "ieler": 11963,
+ "▁duration": 11964,
+ "▁timeout": 11965,
+ "hom": 11966,
+ "▁lux": 11967,
+ "▁trab": 11968,
+ "itary": 11969,
+ "ње": 11970,
+ "▁inspired": 11971,
+ "})\\": 11972,
+ "isely": 11973,
+ "ials": 11974,
+ "▁Vor": 11975,
+ "▁enhance": 11976,
+ "▁lucky": 11977,
+ "World": 11978,
+ "elo": 11979,
+ "ifiers": 11980,
+ "▁facing": 11981,
+ "▁appreciate": 11982,
+ "▁être": 11983,
+ "▁bench": 11984,
+ "atted": 11985,
+ "gence": 11986,
+ "course": 11987,
+ "▁tub": 11988,
+ "▁lors": 11989,
+ "▁mistake": 11990,
+ "nom": 11991,
+ "▁paus": 11992,
+ "▁\"\";": 11993,
+ "▁subs": 11994,
+ "▁stato": 11995,
+ "$)": 11996,
+ "▁gay": 11997,
+ "orry": 11998,
+ "▁vehicles": 11999,
+ "▁brill": 12000,
+ "may": 12001,
+ "resp": 12002,
+ "▁wore": 12003,
+ "ją": 12004,
+ "bp": 12005,
+ "onel": 12006,
+ "▁CR": 12007,
+ "▁diagn": 12008,
+ "mathsf": 12009,
+ "▁holiday": 12010,
+ "▁achieved": 12011,
+ "▁{'": 12012,
+ "▁Resource": 12013,
+ "▁hi": 12014,
+ "▁bra": 12015,
+ "▁CONDITION": 12016,
+ "ctr": 12017,
+ "▁Write": 12018,
+ "ishop": 12019,
+ "OLD": 12020,
+ "▁cpu": 12021,
+ "▁occurs": 12022,
+ "ół": 12023,
+ "straint": 12024,
+ "▁nuclear": 12025,
+ "Area": 12026,
+ "cluster": 12027,
+ "▁surrounding": 12028,
+ "▁Juan": 12029,
+ "▁prima": 12030,
+ "▁Southern": 12031,
+ "itty": 12032,
+ "▁Assembly": 12033,
+ "elem": 12034,
+ "adi": 12035,
+ "éral": 12036,
+ "▁Wat": 12037,
+ "▁Radio": 12038,
+ "▁gegen": 12039,
+ "▁Tony": 12040,
+ "pressed": 12041,
+ "▁Anne": 12042,
+ "▁NS": 12043,
+ "▁Pak": 12044,
+ "▁Civil": 12045,
+ "▁thrown": 12046,
+ "NONE": 12047,
+ "▁pump": 12048,
+ "▁solve": 12049,
+ "ENABLE": 12050,
+ "▁Phys": 12051,
+ "▁],": 12052,
+ "POSE": 12053,
+ "ktet": 12054,
+ "▁Fab": 12055,
+ "validate": 12056,
+ "Iterator": 12057,
+ "condition": 12058,
+ "redu": 12059,
+ "▁negoti": 12060,
+ "anno": 12061,
+ "▁sans": 12062,
+ "▁Ul": 12063,
+ "CHAR": 12064,
+ "▁edition": 12065,
+ "▁spectrum": 12066,
+ "orie": 12067,
+ "▁execution": 12068,
+ "Please": 12069,
+ "▁BO": 12070,
+ "URN": 12071,
+ "▁cow": 12072,
+ "стан": 12073,
+ "istribution": 12074,
+ "Domain": 12075,
+ "▁readers": 12076,
+ "▁consumer": 12077,
+ "▁styles": 12078,
+ "encode": 12079,
+ "▁Cy": 12080,
+ "Common": 12081,
+ "▁Prop": 12082,
+ "▁execute": 12083,
+ "▁eq": 12084,
+ "▁visitors": 12085,
+ "▁Amb": 12086,
+ "udad": 12087,
+ "qquad": 12088,
+ "▁Cert": 12089,
+ "▁trop": 12090,
+ "▁yesterday": 12091,
+ "tain": 12092,
+ "LD": 12093,
+ "atro": 12094,
+ "▁increases": 12095,
+ "▁Wars": 12096,
+ "ned": 12097,
+ "before": 12098,
+ "aupt": 12099,
+ "▁ERR": 12100,
+ "▁Ford": 12101,
+ "▁dalla": 12102,
+ "ULAR": 12103,
+ "▁strike": 12104,
+ "Arr": 12105,
+ "▁recovery": 12106,
+ "▁Response": 12107,
+ "▁strategies": 12108,
+ "▁ін": 12109,
+ "▁rear": 12110,
+ "▁adults": 12111,
+ "▁Не": 12112,
+ "windows": 12113,
+ "decl": 12114,
+ "olen": 12115,
+ "▁Jord": 12116,
+ "▁Kal": 12117,
+ "▁cui": 12118,
+ "▁Про": 12119,
+ "▁Sever": 12120,
+ "▁ale": 12121,
+ "▁peut": 12122,
+ "Stats": 12123,
+ "▁Ross": 12124,
+ "arten": 12125,
+ "shall": 12126,
+ "▁entertain": 12127,
+ "▁parking": 12128,
+ "нови": 12129,
+ "erre": 12130,
+ "▁funding": 12131,
+ "▁Cle": 12132,
+ "▁Ot": 12133,
+ "unst": 12134,
+ "assertEquals": 12135,
+ "▁cancell": 12136,
+ "TAG": 12137,
+ "▁Early": 12138,
+ "▁feedback": 12139,
+ "▁pand": 12140,
+ "yo": 12141,
+ "▁mirror": 12142,
+ "▁verb": 12143,
+ "▁highlight": 12144,
+ "erialize": 12145,
+ "▁grade": 12146,
+ "лась": 12147,
+ "▁Brook": 12148,
+ "▁LI": 12149,
+ "▁implies": 12150,
+ "▁enorm": 12151,
+ "ają": 12152,
+ "▁Wer": 12153,
+ "away": 12154,
+ "▁machines": 12155,
+ "▁dent": 12156,
+ "Idx": 12157,
+ "▁tid": 12158,
+ ")\"": 12159,
+ "▁mole": 12160,
+ "bold": 12161,
+ "CONT": 12162,
+ "▁ép": 12163,
+ "▁cutting": 12164,
+ "▁Neg": 12165,
+ "▁tong": 12166,
+ "▁networks": 12167,
+ "▁Fall": 12168,
+ "generated": 12169,
+ "▁Pri": 12170,
+ "UEST": 12171,
+ "▁Belg": 12172,
+ "▁sheet": 12173,
+ "кси": 12174,
+ "▁†": 12175,
+ "▁yeah": 12176,
+ "▁Victor": 12177,
+ "▁Rub": 12178,
+ "▁candidates": 12179,
+ "prés": 12180,
+ "▁EU": 12181,
+ "etr": 12182,
+ "▁rolled": 12183,
+ "▁Pas": 12184,
+ "▁Arthur": 12185,
+ "Arch": 12186,
+ "▁Mann": 12187,
+ "American": 12188,
+ "zes": 12189,
+ "inners": 12190,
+ "▁Auto": 12191,
+ "▁professor": 12192,
+ "▁);\r": 12193,
+ "▁addr": 12194,
+ "▁Medical": 12195,
+ "▁fired": 12196,
+ "▁Core": 12197,
+ "▁CONFIG": 12198,
+ "▁sql": 12199,
+ "▁Conserv": 12200,
+ "ichen": 12201,
+ "Vertex": 12202,
+ "▁HO": 12203,
+ "Yeah": 12204,
+ "Note": 12205,
+ "▁OK": 12206,
+ "mus": 12207,
+ "focus": 12208,
+ "aja": 12209,
+ "rá": 12210,
+ "▁hence": 12211,
+ "▁executive": 12212,
+ "▁liquid": 12213,
+ "uje": 12214,
+ "▁driven": 12215,
+ "igue": 12216,
+ "▁Wik": 12217,
+ "Rate": 12218,
+ "rand": 12219,
+ "Results": 12220,
+ "▁copies": 12221,
+ "▁tan": 12222,
+ "riteria": 12223,
+ "enen": 12224,
+ "}_\\": 12225,
+ "▁pobl": 12226,
+ "▁southern": 12227,
+ "eln": 12228,
+ "▁zwei": 12229,
+ "▁concrete": 12230,
+ "▁CONDITIONS": 12231,
+ "▁dreams": 12232,
+ "▁minim": 12233,
+ "▁employee": 12234,
+ "▁nap": 12235,
+ "▁suspect": 12236,
+ "Mouse": 12237,
+ "▁therapy": 12238,
+ "aval": 12239,
+ "▁Anth": 12240,
+ "START": 12241,
+ "sters": 12242,
+ "ishment": 12243,
+ "finite": 12244,
+ "WA": 12245,
+ "vy": 12246,
+ "▁mood": 12247,
+ "comfort": 12248,
+ "▁shr": 12249,
+ "▁decade": 12250,
+ "ября": 12251,
+ "▁'#": 12252,
+ "▁dot": 12253,
+ "▁hill": 12254,
+ "arry": 12255,
+ "catch": 12256,
+ "▁jQuery": 12257,
+ "▁corporate": 12258,
+ "▁BASIS": 12259,
+ "▁appointed": 12260,
+ "▁embar": 12261,
+ "ographie": 12262,
+ "▁pressed": 12263,
+ "▁champion": 12264,
+ "emit": 12265,
+ "▁Bed": 12266,
+ "вання": 12267,
+ "Gui": 12268,
+ "▁PUR": 12269,
+ "▁urban": 12270,
+ "▁sentence": 12271,
+ "bury": 12272,
+ "▁Video": 12273,
+ "▁regularly": 12274,
+ "vl": 12275,
+ "▁слу": 12276,
+ "ockey": 12277,
+ "evin": 12278,
+ "ultural": 12279,
+ "▁passage": 12280,
+ "▁состав": 12281,
+ "▁largely": 12282,
+ "orters": 12283,
+ "▁connections": 12284,
+ "▁surprising": 12285,
+ "bc": 12286,
+ "▁strongly": 12287,
+ "ansas": 12288,
+ "▁sist": 12289,
+ "▁extreme": 12290,
+ "whel": 12291,
+ "▁dealing": 12292,
+ "ographic": 12293,
+ "▁Republican": 12294,
+ "▁granted": 12295,
+ "▁CL": 12296,
+ "▁Hope": 12297,
+ "lessly": 12298,
+ "▁upload": 12299,
+ "▁-\\": 12300,
+ "нию": 12301,
+ "▁valuable": 12302,
+ "=[": 12303,
+ "Price": 12304,
+ "issance": 12305,
+ "iens": 12306,
+ "heit": 12307,
+ "▁suggests": 12308,
+ "сло": 12309,
+ "▁jur": 12310,
+ "}|": 12311,
+ "lp": 12312,
+ "▁invited": 12313,
+ "▁deriv": 12314,
+ "IMIT": 12315,
+ "rass": 12316,
+ "▁instruct": 12317,
+ "▁courses": 12318,
+ "äch": 12319,
+ "▁fifty": 12320,
+ "DEVICE": 12321,
+ "ASH": 12322,
+ "▁hip": 12323,
+ "Unknown": 12324,
+ "▁Catalogue": 12325,
+ "▁Roll": 12326,
+ "▁tensor": 12327,
+ "bec": 12328,
+ "été": 12329,
+ "Identity": 12330,
+ "&\\": 12331,
+ "▁Stephen": 12332,
+ "nodes": 12333,
+ "Dim": 12334,
+ "▁consists": 12335,
+ "▁normally": 12336,
+ "ubl": 12337,
+ "▁Police": 12338,
+ "▁Games": 12339,
+ "five": 12340,
+ "Have": 12341,
+ "▁padding": 12342,
+ "eres": 12343,
+ "anth": 12344,
+ "▁puts": 12345,
+ "uminate": 12346,
+ "ovie": 12347,
+ "▁Index": 12348,
+ "blue": 12349,
+ "Scal": 12350,
+ "▁giant": 12351,
+ "TF": 12352,
+ "pson": 12353,
+ "▁victim": 12354,
+ "serial": 12355,
+ "▁Sym": 12356,
+ "Single": 12357,
+ "▁md": 12358,
+ "▁attended": 12359,
+ "▁Stra": 12360,
+ "▁Dark": 12361,
+ ")|": 12362,
+ "▁span": 12363,
+ "▁maintenance": 12364,
+ "▁bind": 12365,
+ "Bean": 12366,
+ "ilarly": 12367,
+ "▁convent": 12368,
+ "▁José": 12369,
+ "udd": 12370,
+ "▁poly": 12371,
+ "▁idx": 12372,
+ "▁asks": 12373,
+ "▁enthus": 12374,
+ "▁suck": 12375,
+ "▁Cou": 12376,
+ "▁Corporation": 12377,
+ "usions": 12378,
+ "opher": 12379,
+ "▁symptoms": 12380,
+ "▁Johann": 12381,
+ "▁пу": 12382,
+ "▁html": 12383,
+ "▁ps": 12384,
+ "earing": 12385,
+ "gesch": 12386,
+ "▁Mother": 12387,
+ "RET": 12388,
+ "▁furniture": 12389,
+ "PF": 12390,
+ "▁Guard": 12391,
+ "pattern": 12392,
+ "▁lovely": 12393,
+ "alg": 12394,
+ "edly": 12395,
+ "sex": 12396,
+ "▁finds": 12397,
+ "Buf": 12398,
+ "▁над": 12399,
+ "▁км": 12400,
+ "▁Por": 12401,
+ "СР": 12402,
+ "Enter": 12403,
+ "▁esta": 12404,
+ "▁тре": 12405,
+ "▁\"*": 12406,
+ "▁Fox": 12407,
+ "▁cock": 12408,
+ "Bundle": 12409,
+ "▁puis": 12410,
+ "▁announce": 12411,
+ "▁guid": 12412,
+ "checked": 12413,
+ "icide": 12414,
+ "neg": 12415,
+ "▁Gil": 12416,
+ "schen": 12417,
+ "ologist": 12418,
+ "iso": 12419,
+ "groups": 12420,
+ "▁somebody": 12421,
+ "Day": 12422,
+ "tras": 12423,
+ "▁compact": 12424,
+ "▁organized": 12425,
+ "▁roles": 12426,
+ "▁hint": 12427,
+ "▁så": 12428,
+ "▁pays": 12429,
+ "▁Си": 12430,
+ "▁hoped": 12431,
+ "▁sail": 12432,
+ "▁Vers": 12433,
+ "▁embr": 12434,
+ "▁bot": 12435,
+ "▁exceed": 12436,
+ "BACK": 12437,
+ "▁gaze": 12438,
+ "▁spons": 12439,
+ "AST": 12440,
+ "▁torch": 12441,
+ "▁newspaper": 12442,
+ "▁Dist": 12443,
+ "▁bass": 12444,
+ "▁hanging": 12445,
+ "▁ears": 12446,
+ "ńsk": 12447,
+ "getValue": 12448,
+ "▁unus": 12449,
+ "▁Ele": 12450,
+ "services": 12451,
+ "▁dressed": 12452,
+ "lav": 12453,
+ "▁пла": 12454,
+ "Private": 12455,
+ "mic": 12456,
+ "▁parser": 12457,
+ "▁sections": 12458,
+ "▁fo": 12459,
+ "Errorf": 12460,
+ "inz": 12461,
+ "örd": 12462,
+ "▁metric": 12463,
+ "URI": 12464,
+ "▁vice": 12465,
+ "RED": 12466,
+ "▁nue": 12467,
+ "revs": 12468,
+ "▁collected": 12469,
+ "oose": 12470,
+ "▁mond": 12471,
+ "▁nas": 12472,
+ "▁Насе": 12473,
+ "▁å": 12474,
+ "Drop": 12475,
+ "▁abuse": 12476,
+ "▁sees": 12477,
+ "▁Hence": 12478,
+ "exec": 12479,
+ "}\\,": 12480,
+ "▁arbitr": 12481,
+ "▁Application": 12482,
+ "family": 12483,
+ "üd": 12484,
+ "▁magnetic": 12485,
+ "▁newly": 12486,
+ "▁reprodu": 12487,
+ "▁writers": 12488,
+ "▁headers": 12489,
+ "ší": 12490,
+ "рт": 12491,
+ "YPE": 12492,
+ "▁schema": 12493,
+ "▁Ce": 12494,
+ "▁Jews": 12495,
+ "▁Record": 12496,
+ "present": 12497,
+ "▁также": 12498,
+ "▁labels": 12499,
+ "Socket": 12500,
+ "▁equations": 12501,
+ "▁medicine": 12502,
+ "▁authorities": 12503,
+ "}`": 12504,
+ "стви": 12505,
+ "▁Corn": 12506,
+ "▁environmental": 12507,
+ "WARE": 12508,
+ "Mer": 12509,
+ "▁само": 12510,
+ "▁Technology": 12511,
+ "▁Saf": 12512,
+ "▁conn": 12513,
+ "▁Um": 12514,
+ "▁Pacific": 12515,
+ "тел": 12516,
+ "jan": 12517,
+ "▁uncertain": 12518,
+ "▁belief": 12519,
+ "counter": 12520,
+ "toBe": 12521,
+ "INS": 12522,
+ "weet": 12523,
+ "Light": 12524,
+ "primary": 12525,
+ "▁featured": 12526,
+ "▁touched": 12527,
+ "HTTP": 12528,
+ "▁tact": 12529,
+ "pository": 12530,
+ "▁eines": 12531,
+ "lass": 12532,
+ "ська": 12533,
+ "▁przez": 12534,
+ "▁fuer": 12535,
+ "▁exciting": 12536,
+ "▁Cub": 12537,
+ "agan": 12538,
+ "VO": 12539,
+ "▁'%": 12540,
+ "▁\\{": 12541,
+ "ubble": 12542,
+ "▁Fol": 12543,
+ "▁Kong": 12544,
+ "▁versch": 12545,
+ "FAIL": 12546,
+ "▁naar": 12547,
+ "ös": 12548,
+ "speed": 12549,
+ "▁territor": 12550,
+ "▁wrap": 12551,
+ "▁Jahre": 12552,
+ "lee": 12553,
+ "▁crossed": 12554,
+ "resolve": 12555,
+ "▁stim": 12556,
+ "Native": 12557,
+ "ursor": 12558,
+ "NotNull": 12559,
+ "▁Albert": 12560,
+ "▁signature": 12561,
+ "▁Ru": 12562,
+ "idas": 12563,
+ "▁decent": 12564,
+ "▁faced": 12565,
+ "▁лю": 12566,
+ "▁Spain": 12567,
+ "▁resistance": 12568,
+ "▁Brian": 12569,
+ "kwargs": 12570,
+ "▁interval": 12571,
+ "▁Ле": 12572,
+ "▁explo": 12573,
+ "▁semi": 12574,
+ "▁widely": 12575,
+ "dx": 12576,
+ "kov": 12577,
+ "▁Come": 12578,
+ "▁knife": 12579,
+ "Asp": 12580,
+ "uno": 12581,
+ "lineto": 12582,
+ "▁Bund": 12583,
+ "Cert": 12584,
+ "▁todo": 12585,
+ "tags": 12586,
+ "▁guarantee": 12587,
+ "▁vital": 12588,
+ "▁fought": 12589,
+ "▁Env": 12590,
+ "HD": 12591,
+ "Lower": 12592,
+ "Tx": 12593,
+ "▁Fa": 12594,
+ "▁anticip": 12595,
+ "Timer": 12596,
+ "mediate": 12597,
+ "▁proven": 12598,
+ "▁partir": 12599,
+ "AE": 12600,
+ "cursor": 12601,
+ "▁wooden": 12602,
+ "▁Contact": 12603,
+ "regs": 12604,
+ "▁provinc": 12605,
+ "▁DC": 12606,
+ "▁memories": 12607,
+ "▁ft": 12608,
+ "▁battery": 12609,
+ "utenant": 12610,
+ "Login": 12611,
+ "ountry": 12612,
+ "▁compens": 12613,
+ "operatorname": 12614,
+ "▁Jacob": 12615,
+ "zed": 12616,
+ "ADDR": 12617,
+ "▁quad": 12618,
+ "*).": 12619,
+ "▁coat": 12620,
+ "▁fir": 12621,
+ "▁Michel": 12622,
+ "▁Standard": 12623,
+ "rf": 12624,
+ "mel": 12625,
+ "▁coeff": 12626,
+ "▁Iraq": 12627,
+ "▁Given": 12628,
+ "нима": 12629,
+ "▁FIT": 12630,
+ "▁peu": 12631,
+ "▁ig": 12632,
+ "▁Case": 12633,
+ "mé": 12634,
+ "▁parallel": 12635,
+ "cio": 12636,
+ "kow": 12637,
+ "▁institutions": 12638,
+ "ícul": 12639,
+ "aban": 12640,
+ "UX": 12641,
+ "▁Sarah": 12642,
+ "▁més": 12643,
+ "▁atmos": 12644,
+ "▁släktet": 12645,
+ "▁brothers": 12646,
+ "▁wanting": 12647,
+ "aaaa": 12648,
+ "▁fest": 12649,
+ "=-": 12650,
+ "▁forty": 12651,
+ "▁creates": 12652,
+ "hh": 12653,
+ "▁Android": 12654,
+ "anches": 12655,
+ "BT": 12656,
+ "upload": 12657,
+ "xis": 12658,
+ "Hz": 12659,
+ "бор": 12660,
+ "RAY": 12661,
+ "ntil": 12662,
+ "▁leaned": 12663,
+ "unda": 12664,
+ "▁ultimately": 12665,
+ "▁tok": 12666,
+ "neh": 12667,
+ "▁lawyer": 12668,
+ "hend": 12669,
+ "▁Vin": 12670,
+ "▁facility": 12671,
+ "▁likes": 12672,
+ "ento": 12673,
+ "Nodes": 12674,
+ "▁entrance": 12675,
+ "atto": 12676,
+ "rett": 12677,
+ "accept": 12678,
+ "theme": 12679,
+ "тан": 12680,
+ "osi": 12681,
+ "▁{},": 12682,
+ "pgfpathlineto": 12683,
+ "good": 12684,
+ "slot": 12685,
+ "▁innoc": 12686,
+ "▁proport": 12687,
+ "▁arrive": 12688,
+ "ého": 12689,
+ "▁pairs": 12690,
+ "▁wrapped": 12691,
+ "▁unw": 12692,
+ "▁explos": 12693,
+ "▁gel": 12694,
+ "Will": 12695,
+ "▁Zealand": 12696,
+ "ías": 12697,
+ "▁Jr": 12698,
+ "▁Fra": 12699,
+ "▁legit": 12700,
+ "▁illegal": 12701,
+ "клю": 12702,
+ "▁tort": 12703,
+ "▁pron": 12704,
+ "Fi": 12705,
+ "▁forg": 12706,
+ "export": 12707,
+ "▁Children": 12708,
+ "▁Abs": 12709,
+ "▁Send": 12710,
+ "▁discount": 12711,
+ "▁poster": 12712,
+ "ented": 12713,
+ "anim": 12714,
+ "verb": 12715,
+ "sto": 12716,
+ "▁Bible": 12717,
+ "pending": 12718,
+ "▁Phot": 12719,
+ "strap": 12720,
+ "ieron": 12721,
+ "PG": 12722,
+ "cular": 12723,
+ "crit": 12724,
+ "urd": 12725,
+ "ENO": 12726,
+ "▁northern": 12727,
+ "▁naturally": 12728,
+ "<'": 12729,
+ "weg": 12730,
+ "▁drunk": 12731,
+ "▁Dal": 12732,
+ "▁mouse": 12733,
+ "▁continuous": 12734,
+ "▁initially": 12735,
+ "agu": 12736,
+ "мпи": 12737,
+ "ANT": 12738,
+ "Div": 12739,
+ "▁recording": 12740,
+ "Bind": 12741,
+ "▁correctly": 12742,
+ "initial": 12743,
+ "▁Rights": 12744,
+ "▁debate": 12745,
+ "WRITE": 12746,
+ "built": 12747,
+ "▁permit": 12748,
+ "▁professionals": 12749,
+ "cv": 12750,
+ "▁DI": 12751,
+ "▁handed": 12752,
+ "▁Cu": 12753,
+ "▁Hospital": 12754,
+ "▁beskrevs": 12755,
+ "ней": 12756,
+ "ност": 12757,
+ "▁anxiety": 12758,
+ "▁heavily": 12759,
+ "▁Var": 12760,
+ "▁dispos": 12761,
+ "+\"": 12762,
+ "▁Ever": 12763,
+ "izon": 12764,
+ "▁operators": 12765,
+ "nego": 12766,
+ "▁Bry": 12767,
+ "▁votes": 12768,
+ "izione": 12769,
+ "▁рай": 12770,
+ "▁feat": 12771,
+ "▁western": 12772,
+ "▁confront": 12773,
+ "▁stronger": 12774,
+ "▁фа": 12775,
+ "stre": 12776,
+ "▁Valid": 12777,
+ "▁nad": 12778,
+ "▁checking": 12779,
+ "▁birds": 12780,
+ "▁Northern": 12781,
+ "▁intention": 12782,
+ "uce": 12783,
+ "▁covers": 12784,
+ "▁wondering": 12785,
+ "▁Optional": 12786,
+ "protocol": 12787,
+ "▁aggress": 12788,
+ "——": 12789,
+ "Vec": 12790,
+ "▁dates": 12791,
+ "quot": 12792,
+ "▁bom": 12793,
+ "▁scan": 12794,
+ "▁Item": 12795,
+ "▁Navy": 12796,
+ "▁Gran": 12797,
+ "▁everybody": 12798,
+ "▁unexpected": 12799,
+ "▁divor": 12800,
+ "▁ease": 12801,
+ "umbled": 12802,
+ "^+": 12803,
+ "cuss": 12804,
+ "▁pale": 12805,
+ "▁Inga": 12806,
+ "▁Broad": 12807,
+ "▁Medic": 12808,
+ "▁Roy": 12809,
+ "▁Inn": 12810,
+ "▁pens": 12811,
+ "PN": 12812,
+ ".:": 12813,
+ "▁principle": 12814,
+ "▁letting": 12815,
+ "▁conducted": 12816,
+ "FALSE": 12817,
+ "▁OS": 12818,
+ "Focus": 12819,
+ "▁measured": 12820,
+ "▁Democratic": 12821,
+ "High": 12822,
+ "▁pré": 12823,
+ "ennes": 12824,
+ "▁indicates": 12825,
+ "▁ending": 12826,
+ "▁Small": 12827,
+ "▁": 26345,
+ "olent": 26346,
+ "▁этого": 26347,
+ "▁Generic": 26348,
+ "▁*/,": 26349,
+ "▁combinations": 26350,
+ "▁rejo": 26351,
+ "спубли": 26352,
+ "capacity": 26353,
+ "▁traces": 26354,
+ "▁opacity": 26355,
+ "▁Official": 26356,
+ "icion": 26357,
+ "▁emotionally": 26358,
+ "▁Joel": 26359,
+ "ському": 26360,
+ "▁legendary": 26361,
+ "▁pam": 26362,
+ "▁También": 26363,
+ ".<": 26364,
+ "iba": 26365,
+ "midt": 26366,
+ "бом": 26367,
+ "▁ensuite": 26368,
+ "Authorization": 26369,
+ "Pag": 26370,
+ "▁helmet": 26371,
+ "▁territo": 26372,
+ "secondary": 26373,
+ "▁segunda": 26374,
+ "▁Wire": 26375,
+ "recated": 26376,
+ "▁invoked": 26377,
+ "▁ValueError": 26378,
+ "▁фо": 26379,
+ "ALIGN": 26380,
+ "CURRENT": 26381,
+ "\\+\\_\\": 26382,
+ "▁compilation": 26383,
+ "ær": 26384,
+ "▁Palmar": 26385,
+ "▁influences": 26386,
+ "/:": 26387,
+ "Mix": 26388,
+ "NOP": 26389,
+ "econom": 26390,
+ "▁tucked": 26391,
+ "▁});\r": 26392,
+ "ANK": 26393,
+ "reject": 26394,
+ "▁pension": 26395,
+ "▁generates": 26396,
+ "чё": 26397,
+ "▁incap": 26398,
+ "▁clicked": 26399,
+ "▁fus": 26400,
+ "ourses": 26401,
+ "▁Easter": 26402,
+ "%;": 26403,
+ "zin": 26404,
+ "▁obligations": 26405,
+ "▁Tips": 26406,
+ "};\r": 26407,
+ ".\"_": 26408,
+ "▁BSD": 26409,
+ "ática": 26410,
+ "▁expose": 26411,
+ "Pars": 26412,
+ "▁Amanda": 26413,
+ "куп": 26414,
+ "▁guessed": 26415,
+ "dsi": 26416,
+ "▁Leip": 26417,
+ "Broad": 26418,
+ "▁Hughes": 26419,
+ "ié": 26420,
+ "▁Wahl": 26421,
+ "▁formerly": 26422,
+ "Relative": 26423,
+ "▁Yu": 26424,
+ "▁Mountains": 26425,
+ "▁Enum": 26426,
+ "▁strang": 26427,
+ "_-": 26428,
+ "recht": 26429,
+ "viv": 26430,
+ "pause": 26431,
+ "▁Londres": 26432,
+ "▁elbow": 26433,
+ "▁Hawaii": 26434,
+ "▁Casino": 26435,
+ "Threshold": 26436,
+ "Units": 26437,
+ "Include": 26438,
+ "ито": 26439,
+ "asury": 26440,
+ "▁steht": 26441,
+ "▁damned": 26442,
+ "▁packets": 26443,
+ "▁Werk": 26444,
+ "▁elevator": 26445,
+ "iedad": 26446,
+ "govern": 26447,
+ "▁CONTRACT": 26448,
+ "mals": 26449,
+ "▁remem": 26450,
+ "▁entonces": 26451,
+ "▁vas": 26452,
+ "▁sympathy": 26453,
+ "▁befindet": 26454,
+ "incing": 26455,
+ "DataSet": 26456,
+ "▁additionally": 26457,
+ "▁musician": 26458,
+ "шего": 26459,
+ "▁listop": 26460,
+ ">\")": 26461,
+ "Printf": 26462,
+ "▁Felix": 26463,
+ "▁carved": 26464,
+ "▁nicely": 26465,
+ "гом": 26466,
+ "chap": 26467,
+ "▁Nieder": 26468,
+ "▁Lav": 26469,
+ "▁modifications": 26470,
+ "moment": 26471,
+ "▁balcon": 26472,
+ "▁dependency": 26473,
+ "CKET": 26474,
+ "▁vanished": 26475,
+ "▁fighters": 26476,
+ "▁zunächst": 26477,
+ "ioctl": 26478,
+ "▁defens": 26479,
+ "▁Nem": 26480,
+ "Utility": 26481,
+ "▁curv": 26482,
+ "▁DAMAGES": 26483,
+ "▁Rogers": 26484,
+ "▁gratitude": 26485,
+ "▁Denmark": 26486,
+ "рая": 26487,
+ "grpc": 26488,
+ "▁juni": 26489,
+ "▁октября": 26490,
+ "▁immense": 26491,
+ "▁prevented": 26492,
+ "▁foam": 26493,
+ "▁Extra": 26494,
+ "aimed": 26495,
+ "▁Criteria": 26496,
+ "▁Simply": 26497,
+ "boxes": 26498,
+ "▁Legend": 26499,
+ "▁Players": 26500,
+ "▁Mercedes": 26501,
+ "▁Branch": 26502,
+ "TERN": 26503,
+ "omena": 26504,
+ "▁incorporate": 26505,
+ "conde": 26506,
+ "▁Estado": 26507,
+ "▁wasted": 26508,
+ "▁complaining": 26509,
+ "▁warriors": 26510,
+ "oter": 26511,
+ "▁этом": 26512,
+ "▁conten": 26513,
+ "▁machinery": 26514,
+ "▁technological": 26515,
+ "▁TD": 26516,
+ "▁gras": 26517,
+ "▁minimize": 26518,
+ "▁Door": 26519,
+ "▁bzw": 26520,
+ "▁prac": 26521,
+ "TREE": 26522,
+ "▁Wing": 26523,
+ "▁Transaction": 26524,
+ "▁MVT": 26525,
+ "▁Klein": 26526,
+ "commons": 26527,
+ "▁}{": 26528,
+ "▁Heritage": 26529,
+ "▁fade": 26530,
+ "рок": 26531,
+ "setValue": 26532,
+ "▁Wallace": 26533,
+ "MX": 26534,
+ "▁ACT": 26535,
+ "▁footage": 26536,
+ "▁entstand": 26537,
+ "arga": 26538,
+ "▁nails": 26539,
+ "▁capitalism": 26540,
+ "▁Garc": 26541,
+ "▁suspension": 26542,
+ "ilis": 26543,
+ "▁Mov": 26544,
+ "uffled": 26545,
+ "Arc": 26546,
+ "▁Beautiful": 26547,
+ "WAY": 26548,
+ "Parallel": 26549,
+ "XXXX": 26550,
+ "diag": 26551,
+ "▁DT": 26552,
+ "mq": 26553,
+ "TextView": 26554,
+ "MLE": 26555,
+ "ennen": 26556,
+ "▁infected": 26557,
+ "▁therapist": 26558,
+ "INGS": 26559,
+ "▁cidade": 26560,
+ "ън": 26561,
+ "▁pdf": 26562,
+ "▁bump": 26563,
+ "CTX": 26564,
+ "▁INCLUDING": 26565,
+ "▁Gef": 26566,
+ "ENTIAL": 26567,
+ "▁handy": 26568,
+ "▁temporal": 26569,
+ "AtA": 26570,
+ "ISH": 26571,
+ "▁Pattern": 26572,
+ "▁lan": 26573,
+ "ependant": 26574,
+ "▁shining": 26575,
+ "idy": 26576,
+ "▁NT": 26577,
+ "▁Fran": 26578,
+ "▁nurses": 26579,
+ "▁betray": 26580,
+ "▁sensible": 26581,
+ "▁апреля": 26582,
+ "▁'[": 26583,
+ "▁thirteen": 26584,
+ ")}_{": 26585,
+ "▁Noah": 26586,
+ "INSERT": 26587,
+ "istically": 26588,
+ "▁Appendix": 26589,
+ "▁recher": 26590,
+ "Receiver": 26591,
+ "▁dernier": 26592,
+ "лла": 26593,
+ "лиза": 26594,
+ "▁Partido": 26595,
+ "▁maximal": 26596,
+ "snap": 26597,
+ "▁часть": 26598,
+ "STOP": 26599,
+ "▁ultra": 26600,
+ "▁développ": 26601,
+ "▁tegen": 26602,
+ "▁Чи": 26603,
+ "LIB": 26604,
+ "▁baseline": 26605,
+ "reload": 26606,
+ "▁Arbitro": 26607,
+ "▁kall": 26608,
+ "capture": 26609,
+ "Arm": 26610,
+ "quin": 26611,
+ "impse": 26612,
+ "zas": 26613,
+ "▁Cand": 26614,
+ "▁brains": 26615,
+ "▁hostile": 26616,
+ "▁marble": 26617,
+ "oons": 26618,
+ "▁Loss": 26619,
+ "MetaData": 26620,
+ "▁República": 26621,
+ "▁andra": 26622,
+ "oden": 26623,
+ "▁documented": 26624,
+ "▁Moses": 26625,
+ "odd": 26626,
+ "▁wax": 26627,
+ "usch": 26628,
+ "▁diagnosed": 26629,
+ "inkle": 26630,
+ "▁Xbox": 26631,
+ "▁seventy": 26632,
+ "cias": 26633,
+ "▁noviembre": 26634,
+ "Compute": 26635,
+ "});\r": 26636,
+ "▁Philippe": 26637,
+ "▁För": 26638,
+ "Leave": 26639,
+ "▁sage": 26640,
+ "▁unpre": 26641,
+ "▁Fortunately": 26642,
+ "▁apost": 26643,
+ "entities": 26644,
+ "▁ellos": 26645,
+ "authorized": 26646,
+ "GBT": 26647,
+ "▁insist": 26648,
+ "▁inspire": 26649,
+ "Mass": 26650,
+ "▁rôle": 26651,
+ "fee": 26652,
+ "ipart": 26653,
+ "цер": 26654,
+ "unate": 26655,
+ "▁CNN": 26656,
+ ":}": 26657,
+ "▁unhappy": 26658,
+ "▁imported": 26659,
+ "HIGH": 26660,
+ "rings": 26661,
+ "▁Instance": 26662,
+ "Bay": 26663,
+ "agles": 26664,
+ "mee": 26665,
+ "bery": 26666,
+ "▁Stories": 26667,
+ "▁Chase": 26668,
+ "▁carriage": 26669,
+ "▁misunder": 26670,
+ "▁imagin": 26671,
+ "pw": 26672,
+ "▁Meter": 26673,
+ "▁crowds": 26674,
+ "▁Fame": 26675,
+ "skill": 26676,
+ "▁comed": 26677,
+ "▁ranch": 26678,
+ "▁lacking": 26679,
+ "▁submar": 26680,
+ "iante": 26681,
+ "▁lanz": 26682,
+ "▁служ": 26683,
+ "-----------": 26684,
+ "▁obten": 26685,
+ "▁downstairs": 26686,
+ "YN": 26687,
+ "rotation": 26688,
+ "▁Jesse": 26689,
+ "$(\"#": 26690,
+ "▁puls": 26691,
+ "irling": 26692,
+ "▁Schaus": 26693,
+ "▁deployed": 26694,
+ "▁{}\",": 26695,
+ "▁Marvel": 26696,
+ "ENUM": 26697,
+ "▁Mathemat": 26698,
+ "▁nn": 26699,
+ "compet": 26700,
+ "ków": 26701,
+ "bil": 26702,
+ "Which": 26703,
+ "isine": 26704,
+ "▁rude": 26705,
+ "▁niveau": 26706,
+ "▁área": 26707,
+ "▁près": 26708,
+ "atis": 26709,
+ "▁[...]": 26710,
+ "fur": 26711,
+ "omm": 26712,
+ "packed": 26713,
+ "мене": 26714,
+ "scriptstyle": 26715,
+ "▁Ath": 26716,
+ "▁desp": 26717,
+ "eltemperaturen": 26718,
+ "▁talents": 26719,
+ "ocy": 26720,
+ "▁raises": 26721,
+ "LIMIT": 26722,
+ "▁editorial": 26723,
+ "▁Animal": 26724,
+ "drive": 26725,
+ "▁работа": 26726,
+ "bss": 26727,
+ "▁Sev": 26728,
+ "epoch": 26729,
+ "▁RC": 26730,
+ "UNUSED": 26731,
+ "▁mandatory": 26732,
+ "(?:": 26733,
+ "▁Bin": 26734,
+ "▁synthetic": 26735,
+ "▁gown": 26736,
+ "▁Dob": 26737,
+ "kap": 26738,
+ "▁harmon": 26739,
+ "▁liberty": 26740,
+ "▁Rice": 26741,
+ "▁prayers": 26742,
+ "▁mise": 26743,
+ "▁confusing": 26744,
+ "▁leap": 26745,
+ "▁arrives": 26746,
+ "kamp": 26747,
+ "▁thats": 26748,
+ "ACC": 26749,
+ "▁Parameters": 26750,
+ "▁одно": 26751,
+ "▁Bio": 26752,
+ "density": 26753,
+ "▁glimpse": 26754,
+ "FORE": 26755,
+ "▁Listen": 26756,
+ "Prev": 26757,
+ "}\\,\\": 26758,
+ "куль": 26759,
+ "▁SEC": 26760,
+ "▁explored": 26761,
+ "▁meantime": 26762,
+ "AIL": 26763,
+ "▁WP": 26764,
+ "▁raison": 26765,
+ "▁existe": 26766,
+ "▁lesser": 26767,
+ "▁Validate": 26768,
+ "▁caution": 26769,
+ "usta": 26770,
+ "heading": 26771,
+ "EFF": 26772,
+ ".'\"": 26773,
+ "▁Gilbert": 26774,
+ "▁limitation": 26775,
+ "▁retour": 26776,
+ "▁Commonwealth": 26777,
+ "▁gewann": 26778,
+ "▁miserable": 26779,
+ "▁networking": 26780,
+ "▁ottobre": 26781,
+ "▁Dise": 26782,
+ "edges": 26783,
+ "▁sede": 26784,
+ "вича": 26785,
+ "uniform": 26786,
+ "▁деятель": 26787,
+ "iros": 26788,
+ "▁desen": 26789,
+ "▁parc": 26790,
+ "▁Rico": 26791,
+ "Ns": 26792,
+ "guid": 26793,
+ "orio": 26794,
+ "avelength": 26795,
+ "▁Gle": 26796,
+ "inceton": 26797,
+ "Amaz": 26798,
+ "Construct": 26799,
+ "▁mx": 26800,
+ "▁Vern": 26801,
+ "▁Generation": 26802,
+ "Jack": 26803,
+ "romag": 26804,
+ "▁viagra": 26805,
+ "▁Peg": 26806,
+ "▁Updated": 26807,
+ "▁overlap": 26808,
+ "EventArgs": 26809,
+ "кро": 26810,
+ "▁*«": 26811,
+ "▁questioned": 26812,
+ "South": 26813,
+ "notice": 26814,
+ "▁permanently": 26815,
+ "lst": 26816,
+ "ficie": 26817,
+ "▁quella": 26818,
+ "▁colleges": 26819,
+ "▁disappointment": 26820,
+ "▁Luft": 26821,
+ "imgur": 26822,
+ "▁transitions": 26823,
+ "▁seller": 26824,
+ "▁июня": 26825,
+ "▁Og": 26826,
+ "▁ADD": 26827,
+ "▁Pays": 26828,
+ "COMMAND": 26829,
+ "grades": 26830,
+ "▁febbra": 26831,
+ "▁Cyr": 26832,
+ "▁febbraio": 26833,
+ "eti": 26834,
+ "▁arom": 26835,
+ "▁Claude": 26836,
+ "▁UEFA": 26837,
+ "▁живе": 26838,
+ "▁Victorian": 26839,
+ "keeping": 26840,
+ "ên": 26841,
+ "▁FIXME": 26842,
+ "itime": 26843,
+ "chestr": 26844,
+ "▁Samsung": 26845,
+ "▁doctrine": 26846,
+ "▁pear": 26847,
+ "▁Mediterranean": 26848,
+ "▁Ya": 26849,
+ "▁vault": 26850,
+ "▁Historic": 26851,
+ "▁sedan": 26852,
+ "▁heated": 26853,
+ "▁política": 26854,
+ "Proof": 26855,
+ ":{": 26856,
+ "fem": 26857,
+ "▁Frankfurt": 26858,
+ "pectives": 26859,
+ "MG": 26860,
+ "▁Eye": 26861,
+ "dai": 26862,
+ "▁reserves": 26863,
+ "NER": 26864,
+ "▁tobacco": 26865,
+ "▁fragments": 26866,
+ "icc": 26867,
+ "▁booth": 26868,
+ "▁cruise": 26869,
+ "▁Testament": 26870,
+ "cola": 26871,
+ "▁Leop": 26872,
+ "▁noon": 26873,
+ "▁terrified": 26874,
+ "vb": 26875,
+ "intel": 26876,
+ "alie": 26877,
+ "▁verification": 26878,
+ "yster": 26879,
+ "ADER": 26880,
+ "chied": 26881,
+ "▁datasets": 26882,
+ "▁зі": 26883,
+ "▁miem": 26884,
+ "ulates": 26885,
+ "▁uuid": 26886,
+ "▁Pictures": 26887,
+ "▁Brend": 26888,
+ "Billboard": 26889,
+ "▁stern": 26890,
+ "▁denom": 26891,
+ "▁accidents": 26892,
+ "сня": 26893,
+ "▁packing": 26894,
+ "ција": 26895,
+ "iblical": 26896,
+ "▁Так": 26897,
+ "▁whisk": 26898,
+ "▁luego": 26899,
+ "▁rectangle": 26900,
+ "▁hooks": 26901,
+ "▁neglect": 26902,
+ "▁sober": 26903,
+ "proposition": 26904,
+ "Multiple": 26905,
+ ":\",": 26906,
+ "▁bapt": 26907,
+ "Parts": 26908,
+ "▁Selection": 26909,
+ "▁Alpha": 26910,
+ "weights": 26911,
+ "hall": 26912,
+ "соб": 26913,
+ "▁lur": 26914,
+ "▁época": 26915,
+ "▁rested": 26916,
+ "ambigu": 26917,
+ "▁tastes": 26918,
+ "amazonaws": 26919,
+ "▁confess": 26920,
+ "▁diciembre": 26921,
+ "implement": 26922,
+ "▁absorption": 26923,
+ "Hal": 26924,
+ "LEAN": 26925,
+ "▁Zach": 26926,
+ "▁freeze": 26927,
+ "LBL": 26928,
+ "STM": 26929,
+ "▁calc": 26930,
+ "={()": 26931,
+ "=*/": 26932,
+ "▁bt": 26933,
+ "Reb": 26934,
+ "▁Wien": 26935,
+ "anska": 26936,
+ "▁surn": 26937,
+ "iative": 26938,
+ "▁invån": 26939,
+ "CY": 26940,
+ "▁là": 26941,
+ "amba": 26942,
+ "leen": 26943,
+ "wahl": 26944,
+ "▁functioning": 26945,
+ "ția": 26946,
+ "getContext": 26947,
+ "gart": 26948,
+ "▁обе": 26949,
+ "Pen": 26950,
+ "vik": 26951,
+ "Slider": 26952,
+ "▁Accept": 26953,
+ "Gap": 26954,
+ "▁Jorge": 26955,
+ "SIG": 26956,
+ "▁вос": 26957,
+ "▁голо": 26958,
+ "▁periodo": 26959,
+ "шта": 26960,
+ "▁patches": 26961,
+ "кої": 26962,
+ "äre": 26963,
+ "engono": 26964,
+ "lista": 26965,
+ "horn": 26966,
+ "▁Complex": 26967,
+ "Sent": 26968,
+ "trfs": 26969,
+ "▁convex": 26970,
+ "Generation": 26971,
+ "▁місце": 26972,
+ "compress": 26973,
+ "▁Sax": 26974,
+ "▁uid": 26975,
+ "▁Lebens": 26976,
+ "Completion": 26977,
+ "\\|_{": 26978,
+ "insky": 26979,
+ "▁schon": 26980,
+ "▁masters": 26981,
+ "independ": 26982,
+ "neys": 26983,
+ "▁lied": 26984,
+ "▁aspir": 26985,
+ "чні": 26986,
+ "▁breakdown": 26987,
+ "▁Harm": 26988,
+ "▁designing": 26989,
+ "hf": 26990,
+ "▁Angela": 26991,
+ "▁confer": 26992,
+ "▁partido": 26993,
+ "▁interference": 26994,
+ "mao": 26995,
+ "▁absorbed": 26996,
+ "▁Vall": 26997,
+ "ErrorCode": 26998,
+ "▁Publishing": 26999,
+ "vano": 27000,
+ "BITS": 27001,
+ "▁deer": 27002,
+ "▁Campaign": 27003,
+ "▁graz": 27004,
+ "CHANGE": 27005,
+ "▁feder": 27006,
+ "iffe": 27007,
+ "handed": 27008,
+ "cq": 27009,
+ "umbing": 27010,
+ "▁unre": 27011,
+ "▁siendo": 27012,
+ "▁simpler": 27013,
+ "why": 27014,
+ "arettes": 27015,
+ "anst": 27016,
+ "▁hass": 27017,
+ "▁Enterprise": 27018,
+ "▁mois": 27019,
+ "▁Fo": 27020,
+ "▁участ": 27021,
+ "ffen": 27022,
+ "▁MODULE": 27023,
+ "▁activated": 27024,
+ "▁internacional": 27025,
+ "▁Mittel": 27026,
+ "degree": 27027,
+ "▁откры": 27028,
+ "▁&(": 27029,
+ "getProperty": 27030,
+ "isz": 27031,
+ "cedure": 27032,
+ "▁enters": 27033,
+ "▁Sally": 27034,
+ "▁Train": 27035,
+ "▁logged": 27036,
+ "▁Rav": 27037,
+ "▁Avoid": 27038,
+ "▁Kaiser": 27039,
+ "▁expend": 27040,
+ "aphor": 27041,
+ "▁brass": 27042,
+ "▁melod": 27043,
+ "▁attitudes": 27044,
+ "*\"": 27045,
+ "Wall": 27046,
+ "▁owe": 27047,
+ "▁bamb": 27048,
+ "shader": 27049,
+ "cester": 27050,
+ "▁PP": 27051,
+ "▁migrations": 27052,
+ "entric": 27053,
+ "▁Setup": 27054,
+ "▁Artist": 27055,
+ "hre": 27056,
+ "▁polite": 27057,
+ "ahan": 27058,
+ "▁luglio": 27059,
+ "▁predecess": 27060,
+ "▁SIG": 27061,
+ "тів": 27062,
+ "▁RF": 27063,
+ "▁Dry": 27064,
+ "▁maker": 27065,
+ "шим": 27066,
+ "▁Sounds": 27067,
+ "▁implementing": 27068,
+ "▁ah": 27069,
+ "▁gev": 27070,
+ "▁duplicate": 27071,
+ "▁Logan": 27072,
+ "▁Grade": 27073,
+ "DUCT": 27074,
+ "íses": 27075,
+ "ért": 27076,
+ "▁nonsense": 27077,
+ "backup": 27078,
+ "Attachment": 27079,
+ "▁ecc": 27080,
+ "▁Squadron": 27081,
+ "learn": 27082,
+ "deprecated": 27083,
+ "▁Aub": 27084,
+ "▁Gol": 27085,
+ "▁overl": 27086,
+ "SERVICE": 27087,
+ "▁beautifully": 27088,
+ "REL": 27089,
+ "▁Gian": 27090,
+ "▁Papa": 27091,
+ "respond": 27092,
+ "▁Caribbean": 27093,
+ "rn": 27094,
+ "▁худож": 27095,
+ "Cfg": 27096,
+ "rai": 27097,
+ "▁sniff": 27098,
+ "tto": 27099,
+ "ологи": 27100,
+ "▁rb": 27101,
+ "▁incidents": 27102,
+ "▁duck": 27103,
+ "▁PROVIDED": 27104,
+ "Sources": 27105,
+ "▁Chelsea": 27106,
+ "▁tek": 27107,
+ "▁налази": 27108,
+ "▁pilots": 27109,
+ "тки": 27110,
+ "▁traded": 27111,
+ "▁Beijing": 27112,
+ "▁Gregory": 27113,
+ "scalar": 27114,
+ "▁inclined": 27115,
+ "▁Kamp": 27116,
+ "▁Marian": 27117,
+ "▁fierce": 27118,
+ "▁theft": 27119,
+ "ющих": 27120,
+ "▁Into": 27121,
+ "constraint": 27122,
+ "parentNode": 27123,
+ "idental": 27124,
+ "▁gouvernement": 27125,
+ "▁SND": 27126,
+ "▁Ruby": 27127,
+ "▁monaster": 27128,
+ "Records": 27129,
+ "▁Kab": 27130,
+ "▁Universe": 27131,
+ "▁approximate": 27132,
+ "Water": 27133,
+ "▁Physical": 27134,
+ "appers": 27135,
+ "oubtedly": 27136,
+ "ложен": 27137,
+ "▁towel": 27138,
+ "▁siblings": 27139,
+ "eph": 27140,
+ "icios": 27141,
+ "рами": 27142,
+ "▁outrage": 27143,
+ "▁també": 27144,
+ "SRC": 27145,
+ "телем": 27146,
+ "Vi": 27147,
+ ".');": 27148,
+ "LM": 27149,
+ "▁mitt": 27150,
+ "▁weed": 27151,
+ "▁crops": 27152,
+ "iman": 27153,
+ "Claim": 27154,
+ "insula": 27155,
+ "▁(“": 27156,
+ "▁Changes": 27157,
+ "▁invånare": 27158,
+ "again": 27159,
+ "▁cnt": 27160,
+ "▁Gaz": 27161,
+ "▁austral": 27162,
+ "overlay": 27163,
+ "▁Mechan": 27164,
+ "▁slammed": 27165,
+ "▁trailing": 27166,
+ "▁Biography": 27167,
+ "▁appealing": 27168,
+ "IVER": 27169,
+ "▁Ave": 27170,
+ "▁Plot": 27171,
+ "voj": 27172,
+ "▁sung": 27173,
+ "▁unos": 27174,
+ "Effects": 27175,
+ "vv": 27176,
+ "cook": 27177,
+ "Buttons": 27178,
+ "▁transm": 27179,
+ "ierto": 27180,
+ "CONTEXT": 27181,
+ "▁dignity": 27182,
+ "aired": 27183,
+ "javax": 27184,
+ "▁Alberto": 27185,
+ "▁Recently": 27186,
+ "▁facial": 27187,
+ "mathop": 27188,
+ "ało": 27189,
+ "вид": 27190,
+ "cott": 27191,
+ "Variables": 27192,
+ "▁Ran": 27193,
+ "▁bunk": 27194,
+ "amiliar": 27195,
+ "CAST": 27196,
+ "▁frü": 27197,
+ "VED": 27198,
+ "▁NOTICE": 27199,
+ "▁turno": 27200,
+ "validator": 27201,
+ "▁Portuguese": 27202,
+ "▁questioning": 27203,
+ "}})": 27204,
+ "▁lear": 27205,
+ "Xamarin": 27206,
+ "▁disadv": 27207,
+ "encoded": 27208,
+ "▁Kot": 27209,
+ "rated": 27210,
+ "▁Theory": 27211,
+ "cius": 27212,
+ "▁Darwin": 27213,
+ "ђе": 27214,
+ "▁décl": 27215,
+ "▁область": 27216,
+ "рович": 27217,
+ "▁mobility": 27218,
+ "VF": 27219,
+ "▁хи": 27220,
+ "until": 27221,
+ "▁barriers": 27222,
+ "gif": 27223,
+ "▁Roh": 27224,
+ "▁aging": 27225,
+ "▁Widget": 27226,
+ "olk": 27227,
+ "▁farms": 27228,
+ "Checker": 27229,
+ "Introduction": 27230,
+ "смо": 27231,
+ "▁Russians": 27232,
+ "naments": 27233,
+ "▁Insert": 27234,
+ "▁Whenever": 27235,
+ "erset": 27236,
+ "itori": 27237,
+ "▁Dort": 27238,
+ "▁costume": 27239,
+ "▁mathematical": 27240,
+ "▁Bast": 27241,
+ "▁nominated": 27242,
+ "▁restoration": 27243,
+ "posal": 27244,
+ "▁unfortunate": 27245,
+ "Ps": 27246,
+ "LIN": 27247,
+ "▁intact": 27248,
+ "▁provoc": 27249,
+ "▁située": 27250,
+ "▁ноября": 27251,
+ "ermo": 27252,
+ "▁fisher": 27253,
+ "гля": 27254,
+ "▁conting": 27255,
+ "▁Doug": 27256,
+ "\"?": 27257,
+ "▁Eva": 27258,
+ "▁tops": 27259,
+ "▁Remote": 27260,
+ "▁artwork": 27261,
+ "▁artillery": 27262,
+ "quick": 27263,
+ "▁Arabia": 27264,
+ "▁SDValue": 27265,
+ "▁Dakota": 27266,
+ "iated": 27267,
+ "▁Optim": 27268,
+ "buttons": 27269,
+ "▁cottage": 27270,
+ "▁wherein": 27271,
+ "▁tutorial": 27272,
+ "▁Scre": 27273,
+ "▁sweep": 27274,
+ "▁Coffee": 27275,
+ "})}": 27276,
+ "▁музы": 27277,
+ "hostname": 27278,
+ "▁Temp": 27279,
+ "▁Fut": 27280,
+ "respect": 27281,
+ "ocz": 27282,
+ "▁predomin": 27283,
+ "Indicator": 27284,
+ "encial": 27285,
+ "UMENT": 27286,
+ "▁SHALL": 27287,
+ "▁commanded": 27288,
+ "▁withdrawal": 27289,
+ "iour": 27290,
+ "REGION": 27291,
+ "sprintf": 27292,
+ "▁вме": 27293,
+ "▁Payment": 27294,
+ "▁Anim": 27295,
+ "publish": 27296,
+ "▁seeks": 27297,
+ "ouw": 27298,
+ "▁GM": 27299,
+ "rugu": 27300,
+ "ustain": 27301,
+ "▁))": 27302,
+ "▁consulting": 27303,
+ "▁Dialog": 27304,
+ "▁Lars": 27305,
+ "▁critique": 27306,
+ "▁circulation": 27307,
+ "▁landsc": 27308,
+ "managed": 27309,
+ "▁Craft": 27310,
+ "▁herman": 27311,
+ "afi": 27312,
+ "amy": 27313,
+ "▁discour": 27314,
+ "<>(": 27315,
+ "▁Steph": 27316,
+ "▁tolerance": 27317,
+ "typename": 27318,
+ "ventions": 27319,
+ "ział": 27320,
+ "стов": 27321,
+ "▁sticking": 27322,
+ "ASC": 27323,
+ "ISO": 27324,
+ "▁Spencer": 27325,
+ "▁Didn": 27326,
+ "gomery": 27327,
+ "imiter": 27328,
+ "dru": 27329,
+ "Clause": 27330,
+ "▁slides": 27331,
+ "###": 27332,
+ "▁Sugar": 27333,
+ "HY": 27334,
+ "▁эти": 27335,
+ "▁Edwards": 27336,
+ "▁cents": 27337,
+ "oya": 27338,
+ "serts": 27339,
+ "▁Hass": 27340,
+ "▁ingen": 27341,
+ "стри": 27342,
+ "▁saddle": 27343,
+ "solid": 27344,
+ "▁champions": 27345,
+ "-)": 27346,
+ "▁Slov": 27347,
+ "▁shiny": 27348,
+ "▁*)&": 27349,
+ "▁Define": 27350,
+ "če": 27351,
+ "▁scrut": 27352,
+ "onden": 27353,
+ "'\",": 27354,
+ "uffs": 27355,
+ "▁olymp": 27356,
+ "idential": 27357,
+ "wand": 27358,
+ "▁annually": 27359,
+ "▁Arkansas": 27360,
+ "▁saint": 27361,
+ "▁gleich": 27362,
+ "▁perfection": 27363,
+ ")>": 27364,
+ "▁shorts": 27365,
+ "▁justified": 27366,
+ "peated": 27367,
+ "packages": 27368,
+ "driven": 27369,
+ "▁Liberty": 27370,
+ "▁stripped": 27371,
+ "шение": 27372,
+ "▁fünf": 27373,
+ "▁ecosystem": 27374,
+ "ixa": 27375,
+ "▁Fresh": 27376,
+ "vart": 27377,
+ "▁treats": 27378,
+ "▁stance": 27379,
+ "чёт": 27380,
+ "▁pity": 27381,
+ "adém": 27382,
+ "▁окон": 27383,
+ "▁Chand": 27384,
+ "rab": 27385,
+ "вший": 27386,
+ "inski": 27387,
+ "▁continually": 27388,
+ "▁Daddy": 27389,
+ "▁nightmare": 27390,
+ "icional": 27391,
+ "▁efect": 27392,
+ "ueblo": 27393,
+ "▁lanç": 27394,
+ "▁Collections": 27395,
+ "due": 27396,
+ "ampton": 27397,
+ "▁memcpy": 27398,
+ "▁**(": 27399,
+ "issent": 27400,
+ "▁Insp": 27401,
+ "▁Glasgow": 27402,
+ "▁furono": 27403,
+ "▁kindness": 27404,
+ "Bi": 27405,
+ "▁competed": 27406,
+ "▁oak": 27407,
+ "Large": 27408,
+ "▁disgu": 27409,
+ "▁kings": 27410,
+ "тами": 27411,
+ "▁stuffed": 27412,
+ "▁hilar": 27413,
+ "published": 27414,
+ "▁stressed": 27415,
+ "▁Peak": 27416,
+ "▁loader": 27417,
+ "Keyboard": 27418,
+ "▁reconstruction": 27419,
+ "▁vod": 27420,
+ "▁dun": 27421,
+ "▁understands": 27422,
+ "tenant": 27423,
+ "▁chaque": 27424,
+ "▁prejud": 27425,
+ "utat": 27426,
+ "▁uso": 27427,
+ "▁Heavy": 27428,
+ "▁cuatro": 27429,
+ "▁sidewalk": 27430,
+ "▁Bug": 27431,
+ "▁månaden": 27432,
+ "geo": 27433,
+ "▁united": 27434,
+ "▁Files": 27435,
+ "▁Аль": 27436,
+ "▁rugby": 27437,
+ "▁financing": 27438,
+ "▁comply": 27439,
+ "": 27440,
+ "▁rushing": 27441,
+ "▁fen": 27442,
+ "mong": 27443,
+ "▁spé": 27444,
+ "▁presenting": 27445,
+ "INCLUDING": 27446,
+ "ěl": 27447,
+ "zeichnung": 27448,
+ "Backup": 27449,
+ "▁petit": 27450,
+ "▁allerg": 27451,
+ "нут": 27452,
+ "▁worrying": 27453,
+ "▁mamm": 27454,
+ "▁operand": 27455,
+ ":%.*]]": 27456,
+ "▁realise": 27457,
+ "Commands": 27458,
+ "▁Bew": 27459,
+ "▁assumes": 27460,
+ "▁Covid": 27461,
+ "▁quand": 27462,
+ "tyard": 27463,
+ "▁Mono": 27464,
+ "linked": 27465,
+ "MARK": 27466,
+ "Esp": 27467,
+ "▁blessing": 27468,
+ "▁eyebrows": 27469,
+ "▁NV": 27470,
+ "▁стру": 27471,
+ "▁modeling": 27472,
+ "▁greeted": 27473,
+ "Workspace": 27474,
+ "▁pedest": 27475,
+ "▁неза": 27476,
+ "lemagne": 27477,
+ "Statistics": 27478,
+ "▁aument": 27479,
+ "▁speeds": 27480,
+ "▁syndrome": 27481,
+ "CONNECT": 27482,
+ "zahl": 27483,
+ "verso": 27484,
+ "ército": 27485,
+ "▁astronom": 27486,
+ "▁aprile": 27487,
+ "žen": 27488,
+ "веро": 27489,
+ "draft": 27490,
+ "▁gioc": 27491,
+ "▁comport": 27492,
+ "▁variance": 27493,
+ "▁realizing": 27494,
+ "EDIT": 27495,
+ "олові": 27496,
+ "▁estar": 27497,
+ "▁sost": 27498,
+ "NORMAL": 27499,
+ "▁ó": 27500,
+ "▁Andr": 27501,
+ "ATTRIB": 27502,
+ "▁rede": 27503,
+ "▁toes": 27504,
+ "▁advances": 27505,
+ "▁Against": 27506,
+ "TOM": 27507,
+ "rss": 27508,
+ "MMMM": 27509,
+ "▁newest": 27510,
+ "▁VER": 27511,
+ "▁phrases": 27512,
+ "anter": 27513,
+ "Launch": 27514,
+ "▁chr": 27515,
+ "▁manufactured": 27516,
+ "$),": 27517,
+ "rollment": 27518,
+ "eston": 27519,
+ "▁peint": 27520,
+ "”)": 27521,
+ "endet": 27522,
+ "▁Hair": 27523,
+ "ivalent": 27524,
+ "▁upright": 27525,
+ "gren": 27526,
+ "anked": 27527,
+ "wright": 27528,
+ "▁mast": 27529,
+ "▁onChange": 27530,
+ "▁debris": 27531,
+ "▁grap": 27532,
+ "etry": 27533,
+ "▁(__": 27534,
+ "▁Commerce": 27535,
+ "BOX": 27536,
+ "Tax": 27537,
+ "▁отри": 27538,
+ "▁prevention": 27539,
+ "▁Feel": 27540,
+ "▁exotic": 27541,
+ "▁Bark": 27542,
+ "▁Steam": 27543,
+ "fon": 27544,
+ "olin": 27545,
+ "▁eliminated": 27546,
+ "▁bc": 27547,
+ "▁Cycl": 27548,
+ "▁$(\"#": 27549,
+ "▁Parl": 27550,
+ "manuel": 27551,
+ "ospher": 27552,
+ "WF": 27553,
+ "Analy": 27554,
+ "▁navig": 27555,
+ "▁renown": 27556,
+ "Rx": 27557,
+ "▁Walt": 27558,
+ "uffed": 27559,
+ "▁foster": 27560,
+ "$:": 27561,
+ "shore": 27562,
+ "Connector": 27563,
+ "фика": 27564,
+ "▁realization": 27565,
+ "Li": 27566,
+ "ctxt": 27567,
+ "ahoo": 27568,
+ "▁miracle": 27569,
+ "▁ET": 27570,
+ "▁GPS": 27571,
+ "▁Observable": 27572,
+ "▁hf": 27573,
+ "▁magnificent": 27574,
+ "него": 27575,
+ "BIN": 27576,
+ "▁Dorf": 27577,
+ "ieck": 27578,
+ "vee": 27579,
+ "▁Craw": 27580,
+ "/#": 27581,
+ "▁pci": 27582,
+ "ippet": 27583,
+ "▁Hillary": 27584,
+ "▁gir": 27585,
+ "▁rand": 27586,
+ "▁laying": 27587,
+ "▁Different": 27588,
+ "boys": 27589,
+ "virt": 27590,
+ "▁encryption": 27591,
+ "ász": 27592,
+ "пор": 27593,
+ "▁smelled": 27594,
+ "▁suscept": 27595,
+ "cluded": 27596,
+ "▁Carn": 27597,
+ "igten": 27598,
+ "▁Chuck": 27599,
+ "▁Provinc": 27600,
+ "▁perí": 27601,
+ "▁Marshal": 27602,
+ "мож": 27603,
+ "gfx": 27604,
+ "oshi": 27605,
+ "▁WHE": 27606,
+ "▁relaxation": 27607,
+ ",.": 27608,
+ "were": 27609,
+ "▁varieties": 27610,
+ "▁Won": 27611,
+ "▁gaps": 27612,
+ "▁stole": 27613,
+ "igua": 27614,
+ "ющие": 27615,
+ "▁Hampshire": 27616,
+ "phrase": 27617,
+ "▁película": 27618,
+ "Processing": 27619,
+ "▁initialization": 27620,
+ "oustic": 27621,
+ "▁Josef": 27622,
+ "icating": 27623,
+ "▁goodness": 27624,
+ "TES": 27625,
+ "▁cope": 27626,
+ "▁ignorance": 27627,
+ "▁Brist": 27628,
+ "▁paras": 27629,
+ "▁accidentally": 27630,
+ "▁tand": 27631,
+ "ittest": 27632,
+ "▁ули": 27633,
+ "▁shipped": 27634,
+ "▁ост": 27635,
+ "elseif": 27636,
+ "▁usize": 27637,
+ "horizontal": 27638,
+ "▁Carr": 27639,
+ "▁precip": 27640,
+ "roz": 27641,
+ "pathetic": 27642,
+ "rived": 27643,
+ "rok": 27644,
+ "▁digging": 27645,
+ "мом": 27646,
+ "▁Mull": 27647,
+ "▁XIII": 27648,
+ "▁peas": 27649,
+ "▁foul": 27650,
+ "▁travels": 27651,
+ "▁Ng": 27652,
+ "▁составе": 27653,
+ "Mont": 27654,
+ "arde": 27655,
+ "▁Stefan": 27656,
+ "^^^^": 27657,
+ "▁Kiss": 27658,
+ "▁Ek": 27659,
+ "▁oktober": 27660,
+ "▁memorable": 27661,
+ "')).": 27662,
+ "▁Vision": 27663,
+ "▁Nina": 27664,
+ "▁Solar": 27665,
+ "▁highlighted": 27666,
+ "▁memo": 27667,
+ "meisterschaft": 27668,
+ "sidebar": 27669,
+ "SEE": 27670,
+ "▁Nevada": 27671,
+ "Da": 27672,
+ "▁drawer": 27673,
+ "astically": 27674,
+ "elde": 27675,
+ "scribed": 27676,
+ "▁priests": 27677,
+ "▁hommes": 27678,
+ "▁instructor": 27679,
+ "клад": 27680,
+ "▁spett": 27681,
+ "\\-": 27682,
+ "▁мира": 27683,
+ "▁Looks": 27684,
+ "▁sleeve": 27685,
+ "▁strongest": 27686,
+ "▁tête": 27687,
+ "▁Nicole": 27688,
+ "imper": 27689,
+ "нача": 27690,
+ "ipper": 27691,
+ "▁inwon": 27692,
+ "ilers": 27693,
+ "▁Deputy": 27694,
+ "oge": 27695,
+ "▁depressed": 27696,
+ "▁arte": 27697,
+ "▁combining": 27698,
+ "LAST": 27699,
+ "inted": 27700,
+ "▁Average": 27701,
+ "▁pollution": 27702,
+ "▁Phillips": 27703,
+ "▁WM": 27704,
+ "}}}\\": 27705,
+ "Added": 27706,
+ "▁peripher": 27707,
+ "Creation": 27708,
+ "▁italien": 27709,
+ "▁Choice": 27710,
+ "▁EXPRESS": 27711,
+ "▁Struct": 27712,
+ "ysz": 27713,
+ "Resize": 27714,
+ "ARGS": 27715,
+ "▁repo": 27716,
+ "▁чтобы": 27717,
+ "▁pref": 27718,
+ "▁earthqu": 27719,
+ "▁Мекси": 27720,
+ "▁Finale": 27721,
+ "▁hecho": 27722,
+ "requests": 27723,
+ "Cut": 27724,
+ "▁deserved": 27725,
+ "гово": 27726,
+ "▁Recent": 27727,
+ "▁дивизи": 27728,
+ "▁supportive": 27729,
+ "прави": 27730,
+ "▁irrelevant": 27731,
+ "'\r": 27732,
+ "▁ctrl": 27733,
+ "▁Deal": 27734,
+ "izada": 27735,
+ "uo": 27736,
+ "▁nort": 27737,
+ "geometry": 27738,
+ "▁Individual": 27739,
+ "ereg": 27740,
+ "▁приня": 27741,
+ "cref": 27742,
+ "══": 27743,
+ "▁comerc": 27744,
+ "=_": 27745,
+ "bund": 27746,
+ "тах": 27747,
+ "ilen": 27748,
+ "чита": 27749,
+ "▁corporation": 27750,
+ "esz": 27751,
+ "▁==>": 27752,
+ "ablish": 27753,
+ "Apr": 27754,
+ "▁ripped": 27755,
+ "Vars": 27756,
+ "stret": 27757,
+ "▁Francesco": 27758,
+ "NaN": 27759,
+ "▁anytime": 27760,
+ "▁automated": 27761,
+ "ostream": 27762,
+ "▁drawings": 27763,
+ "▁enhancement": 27764,
+ "okrat": 27765,
+ "▁Issue": 27766,
+ "вра": 27767,
+ "Currency": 27768,
+ "▁wyn": 27769,
+ "izarre": 27770,
+ "ético": 27771,
+ "multiple": 27772,
+ "▁Rate": 27773,
+ "▁Ich": 27774,
+ "▁Auss": 27775,
+ "▁Former": 27776,
+ "Curve": 27777,
+ "▁marvel": 27778,
+ "attro": 27779,
+ "▁сп": 27780,
+ "BOOL": 27781,
+ "сия": 27782,
+ "gold": 27783,
+ "▁Nintendo": 27784,
+ "▁Salvador": 27785,
+ "▁Solution": 27786,
+ "ADC": 27787,
+ "бора": 27788,
+ "▁Bennett": 27789,
+ "▁FR": 27790,
+ "▁pueden": 27791,
+ "patient": 27792,
+ "▁PG": 27793,
+ "▁Jin": 27794,
+ "▁crashed": 27795,
+ "▁denen": 27796,
+ "▁Sample": 27797,
+ "▁Quebec": 27798,
+ "itories": 27799,
+ "▁blinked": 27800,
+ "▁lion": 27801,
+ "▁voce": 27802,
+ "▁Impact": 27803,
+ "▁Mau": 27804,
+ "▁Nie": 27805,
+ "▁lob": 27806,
+ "▁две": 27807,
+ "orneys": 27808,
+ "▁coastal": 27809,
+ "▁sensors": 27810,
+ "▁XII": 27811,
+ "▁illusion": 27812,
+ "oji": 27813,
+ "▁INC": 27814,
+ "▁Duncan": 27815,
+ "yk": 27816,
+ "▁affecting": 27817,
+ "pul": 27818,
+ "▁Napoleon": 27819,
+ "▁акаде": 27820,
+ "▁compt": 27821,
+ "▁profitable": 27822,
+ "loe": 27823,
+ "▁deuxième": 27824,
+ "▁WC": 27825,
+ "▁viable": 27826,
+ "▁Drug": 27827,
+ "TextBox": 27828,
+ "▁luminos": 27829,
+ "auté": 27830,
+ "yc": 27831,
+ "ště": 27832,
+ "▁affiliates": 27833,
+ "ilda": 27834,
+ "conduct": 27835,
+ "▁ebenfalls": 27836,
+ "▁AMD": 27837,
+ "▁Monitor": 27838,
+ "▁Companies": 27839,
+ "▁corrected": 27840,
+ "äck": 27841,
+ "SYSTEM": 27842,
+ "otherapy": 27843,
+ "▁перед": 27844,
+ "▁blues": 27845,
+ "atisf": 27846,
+ "although": 27847,
+ "rost": 27848,
+ "SCAN": 27849,
+ "▁RAM": 27850,
+ "ціональ": 27851,
+ "▁vendors": 27852,
+ "▁customs": 27853,
+ "▁activate": 27854,
+ "▁blogs": 27855,
+ "▁brace": 27856,
+ "▁strat": 27857,
+ "anje": 27858,
+ "щё": 27859,
+ "▁tide": 27860,
+ "▁Brigade": 27861,
+ "getOperand": 27862,
+ "▁aliment": 27863,
+ "▁achievements": 27864,
+ "▁suspicion": 27865,
+ "▁touchdown": 27866,
+ "broad": 27867,
+ "iore": 27868,
+ "Comparison": 27869,
+ "▁mum": 27870,
+ "English": 27871,
+ "▁Picture": 27872,
+ "▁Mouse": 27873,
+ "amd": 27874,
+ "▁[`": 27875,
+ "▁denomin": 27876,
+ "▁Aleks": 27877,
+ "▁prevents": 27878,
+ "ób": 27879,
+ "fed": 27880,
+ "▁Pray": 27881,
+ "▁shine": 27882,
+ "▁clutch": 27883,
+ "mux": 27884,
+ "Appro": 27885,
+ "▁notably": 27886,
+ "chio": 27887,
+ "nage": 27888,
+ "HAS": 27889,
+ "▁')": 27890,
+ "▁Miche": 27891,
+ "tg": 27892,
+ "::~": 27893,
+ "▁amely": 27894,
+ "▁rodz": 27895,
+ "zs": 27896,
+ "trait": 27897,
+ "▁klass": 27898,
+ "fö": 27899,
+ "▁destac": 27900,
+ "▁Clara": 27901,
+ "frequency": 27902,
+ "▁Git": 27903,
+ "▁поль": 27904,
+ "▁frequencies": 27905,
+ "▁febrero": 27906,
+ "▁stumbled": 27907,
+ "кою": 27908,
+ "▁Names": 27909,
+ "▁Flight": 27910,
+ "▁prey": 27911,
+ "▁medio": 27912,
+ "▁VAR": 27913,
+ "▁Float": 27914,
+ "▁Ernest": 27915,
+ "▁Marcatori": 27916,
+ "oport": 27917,
+ "▁cancellation": 27918,
+ "▁Bryan": 27919,
+ "————": 27920,
+ "Luc": 27921,
+ "▁libre": 27922,
+ "▁título": 27923,
+ "*>": 27924,
+ "▁Sandy": 27925,
+ "▁Marina": 27926,
+ "Been": 27927,
+ "▁wal": 27928,
+ "▁Kultur": 27929,
+ "▁explode": 27930,
+ "▁limiting": 27931,
+ "▁presumably": 27932,
+ "▁pb": 27933,
+ "▁Merc": 27934,
+ "▁реки": 27935,
+ "learning": 27936,
+ "Catalog": 27937,
+ "▁Census": 27938,
+ "lte": 27939,
+ "▁NET": 27940,
+ "raising": 27941,
+ "ське": 27942,
+ "staff": 27943,
+ "▁Quinn": 27944,
+ "▁memorial": 27945,
+ "пня": 27946,
+ "▁cuenta": 27947,
+ "▁XI": 27948,
+ "lbl": 27949,
+ "▁varies": 27950,
+ "▁fluctuations": 27951,
+ "▁долж": 27952,
+ "▁особи": 27953,
+ "▁warehouse": 27954,
+ "However": 27955,
+ "▁corrections": 27956,
+ "dhd": 27957,
+ "▁fals": 27958,
+ "▁controversy": 27959,
+ "▁curse": 27960,
+ "▁télé": 27961,
+ "řed": 27962,
+ "▁AU": 27963,
+ "▁тор": 27964,
+ "▁crít": 27965,
+ "idan": 27966,
+ "iliary": 27967,
+ "▁Panel": 27968,
+ "cule": 27969,
+ "▁Poor": 27970,
+ "▁BA": 27971,
+ "▁ignorant": 27972,
+ "èmes": 27973,
+ "▁aesthetic": 27974,
+ "Linked": 27975,
+ "getInt": 27976,
+ "Unicode": 27977,
+ "[@": 27978,
+ "▁Zent": 27979,
+ "Manifest": 27980,
+ "▁vars": 27981,
+ "PB": 27982,
+ "▁ву": 27983,
+ "▁Describe": 27984,
+ "▁Anything": 27985,
+ "oirs": 27986,
+ "▁socks": 27987,
+ "▁imped": 27988,
+ "▁neue": 27989,
+ "▁dispers": 27990,
+ "Collect": 27991,
+ "filer": 27992,
+ "▁Frau": 27993,
+ "▁Hockey": 27994,
+ "▁teens": 27995,
+ "▁Roberto": 27996,
+ "lauf": 27997,
+ "вать": 27998,
+ "▁ско": 27999,
+ "isArray": 28000,
+ "▁teenager": 28001,
+ "Built": 28002,
+ "▁loudly": 28003,
+ "Capacity": 28004,
+ "▁adventures": 28005,
+ "▁Molly": 28006,
+ "recogn": 28007,
+ "bars": 28008,
+ "▁Lor": 28009,
+ "▁può": 28010,
+ "▁mong": 28011,
+ "inement": 28012,
+ "Assignment": 28013,
+ "▁diz": 28014,
+ "lessness": 28015,
+ "▁Halloween": 28016,
+ "▁bitmap": 28017,
+ "Rom": 28018,
+ "нар": 28019,
+ "▁rebel": 28020,
+ "▁radial": 28021,
+ "measure": 28022,
+ "nit": 28023,
+ "▁Assume": 28024,
+ "▁assignments": 28025,
+ "▁Isn": 28026,
+ "▁altre": 28027,
+ "ßer": 28028,
+ "наль": 28029,
+ "▁flies": 28030,
+ "▁droit": 28031,
+ "▁thickness": 28032,
+ "▁enjo": 28033,
+ "▁dwell": 28034,
+ "▁homosexual": 28035,
+ "▁eval": 28036,
+ "$_{": 28037,
+ "asia": 28038,
+ "▁philos": 28039,
+ "getCurrent": 28040,
+ "▁veterans": 28041,
+ "▁Berkeley": 28042,
+ "▁wildlife": 28043,
+ "Cop": 28044,
+ "vern": 28045,
+ "▁Ú": 28046,
+ "tos": 28047,
+ "▁Led": 28048,
+ "▁keywords": 28049,
+ "▁medications": 28050,
+ "neum": 28051,
+ "▁jamais": 28052,
+ "▁Buc": 28053,
+ "▁PD": 28054,
+ "▁Statement": 28055,
+ "▁PI": 28056,
+ "▁Jackie": 28057,
+ "▁ordin": 28058,
+ "▁kör": 28059,
+ "enze": 28060,
+ "▁utilized": 28061,
+ "áct": 28062,
+ "azed": 28063,
+ "▁severely": 28064,
+ "▁även": 28065,
+ "▁libro": 28066,
+ "▁Eu": 28067,
+ "äst": 28068,
+ "PART": 28069,
+ "▁Butler": 28070,
+ "▁puzzle": 28071,
+ "Fall": 28072,
+ "Country": 28073,
+ "pfn": 28074,
+ "▁україн": 28075,
+ "▁Orchestra": 28076,
+ "▁alto": 28077,
+ "▁ancora": 28078,
+ "▁decomposition": 28079,
+ "▁م": 28080,
+ "▁appetite": 28081,
+ "adu": 28082,
+ "▁THAT": 28083,
+ "▁comenz": 28084,
+ "mina": 28085,
+ "▁initiated": 28086,
+ "▁Tat": 28087,
+ "▁sometime": 28088,
+ "rek": 28089,
+ "bread": 28090,
+ "▁Statistics": 28091,
+ "▁Cob": 28092,
+ "Follow": 28093,
+ "▁geometric": 28094,
+ "шла": 28095,
+ "▁proceedings": 28096,
+ "Dlg": 28097,
+ "seven": 28098,
+ "▁[-": 28099,
+ "▁Buffalo": 28100,
+ "▁blacks": 28101,
+ "▁sov": 28102,
+ "▁custody": 28103,
+ "▁ras": 28104,
+ "▁tattoo": 28105,
+ "öffentlicht": 28106,
+ "Blo": 28107,
+ "Austral": 28108,
+ "▁recuper": 28109,
+ "лев": 28110,
+ "▁bem": 28111,
+ "▁thou": 28112,
+ "oriented": 28113,
+ "vir": 28114,
+ "▁colony": 28115,
+ "▁Stanford": 28116,
+ "Absolute": 28117,
+ "adrat": 28118,
+ "▁Situ": 28119,
+ "▁souvent": 28120,
+ "EXEC": 28121,
+ "▁mű": 28122,
+ "▁apartments": 28123,
+ "▁случа": 28124,
+ "▁ano": 28125,
+ "WINDO": 28126,
+ "acci": 28127,
+ "▁Lau": 28128,
+ "court": 28129,
+ "▁manifold": 28130,
+ "▁coalition": 28131,
+ "▁XIV": 28132,
+ "Attrib": 28133,
+ "ascade": 28134,
+ "▁wheat": 28135,
+ "▁strengths": 28136,
+ "FREE": 28137,
+ "EMPTY": 28138,
+ "▁hey": 28139,
+ "ascular": 28140,
+ "▁plasma": 28141,
+ "▁bob": 28142,
+ "Separator": 28143,
+ "=\"${": 28144,
+ "▁Zag": 28145,
+ "▁projet": 28146,
+ "▁smoothly": 28147,
+ "SEQU": 28148,
+ "analy": 28149,
+ "attachment": 28150,
+ "▁ES": 28151,
+ "▁popped": 28152,
+ "ős": 28153,
+ "tom": 28154,
+ "▁són": 28155,
+ "▁rott": 28156,
+ "Utilities": 28157,
+ "hadoop": 28158,
+ "▁sotto": 28159,
+ "autor": 28160,
+ "▁Georges": 28161,
+ "▁který": 28162,
+ "▁gruppo": 28163,
+ "▁когда": 28164,
+ "▁меда": 28165,
+ "▁instrumental": 28166,
+ "▁Writer": 28167,
+ "▁setTimeout": 28168,
+ "ikk": 28169,
+ "▁Dopo": 28170,
+ "]);\r": 28171,
+ "▁practicing": 28172,
+ "▁Ronald": 28173,
+ "▁уби": 28174,
+ "▁agrees": 28175,
+ "▁denoted": 28176,
+ "ismiss": 28177,
+ "▁interviewed": 28178,
+ "templates": 28179,
+ "ři": 28180,
+ "administr": 28181,
+ "▁Butter": 28182,
+ "▁XVII": 28183,
+ "▁positioned": 28184,
+ "▁Fourth": 28185,
+ "▁overwhelmed": 28186,
+ "▁Regular": 28187,
+ "▁reprezent": 28188,
+ "кономи": 28189,
+ "▁expects": 28190,
+ "Indices": 28191,
+ "▁marijuana": 28192,
+ "▁zaj": 28193,
+ "▁Bren": 28194,
+ "▁begg": 28195,
+ "▁nahm": 28196,
+ "▁interrog": 28197,
+ "тие": 28198,
+ "▁Bun": 28199,
+ "▁серед": 28200,
+ "▁shelves": 28201,
+ "▁которых": 28202,
+ "▁Frauen": 28203,
+ "▁Sergeant": 28204,
+ "▁успе": 28205,
+ "matched": 28206,
+ "▁donne": 28207,
+ "▁touches": 28208,
+ "abort": 28209,
+ "▁vale": 28210,
+ "▁institutional": 28211,
+ "▁Mons": 28212,
+ "▁ambitious": 28213,
+ "▁nonetheless": 28214,
+ "jd": 28215,
+ "пей": 28216,
+ "▁backpack": 28217,
+ "dao": 28218,
+ "вия": 28219,
+ "▁surroundings": 28220,
+ "|_{": 28221,
+ "▁gegründ": 28222,
+ "disp": 28223,
+ "▁moisture": 28224,
+ "▁wyd": 28225,
+ "▁traders": 28226,
+ "▁Erst": 28227,
+ "▁Galaxy": 28228,
+ "▁воло": 28229,
+ "▁Peru": 28230,
+ "▁priorities": 28231,
+ "▁pronounced": 28232,
+ "▁CBS": 28233,
+ "▁Palm": 28234,
+ "▁expans": 28235,
+ "▁energet": 28236,
+ "▁Condition": 28237,
+ "▁Sver": 28238,
+ "nested": 28239,
+ "▁февраля": 28240,
+ "hero": 28241,
+ "▁коло": 28242,
+ "▁Films": 28243,
+ "Bon": 28244,
+ "éal": 28245,
+ "ployed": 28246,
+ "trained": 28247,
+ "▁első": 28248,
+ "▁lust": 28249,
+ "atinum": 28250,
+ "oyle": 28251,
+ "▁Jet": 28252,
+ "ждения": 28253,
+ "▁surveys": 28254,
+ "bee": 28255,
+ "workers": 28256,
+ "records": 28257,
+ "calendar": 28258,
+ "bbing": 28259,
+ "regation": 28260,
+ "dashboard": 28261,
+ "King": 28262,
+ "▁vista": 28263,
+ "▁depicted": 28264,
+ "▁occurring": 28265,
+ "▁офи": 28266,
+ "▁sandwich": 28267,
+ "rcu": 28268,
+ "kern": 28269,
+ "▁minut": 28270,
+ "▁смер": 28271,
+ "▁td": 28272,
+ "solete": 28273,
+ "Complex": 28274,
+ "▁tunn": 28275,
+ "▁scarc": 28276,
+ "stead": 28277,
+ "▁Fail": 28278,
+ "▁Rs": 28279,
+ "▁trails": 28280,
+ "kem": 28281,
+ "▁Romans": 28282,
+ "ativity": 28283,
+ "Previous": 28284,
+ "▁depress": 28285,
+ "▁resigned": 28286,
+ "getDefault": 28287,
+ "▁Tibet": 28288,
+ "▁Franco": 28289,
+ "\")));": 28290,
+ "▁injection": 28291,
+ "removed": 28292,
+ "▁praised": 28293,
+ "▁Asc": 28294,
+ "erase": 28295,
+ "▁commissioned": 28296,
+ "MAIL": 28297,
+ "▁Boh": 28298,
+ "Poly": 28299,
+ "▁cinq": 28300,
+ "▁Above": 28301,
+ "▁Joshua": 28302,
+ "ZERO": 28303,
+ "▁summit": 28304,
+ "▁Urs": 28305,
+ "▁curl": 28306,
+ "▁visa": 28307,
+ "▁resur": 28308,
+ "={'": 28309,
+ "feat": 28310,
+ "▁absorb": 28311,
+ "▁planets": 28312,
+ "▁princess": 28313,
+ "▁Jahrhunderts": 28314,
+ "xp": 28315,
+ "▁NBC": 28316,
+ "▁коми": 28317,
+ "▁FUN": 28318,
+ "▁neuen": 28319,
+ "▁déjà": 28320,
+ "▁Oz": 28321,
+ "bben": 28322,
+ "VIDEO": 28323,
+ "▁ejempl": 28324,
+ "▁considers": 28325,
+ "atri": 28326,
+ "▁arrog": 28327,
+ "ioso": 28328,
+ "▁hace": 28329,
+ "▁contacted": 28330,
+ "▁unple": 28331,
+ "▁sponsored": 28332,
+ "▁trainer": 28333,
+ "sbi": 28334,
+ "▁занима": 28335,
+ "Criterion": 28336,
+ "ното": 28337,
+ "scheme": 28338,
+ "ennial": 28339,
+ "perform": 28340,
+ "▁fixing": 28341,
+ "▁постро": 28342,
+ "arb": 28343,
+ "EXIT": 28344,
+ "▁café": 28345,
+ "ituted": 28346,
+ "riages": 28347,
+ "Tur": 28348,
+ "▁haber": 28349,
+ "elasticsearch": 28350,
+ "▁ал": 28351,
+ "rh": 28352,
+ "▁voll": 28353,
+ "CLU": 28354,
+ "Mil": 28355,
+ "▁membres": 28356,
+ "▁remarked": 28357,
+ "вана": 28358,
+ "=\"_": 28359,
+ "Less": 28360,
+ "(\"\");": 28361,
+ "▁Yale": 28362,
+ "berries": 28363,
+ "▁releasing": 28364,
+ "▁imports": 28365,
+ "idea": 28366,
+ "▁(+": 28367,
+ "▁arqu": 28368,
+ "ificación": 28369,
+ "▁пара": 28370,
+ "▁Rangers": 28371,
+ "Mic": 28372,
+ "▁nederbörd": 28373,
+ "▁imaginary": 28374,
+ "▁specialists": 28375,
+ "▁hoof": 28376,
+ "Modules": 28377,
+ "▁sadly": 28378,
+ "ceil": 28379,
+ "TabIndex": 28380,
+ "ationale": 28381,
+ "▁Partner": 28382,
+ "tbody": 28383,
+ "▁leverage": 28384,
+ "DN": 28385,
+ "▁Prec": 28386,
+ "▁Sé": 28387,
+ "▁Mam": 28388,
+ "▁afin": 28389,
+ "isValid": 28390,
+ "Pse": 28391,
+ "▁сторо": 28392,
+ "▁chopped": 28393,
+ "▁Minor": 28394,
+ "▁dabei": 28395,
+ "David": 28396,
+ "ussia": 28397,
+ "▁деревня": 28398,
+ "▁Identity": 28399,
+ "▁LGBT": 28400,
+ "ције": 28401,
+ "▁Orts": 28402,
+ "▁parti": 28403,
+ "▁Bachelor": 28404,
+ "uga": 28405,
+ "▁OPT": 28406,
+ "▁Seth": 28407,
+ "▁LIABLE": 28408,
+ "▁inaugur": 28409,
+ "▁Shanghai": 28410,
+ "▁relaxing": 28411,
+ "циона": 28412,
+ "\"%": 28413,
+ "▁obey": 28414,
+ "▁Airlines": 28415,
+ "Links": 28416,
+ "▁Celt": 28417,
+ "▁Admin": 28418,
+ "agation": 28419,
+ "▁worries": 28420,
+ "INTE": 28421,
+ "arith": 28422,
+ "Fatalf": 28423,
+ "]])": 28424,
+ "colm": 28425,
+ "▁archae": 28426,
+ "▁brushed": 28427,
+ "▁tät": 28428,
+ "▁structured": 28429,
+ "тии": 28430,
+ "▁homem": 28431,
+ "[:,": 28432,
+ "▁navy": 28433,
+ "getKey": 28434,
+ "powered": 28435,
+ "▁sucked": 28436,
+ "▁zomb": 28437,
+ "issant": 28438,
+ "▁Might": 28439,
+ "▁Pull": 28440,
+ "rir": 28441,
+ "▁пі": 28442,
+ "▁seas": 28443,
+ "▁Wrest": 28444,
+ "▁tense": 28445,
+ "▁atm": 28446,
+ "▁havet": 28447,
+ "▁pierws": 28448,
+ "▁tragic": 28449,
+ "▁Diff": 28450,
+ "▁confidential": 28451,
+ "successful": 28452,
+ "ęż": 28453,
+ "▁Chain": 28454,
+ "▁Kenya": 28455,
+ "Choice": 28456,
+ "ocur": 28457,
+ "aniu": 28458,
+ "▁consultant": 28459,
+ "▁Advis": 28460,
+ "Lif": 28461,
+ "▁Lors": 28462,
+ "avorite": 28463,
+ "▁utilizing": 28464,
+ "▁vintage": 28465,
+ "Matcher": 28466,
+ "▁membre": 28467,
+ "▁Expect": 28468,
+ "▁tracing": 28469,
+ "nog": 28470,
+ "▁dej": 28471,
+ "▁уче": 28472,
+ "▁loops": 28473,
+ "▁onclick": 28474,
+ "▁GPU": 28475,
+ "▁Albums": 28476,
+ "▁Archives": 28477,
+ "вата": 28478,
+ "▁stove": 28479,
+ "шли": 28480,
+ "ancies": 28481,
+ "▁gemeente": 28482,
+ "mob": 28483,
+ "PDF": 28484,
+ "eso": 28485,
+ "▁vég": 28486,
+ "Resolve": 28487,
+ "▁teaches": 28488,
+ "ложе": 28489,
+ "▁ство": 28490,
+ "▁Одна": 28491,
+ "▁fid": 28492,
+ "Something": 28493,
+ "▁nebo": 28494,
+ "▁Valentine": 28495,
+ "rowning": 28496,
+ "▁але": 28497,
+ "awi": 28498,
+ "ishi": 28499,
+ "▁SPI": 28500,
+ "▁spel": 28501,
+ "▁біль": 28502,
+ "▁participant": 28503,
+ "▁Ned": 28504,
+ "▁Gast": 28505,
+ "▁blond": 28506,
+ "▁saves": 28507,
+ "colored": 28508,
+ "▁ACTION": 28509,
+ "▁Politiker": 28510,
+ "}$)": 28511,
+ "▁Dum": 28512,
+ "dentry": 28513,
+ "Student": 28514,
+ "▁~=": 28515,
+ "loads": 28516,
+ "▁Foster": 28517,
+ "一个": 28518,
+ "▁PK": 28519,
+ "▁SB": 28520,
+ "▁Hern": 28521,
+ "▁Exhib": 28522,
+ "Listeners": 28523,
+ "Sun": 28524,
+ "plac": 28525,
+ "▁Bever": 28526,
+ "▁incluy": 28527,
+ "▁dc": 28528,
+ "argc": 28529,
+ "▁ged": 28530,
+ "спа": 28531,
+ "▁Formula": 28532,
+ "▁сем": 28533,
+ "▁empt": 28534,
+ "unregister": 28535,
+ "▁Queensland": 28536,
+ "ández": 28537,
+ "otive": 28538,
+ "▁alley": 28539,
+ "▁Democrat": 28540,
+ "▁travail": 28541,
+ "▁$,": 28542,
+ "RP": 28543,
+ "рое": 28544,
+ "personal": 28545,
+ "▁période": 28546,
+ "HOME": 28547,
+ "omes": 28548,
+ "▁recognised": 28549,
+ "heng": 28550,
+ "▁Jung": 28551,
+ "▁Roland": 28552,
+ "▁convicted": 28553,
+ "Locked": 28554,
+ "▁mari": 28555,
+ "▁Luxem": 28556,
+ "referto": 28557,
+ "Deleted": 28558,
+ "intent": 28559,
+ "▁Staats": 28560,
+ "▁області": 28561,
+ "ит": 28562,
+ "▁саве": 28563,
+ "▁Protocol": 28564,
+ "ając": 28565,
+ "chk": 28566,
+ "TypeInfo": 28567,
+ "▁pkt": 28568,
+ "▁scandal": 28569,
+ "▁individually": 28570,
+ "FMT": 28571,
+ "▁nj": 28572,
+ "abile": 28573,
+ "▁Rivers": 28574,
+ "PROPERTY": 28575,
+ "VB": 28576,
+ "wort": 28577,
+ "▁splitting": 28578,
+ "achten": 28579,
+ "▁ARISING": 28580,
+ "▁sip": 28581,
+ "▁fres": 28582,
+ "▁groom": 28583,
+ "Hol": 28584,
+ "▁canon": 28585,
+ "▁abruptly": 28586,
+ "▁afterward": 28587,
+ "▁Running": 28588,
+ "▁ji": 28589,
+ "▁%,": 28590,
+ "▁Palestinian": 28591,
+ "RW": 28592,
+ "pgfscope": 28593,
+ "▁countryside": 28594,
+ "▁fortunate": 28595,
+ "▁cél": 28596,
+ "▁Pointer": 28597,
+ "ensors": 28598,
+ "rating": 28599,
+ "▁buffers": 28600,
+ "▁remot": 28601,
+ "▁PropTypes": 28602,
+ "▁Nah": 28603,
+ "altern": 28604,
+ "▁easiest": 28605,
+ "▁invas": 28606,
+ "▁clk": 28607,
+ "copyright": 28608,
+ "▁blanc": 28609,
+ "SAMP": 28610,
+ "▁Cohen": 28611,
+ "▁Shell": 28612,
+ "▁destroying": 28613,
+ "▁Zel": 28614,
+ "dater": 28615,
+ "čen": 28616,
+ "▁filing": 28617,
+ "▁integrate": 28618,
+ "xit": 28619,
+ "▁RET": 28620,
+ "lene": 28621,
+ "calls": 28622,
+ "▁slaughter": 28623,
+ "initialized": 28624,
+ "unches": 28625,
+ "▁Trace": 28626,
+ "efficient": 28627,
+ "▁Woods": 28628,
+ "▁longitud": 28629,
+ "GN": 28630,
+ "▁Kont": 28631,
+ "▁chunks": 28632,
+ "ách": 28633,
+ "▁unemployment": 28634,
+ "acom": 28635,
+ "▁slowed": 28636,
+ "▁outlined": 28637,
+ "xffff": 28638,
+ "▁ikke": 28639,
+ "▁workspace": 28640,
+ "Mc": 28641,
+ "▁kicking": 28642,
+ "▁embedding": 28643,
+ "chnitt": 28644,
+ "erten": 28645,
+ "▁Interior": 28646,
+ "▁Songs": 28647,
+ "mmc": 28648,
+ "▁analyzed": 28649,
+ "▁Coupe": 28650,
+ "▁favorites": 28651,
+ "▁tt": 28652,
+ "▁той": 28653,
+ "Routing": 28654,
+ "▁Silva": 28655,
+ "▁anderem": 28656,
+ "▁honom": 28657,
+ "▁использова": 28658,
+ ".\"]": 28659,
+ "▁Wu": 28660,
+ "legt": 28661,
+ "▁spoon": 28662,
+ "▁jap": 28663,
+ "▁Extension": 28664,
+ "erne": 28665,
+ "▁vagy": 28666,
+ "▁села": 28667,
+ "▁функ": 28668,
+ "▁analytics": 28669,
+ "▁sug": 28670,
+ "▁Async": 28671,
+ "▁peaks": 28672,
+ "▁Gym": 28673,
+ "▁lawsuit": 28674,
+ "<>": 28675,
+ "ialis": 28676,
+ "etric": 28677,
+ "faced": 28678,
+ "▁disrupt": 28679,
+ "▁få": 28680,
+ "Inputs": 28681,
+ "`);": 28682,
+ "▁Mend": 28683,
+ "gon": 28684,
+ "▁\",\"": 28685,
+ "▁nerves": 28686,
+ "▁doubts": 28687,
+ "sap": 28688,
+ "▁sow": 28689,
+ ",\\,\\": 28690,
+ "▁BS": 28691,
+ "▁Glad": 28692,
+ "▁aster": 28693,
+ "œuvre": 28694,
+ "▁Bangl": 28695,
+ "▁iPad": 28696,
+ "useppe": 28697,
+ "▁conducting": 28698,
+ "▁({\\": 28699,
+ "▁Harbor": 28700,
+ "psz": 28701,
+ "▁FIFA": 28702,
+ "_**": 28703,
+ "emor": 28704,
+ "▁": 28705,
+ "e": 28706,
+ "t": 28707,
+ "a": 28708,
+ "o": 28709,
+ "i": 28710,
+ "n": 28711,
+ "r": 28712,
+ "s": 28713,
+ "l": 28714,
+ "d": 28715,
+ "h": 28716,
+ "c": 28717,
+ "u": 28718,
+ "m": 28719,
+ "p": 28720,
+ "g": 28721,
+ "f": 28722,
+ ".": 28723,
+ "y": 28724,
+ ",": 28725,
+ "b": 28726,
+ "w": 28727,
+ "v": 28728,
+ "k": 28729,
+ "_": 28730,
+ ")": 28731,
+ "(": 28732,
+ "-": 28733,
+ "0": 28734,
+ "S": 28735,
+ "*": 28736,
+ "I": 28737,
+ "T": 28738,
+ "\"": 28739,
+ "1": 28740,
+ "A": 28741,
+ "'": 28742,
+ "C": 28743,
+ "x": 28744,
+ ";": 28745,
+ "=": 28746,
+ ":": 28747,
+ "/": 28748,
+ "E": 28749,
+ "2": 28750,
+ "{": 28751,
+ "}": 28752,
+ "P": 28753,
+ "R": 28754,
+ "M": 28755,
+ "\\": 28756,
+ "D": 28757,
+ "L": 28758,
+ "N": 28759,
+ "B": 28760,
+ "о": 28761,
+ "O": 28762,
+ "а": 28763,
+ "z": 28764,
+ "F": 28765,
+ "|": 28766,
+ ">": 28767,
+ "j": 28768,
+ "H": 28769,
+ "3": 28770,
+ "#": 28771,
+ "и": 28772,
+ "е": 28773,
+ "9": 28774,
+ "q": 28775,
+ "$": 28776,
+ "G": 28777,
+ "н": 28778,
+ "U": 28779,
+ "W": 28780,
+ "4": 28781,
+ "5": 28782,
+ "8": 28783,
+ "6": 28784,
+ "р": 28785,
+ "т": 28786,
+ "7": 28787,
+ "с": 28788,
+ "<": 28789,
+ "V": 28790,
+ "в": 28791,
+ "[": 28792,
+ "]": 28793,
+ "л": 28794,
+ "к": 28795,
+ "K": 28796,
+ "é": 28797,
+ "J": 28798,
+ "д": 28799,
+ "&": 28800,
+ "\r": 28801,
+ "Y": 28802,
+ "м": 28803,
+ "?": 28804,
+ "у": 28805,
+ "+": 28806,
+ "п": 28807,
+ "!": 28808,
+ "’": 28809,
+ "г": 28810,
+ "я": 28811,
+ "з": 28812,
+ "і": 28813,
+ "X": 28814,
+ "^": 28815,
+ "–": 28816,
+ "б": 28817,
+ "@": 28818,
+ "й": 28819,
+ "á": 28820,
+ "—": 28821,
+ "ь": 28822,
+ "%": 28823,
+ "Q": 28824,
+ "ó": 28825,
+ "ч": 28826,
+ "í": 28827,
+ "Z": 28828,
+ "ы": 28829,
+ "ä": 28830,
+ "х": 28831,
+ "`": 28832,
+ "ц": 28833,
+ "ö": 28834,
+ "“": 28835,
+ "ж": 28836,
+ "ü": 28837,
+ "”": 28838,
+ "à": 28839,
+ "è": 28840,
+ "ш": 28841,
+ "ю": 28842,
+ "ł": 28843,
+ "С": 28844,
+ "~": 28845,
+ "ф": 28846,
+ "П": 28847,
+ "»": 28848,
+ "В": 28849,
+ "«": 28850,
+ "å": 28851,
+ "К": 28852,
+ "щ": 28853,
+ "·": 28854,
+ "ј": 28855,
+ "М": 28856,
+ "ç": 28857,
+ "А": 28858,
+ "Н": 28859,
+ "Р": 28860,
+ "Б": 28861,
+ "č": 28862,
+ "ú": 28863,
+ "ę": 28864,
+ "ã": 28865,
+ "ą": 28866,
+ "ă": 28867,
+ "Д": 28868,
+ "ї": 28869,
+ "ъ": 28870,
+ "ě": 28871,
+ "Г": 28872,
+ "š": 28873,
+ "О": 28874,
+ "Т": 28875,
+ "ê": 28876,
+ "ñ": 28877,
+ "…": 28878,
+ "ž": 28879,
+ "ß": 28880,
+ "ё": 28881,
+ "ż": 28882,
+ "ř": 28883,
+ "ś": 28884,
+ "Л": 28885,
+ "ő": 28886,
+ "„": 28887,
+ "э": 28888,
+ "ý": 28889,
+ "У": 28890,
+ "â": 28891,
+ "И": 28892,
+ "є": 28893,
+ "‘": 28894,
+ "î": 28895,
+ "З": 28896,
+ "Ф": 28897,
+ "ò": 28898,
+ "•": 28899,
+ "ć": 28900,
+ "É": 28901,
+ "°": 28902,
+ "ș": 28903,
+ "Х": 28904,
+ "ț": 28905,
+ "ô": 28906,
+ "Е": 28907,
+ "ń": 28908,
+ "Ч": 28909,
+ "Ш": 28910,
+ "ø": 28911,
+ "ù": 28912,
+ "ů": 28913,
+ "的": 28914,
+ "ا": 28915,
+ "æ": 28916,
+ "њ": 28917,
+ "љ": 28918,
+ "ë": 28919,
+ "ï": 28920,
+ "Э": 28921,
+ "£": 28922,
+ "−": 28923,
+ ",": 28924,
+ "õ": 28925,
+ "ћ": 28926,
+ "": 28927,
+ "Ц": 28928,
+ "І": 28929,
+ "ā": 28930,
+ "ű": 28931,
+ "†": 28932,
+ "ل": 28933,
+ "ō": 28934,
+ "": 28935,
+ "º": 28936,
+ "Я": 28937,
+ "′": 28938,
+ "Á": 28939,
+ "Ö": 28940,
+ "²": 28941,
+ "Ж": 28942,
+ "ì": 28943,
+ "。": 28944,
+ "数": 28945,
+ "×": 28946,
+ "ر": 28947,
+ "α": 28948,
+ "́": 28949,
+ "Ю": 28950,
+ "û": 28951,
+ "œ": 28952,
+ "ı": 28953,
+ "م": 28954,
+ "ن": 28955,
+ "ª": 28956,
+ "ź": 28957,
+ "ο": 28958,
+ "″": 28959,
+ "€": 28960,
+ "Ü": 28961,
+ "و": 28962,
+ "用": 28963,
+ "À": 28964,
+ "Č": 28965,
+ "Š": 28966,
+ "ت": 28967,
+ "د": 28968,
+ "一": 28969,
+ "¿": 28970,
+ "是": 28971,
+ "ي": 28972,
+ "ђ": 28973,
+ "®": 28974,
+ "ی": 28975,
+ "ν": 28976,
+ "đ": 28977,
+ "τ": 28978,
+ "─": 28979,
+ "ι": 28980,
+ "ε": 28981,
+ "→": 28982,
+ "ب": 28983,
+ "Å": 28984,
+ "ū": 28985,
+ "№": 28986,
+ "ş": 28987,
+ "不": 28988,
+ "џ": 28989,
+ "ー": 28990,
+ "中": 28991,
+ "Î": 28992,
+ "の": 28993,
+ ":": 28994,
+ "个": 28995,
+ "Й": 28996,
+ "ρ": 28997,
+ "有": 28998,
+ "Ä": 28999,
+ " ": 29000,
+ "ī": 29001,
+ "©": 29002,
+ "为": 29003,
+ "ه": 29004,
+ "י": 29005,
+ "ו": 29006,
+ "时": 29007,
+ "س": 29008,
+ "Ś": 29009,
+ "在": 29010,
+ "件": 29011,
+ "取": 29012,
+ "ς": 29013,
+ "™": 29014,
+ "이": 29015,
+ "σ": 29016,
+ "μ": 29017,
+ "定": 29018,
+ "文": 29019,
+ "据": 29020,
+ "置": 29021,
+ "Ž": 29022,
+ "±": 29023,
+ "表": 29024,
+ "成": 29025,
+ "ň": 29026,
+ "λ": 29027,
+ "¡": 29028,
+ "È": 29029,
+ "π": 29030,
+ "字": 29031,
+ "│": 29032,
+ "Ј": 29033,
+ "回": 29034,
+ "Є": 29035,
+ "到": 29036,
+ "行": 29037,
+ "§": 29038,
+ "½": 29039,
+ "ع": 29040,
+ "、": 29041,
+ "Ł": 29042,
+ "다": 29043,
+ "ン": 29044,
+ "κ": 29045,
+ "名": 29046,
+ "ה": 29047,
+ "入": 29048,
+ "η": 29049,
+ "大": 29050,
+ "对": 29051,
+ "可": 29052,
+ "Â": 29053,
+ "上": 29054,
+ "█": 29055,
+ "新": 29056,
+ "ف": 29057,
+ "加": 29058,
+ "要": 29059,
+ "Ż": 29060,
+ "下": 29061,
+ "分": 29062,
+ "值": 29063,
+ "ת": 29064,
+ "出": 29065,
+ "类": 29066,
+ "请": 29067,
+ "": 29068,
+ "息": 29069,
+ "Ú": 29070,
+ "υ": 29071,
+ "获": 29072,
+ "示": 29073,
+ "以": 29074,
+ "ר": 29075,
+ "接": 29076,
+ "ל": 29077,
+ "を": 29078,
+ "存": 29079,
+ "信": 29080,
+ "设": 29081,
+ "方": 29082,
+ "ش": 29083,
+ "能": 29084,
+ "点": 29085,
+ "人": 29086,
+ "前": 29087,
+ "ğ": 29088,
+ "作": 29089,
+ "═": 29090,
+ "↘": 29091,
+ "ð": 29092,
+ "理": 29093,
+ "■": 29094,
+ "法": 29095,
+ "️": 29096,
+ "ˈ": 29097,
+ "果": 29098,
+ "发": 29099,
+ "ح": 29100,
+ "γ": 29101,
+ "ɵ": 29102,
+ "า": 29103,
+ "َ": 29104,
+ "了": 29105,
+ "户": 29106,
+ "Í": 29107,
+ "ə": 29108,
+ "ス": 29109,
+ "查": 29110,
+ "し": 29111,
+ "מ": 29112,
+ "单": 29113,
+ "ť": 29114,
+ "ق": 29115,
+ "る": 29116,
+ "间": 29117,
+ "如": 29118,
+ "本": 29119,
+ "后": 29120,
+ "ί": 29121,
+ "式": 29122,
+ "ト": 29123,
+ "Щ": 29124,
+ "Ó": 29125,
+ "す": 29126,
+ "א": 29127,
+ "生": 29128,
+ "动": 29129,
+ "ک": 29130,
+ "和": 29131,
+ "い": 29132,
+ "": 29133,
+ "ა": 29134,
+ "가": 29135,
+ "하": 29136,
+ "�": 29137,
+ "小": 29138,
+ "返": 29139,
+ "否": 29140,
+ "ة": 29141,
+ "日": 29142,
+ "로": 29143,
+ "标": 29144,
+ "码": 29145,
+ "地": 29146,
+ "位": 29147,
+ "에": 29148,
+ " ": 29149,
+ "列": 29150,
+ "수": 29151,
+ "β": 29152,
+ "除": 29153,
+ "使": 29154,
+ "ש": 29155,
+ "ج": 29156,
+ "イ": 29157,
+ "δ": 29158,
+ "自": 29159,
+ "于": 29160,
+ "지": 29161,
+ "当": 29162,
+ "所": 29163,
+ "기": 29164,
+ "ი": 29165,
+ "ב": 29166,
+ "ร": 29167,
+ "★": 29168,
+ "子": 29169,
+ "号": 29170,
+ "ك": 29171,
+ "参": 29172,
+ "型": 29173,
+ "に": 29174,
+ "는": 29175,
+ "这": 29176,
+ "开": 29177,
+ "น": 29178,
+ "会": 29179,
+ "器": 29180,
+ "面": 29181,
+ "ル": 29182,
+ "图": 29183,
+ "度": 29184,
+ ")": 29185,
+ "(": 29186,
+ "의": 29187,
+ "内": 29188,
+ "을": 29189,
+ "最": 29190,
+ "": 29191,
+ "化": 29192,
+ "建": 29193,
+ "니": 29194,
+ "量": 29195,
+ "😂": 29196,
+ "始": 29197,
+ "ē": 29198,
+ "خ": 29199,
+ "를": 29200,
+ "ά": 29201,
+ "过": 29202,
+ "³": 29203,
+ "´": 29204,
+ "组": 29205,
+ "功": 29206,
+ "": 29207,
+ "": 29208,
+ "区": 29209,
+ "ز": 29210,
+ "ґ": 29211,
+ "ό": 29212,
+ "ッ": 29213,
+ "ω": 29214,
+ "Ç": 29215,
+ "选": 29216,
+ "通": 29217,
+ "结": 29218,
+ "录": 29219,
+ "改": 29220,
+ "ク": 29221,
+ "目": 29222,
+ "指": 29223,
+ "务": 29224,
+ "๐": 29225,
+ "输": 29226,
+ "た": 29227,
+ "อ": 29228,
+ "关": 29229,
+ "で": 29230,
+ "调": 29231,
+ "ा": 29232,
+ "정": 29233,
+ "合": 29234,
+ "已": 29235,
+ "시": 29236,
+ "部": 29237,
+ "页": 29238,
+ "━": 29239,
+ "ː": 29240,
+ "ま": 29241,
+ "我": 29242,
+ "求": 29243,
+ "市": 29244,
+ "次": 29245,
+ "נ": 29246,
+ "实": 29247,
+ "将": 29248,
+ "重": 29249,
+ "更": 29250,
+ "制": 29251,
+ "符": 29252,
+ "配": 29253,
+ "象": 29254,
+ "θ": 29255,
+ "ก": 29256,
+ "て": 29257,
+ "进": 29258,
+ "需": 29259,
+ "Đ": 29260,
+ "性": 29261,
+ "认": 29262,
+ "来": 29263,
+ "题": 29264,
+ "程": 29265,
+ "模": 29266,
+ "!": 29267,
+ "失": 29268,
+ "口": 29269,
+ "な": 29270,
+ "έ": 29271,
+ "": 29272,
+ "空": 29273,
+ "": 29274,
+ "期": 29275,
+ "者": 29276,
+ "は": 29277,
+ "Ђ": 29278,
+ "提": 29279,
+ "ή": 29280,
+ "ラ": 29281,
+ "한": 29282,
+ "态": 29283,
+ "复": 29284,
+ "ง": 29285,
+ "ე": 29286,
+ "Ø": 29287,
+ "리": 29288,
+ "修": 29289,
+ "‚": 29290,
+ "得": 29291,
+ "多": 29292,
+ "格": 29293,
+ "자": 29294,
+ "ע": 29295,
+ "่": 29296,
+ "函": 29297,
+ "应": 29298,
+ "↗": 29299,
+ "्": 29300,
+ "เ": 29301,
+ "正": 29302,
+ "注": 29303,
+ "스": 29304,
+ "서": 29305,
+ "リ": 29306,
+ "φ": 29307,
+ "ص": 29308,
+ "が": 29309,
+ "则": 29310,
+ "消": 29311,
+ "节": 29312,
+ "序": 29313,
+ "代": 29314,
+ "사": 29315,
+ "と": 29316,
+ "ד": 29317,
+ "้": 29318,
+ "र": 29319,
+ "此": 29320,
+ "保": 29321,
+ "ア": 29322,
+ "ư": 29323,
+ "인": 29324,
+ "ė": 29325,
+ "处": 29326,
+ "删": 29327,
+ "ɛ": 29328,
+ "容": 29329,
+ "ط": 29330,
+ "": 29331,
+ "之": 29332,
+ "包": 29333,
+ "状": 29334,
+ "ド": 29335,
+ "İ": 29336,
+ "体": 29337,
+ "同": 29338,
+ "事": 29339,
+ "🙂": 29340,
+ "タ": 29341,
+ "χ": 29342,
+ "ʿ": 29343,
+ "Ș": 29344,
+ "主": 29345,
+ "品": 29346,
+ "ק": 29347,
+ "询": 29348,
+ "创": 29349,
+ "该": 29350,
+ " ": 29351,
+ "元": 29352,
+ "第": 29353,
+ "天": 29354,
+ "或": 29355,
+ "年": 29356,
+ "转": 29357,
+ "ח": 29358,
+ "传": 29359,
+ "ţ": 29360,
+ "路": 29361,
+ "例": 29362,
+ "机": 29363,
+ "Ã": 29364,
+ "ď": 29365,
+ "高": 29366,
+ "相": 29367,
+ "โ": 29368,
+ "片": 29369,
+ "―": 29370,
+ "操": 29371,
+ "ա": 29372,
+ "ม": 29373,
+ "全": 29374,
+ "无": 29375,
+ "月": 29376,
+ "称": 29377,
+ "ั": 29378,
+ "就": 29379,
+ "": 29380,
+ "明": 29381,
+ "计": 29382,
+ "你": 29383,
+ "败": 29384,
+ "密": 29385,
+ "解": 29386,
+ "れ": 29387,
+ "أ": 29388,
+ "变": 29389,
+ "段": 29390,
+ "条": 29391,
+ "默": 29392,
+ "●": 29393,
+ "ล": 29394,
+ "色": 29395,
+ "断": 29396,
+ "商": 29397,
+ "ם": 29398,
+ "か": 29399,
+ "里": 29400,
+ "系": 29401,
+ "编": 29402,
+ "错": 29403,
+ "트": 29404,
+ "只": 29405,
+ "县": 29406,
+ "ს": 29407,
+ "常": 29408,
+ "初": 29409,
+ "ɔ": 29410,
+ "Α": 29411,
+ "フ": 29412,
+ "►": 29413,
+ "等": 29414,
+ "일": 29415,
+ "・": 29416,
+ "Ō": 29417,
+ "情": 29418,
+ "现": 29419,
+ "Ř": 29420,
+ "ِ": 29421,
+ "さ": 29422,
+ "ạ": 29423,
+ "용": 29424,
+ "证": 29425,
+ "해": 29426,
+ "手": 29427,
+ "支": 29428,
+ "입": 29429,
+ "服": 29430,
+ "்": 29431,
+ "道": 29432,
+ "어": 29433,
+ "送": 29434,
+ "载": 29435,
+ "限": 29436,
+ "线": 29437,
+ "属": 29438,
+ "": 29439,
+ "他": 29440,
+ "放": 29441,
+ "记": 29442,
+ "公": 29443,
+ "没": 29444,
+ "添": 29445,
+ "显": 29446,
+ "บ": 29447,
+ "ย": 29448,
+ "რ": 29449,
+ "其": 29450,
+ "集": 29451,
+ "金": 29452,
+ "国": 29453,
+ "任": 29454,
+ "ە": 29455,
+ "话": 29456,
+ "并": 29457,
+ "被": 29458,
+ "ύ": 29459,
+ "都": 29460,
+ "گ": 29461,
+ "意": 29462,
+ "כ": 29463,
+ "经": 29464,
+ "성": 29465,
+ "看": 29466,
+ "פ": 29467,
+ "址": 29468,
+ "ס": 29469,
+ "드": 29470,
+ "交": 29471,
+ "¼": 29472,
+ "Џ": 29473,
+ "完": 29474,
+ "Δ": 29475,
+ "义": 29476,
+ "보": 29477,
+ "向": 29478,
+ "换": 29479,
+ "山": 29480,
+ "算": 29481,
+ "二": 29482,
+ "پ": 29483,
+ "⁄": 29484,
+ "判": 29485,
+ "级": 29486,
+ "工": 29487,
+ "ด": 29488,
+ "⠀": 29489,
+ "家": 29490,
+ "レ": 29491,
+ "三": 29492,
+ "原": 29493,
+ "】": 29494,
+ "长": 29495,
+ "া": 29496,
+ "管": 29497,
+ "ѝ": 29498,
+ "क": 29499,
+ "学": 29500,
+ "ロ": 29501,
+ "验": 29502,
+ "写": 29503,
+ "Œ": 29504,
+ "从": 29505,
+ "【": 29506,
+ "收": 29507,
+ "ả": 29508,
+ "未": 29509,
+ "登": 29510,
+ "고": 29511,
+ "源": 29512,
+ "每": 29513,
+ "µ": 29514,
+ "误": 29515,
+ "り": 29516,
+ "요": 29517,
+ "按": 29518,
+ "ว": 29519,
+ "权": 29520,
+ "根": 29521,
+ "プ": 29522,
+ "串": 29523,
+ "ส": 29524,
+ "›": 29525,
+ "제": 29526,
+ "シ": 29527,
+ "Ş": 29528,
+ "确": 29529,
+ "好": 29530,
+ "统": 29531,
+ "效": 29532,
+ "网": 29533,
+ "\u0001": 29534,
+ "物": 29535,
+ "아": 29536,
+ "也": 29537,
+ "은": 29538,
+ "ệ": 29539,
+ "न": 29540,
+ "项": 29541,
+ "资": 29542,
+ "こ": 29543,
+ "引": 29544,
+ "ジ": 29545,
+ "ค": 29546,
+ "版": 29547,
+ "ท": 29548,
+ "平": 29549,
+ "们": 29550,
+ "与": 29551,
+ "き": 29552,
+ "移": 29553,
+ "ि": 29554,
+ "素": 29555,
+ "执": 29556,
+ "주": 29557,
+ "‐": 29558,
+ "Ґ": 29559,
+ "ี": 29560,
+ "板": 29561,
+ "问": 29562,
+ "Ε": 29563,
+ "安": 29564,
+ "면": 29565,
+ "소": 29566,
+ "ต": 29567,
+ "ิ": 29568,
+ "持": 29569,
+ "습": 29570,
+ "Σ": 29571,
+ "ら": 29572,
+ "コ": 29573,
+ "心": 29574,
+ "Π": 29575,
+ "打": 29576,
+ "」": 29577,
+ "상": 29578,
+ "「": 29579,
+ "检": 29580,
+ "库": 29581,
+ "÷": 29582,
+ "으": 29583,
+ "测": 29584,
+ "ん": 29585,
+ "े": 29586,
+ "ُ": 29587,
+ "力": 29588,
+ "直": 29589,
+ "由": 29590,
+ "ى": 29591,
+ "试": 29592,
+ "必": 29593,
+ "端": 29594,
+ "ʻ": 29595,
+ "先": 29596,
+ "↑": 29597,
+ "命": 29598,
+ "도": 29599,
+ "전": 29600,
+ "ห": 29601,
+ "员": 29602,
+ "ɪ": 29603,
+ "있": 29604,
+ "比": 29605,
+ "ṣ": 29606,
+ "時": 29607,
+ "择": 29608,
+ "ذ": 29609,
+ "テ": 29610,
+ "": 29611,
+ "构": 29612,
+ "备": 29613,
+ "그": 29614,
+ "链": 29615,
+ "说": 29616,
+ "ლ": 29617,
+ "ן": 29618,
+ "签": 29619,
+ "う": 29620,
+ "غ": 29621,
+ "ế": 29622,
+ "ض": 29623,
+ "ḥ": 29624,
+ "启": 29625,
+ "력": 29626,
+ "ო": 29627,
+ "付": 29628,
+ "მ": 29629,
+ "索": 29630,
+ "特": 29631,
+ "ג": 29632,
+ "西": 29633,
+ "대": 29634,
+ "├": 29635,
+ "": 29636,
+ "": 29637,
+ "外": 29638,
+ "צ": 29639,
+ "头": 29640,
+ "连": 29641,
+ "流": 29642,
+ "◄": 29643,
+ "デ": 29644,
+ "カ": 29645,
+ "র": 29646,
+ "오": 29647,
+ "找": 29648,
+ "清": 29649,
+ "🤣": 29650,
+ "去": 29651,
+ "₹": 29652,
+ "경": 29653,
+ "グ": 29654,
+ "ْ": 29655,
+ "¢": 29656,
+ "因": 29657,
+ "": 29658,
+ "Κ": 29659,
+ "增": 29660,
+ "知": 29661,
+ "¶": 29662,
+ "像": 29663,
+ "♥": 29664,
+ "터": 29665,
+ "く": 29666,
+ "ậ": 29667,
+ "メ": 29668,
+ "Æ": 29669,
+ "省": 29670,
+ "स": 29671,
+ "म": 29672,
+ "❤": 29673,
+ "あ": 29674,
+ "样": 29675,
+ "起": 29676,
+ "台": 29677,
+ "读": 29678,
+ "角": 29679,
+ "南": 29680,
+ "整": 29681,
+ "订": 29682,
+ "\f": 29683,
+ "ט": 29684,
+ "マ": 29685,
+ "্": 29686,
+ "우": 29687,
+ "ն": 29688,
+ "您": 29689,
+ "ئ": 29690,
+ "基": 29691,
+ "水": 29692,
+ "생": 29693,
+ "‑": 29694,
+ "나": 29695,
+ "画": 29696,
+ "描": 29697,
+ "击": 29698,
+ "っ": 29699,
+ "라": 29700,
+ "ნ": 29701,
+ "ր": 29702,
+ "业": 29703,
+ "ბ": 29704,
+ "别": 29705,
+ "♦": 29706,
+ "ィ": 29707,
+ "त": 29708,
+ "给": 29709,
+ "문": 29710,
+ "形": 29711,
+ "控": 29712,
+ "然": 29713,
+ "동": 29714,
+ "Њ": 29715,
+ "": 29716,
+ "东": 29717,
+ "ป": 29718,
+ "州": 29719,
+ "排": 29720,
+ "세": 29721,
+ "装": 29722,
+ "할": 29723,
+ "Ć": 29724,
+ "∞": 29725,
+ "海": 29726,
+ "城": 29727,
+ "键": 29728,
+ "径": 29729,
+ "호": 29730,
+ "화": 29731,
+ "្": 29732,
+ "料": 29733,
+ "ơ": 29734,
+ "ी": 29735,
+ "ウ": 29736,
+ "具": 29737,
+ "ブ": 29738,
+ "块": 29739,
+ "再": 29740,
+ "ố": 29741,
+ "电": 29742,
+ ";": 29743,
+ "위": 29744,
+ "两": 29745,
+ "而": 29746,
+ "장": 29747,
+ "آ": 29748,
+ "Ț": 29749,
+ "バ": 29750,
+ "还": 29751,
+ "令": 29752,
+ "キ": 29753,
+ "ّ": 29754,
+ "값": 29755,
+ "번": 29756,
+ "만": 29757,
+ "总": 29758,
+ "ल": 29759,
+ "▲": 29760,
+ "异": 29761,
+ "光": 29762,
+ "客": 29763,
+ "非": 29764,
+ "ị": 29765,
+ "": 29766,
+ "þ": 29767,
+ "設": 29768,
+ "述": 29769,
+ "합": 29770,
+ "?": 29771,
+ "✔": 29772,
+ "导": 29773,
+ "ṇ": 29774,
+ "부": 29775,
+ "˙": 29776,
+ "Τ": 29777,
+ "も": 29778,
+ "구": 29779,
+ "镇": 29780,
+ "작": 29781,
+ "░": 29782,
+ "步": 29783,
+ "ộ": 29784,
+ "活": 29785,
+ "พ": 29786,
+ "←": 29787,
+ "ǎ": 29788,
+ "จ": 29789,
+ "束": 29790,
+ "ـ": 29791,
+ "": 29792,
+ "那": 29793,
+ "प": 29794,
+ "エ": 29795,
+ "志": 29796,
+ "么": 29797,
+ "运": 29798,
+ "北": 29799,
+ "超": 29800,
+ "་": 29801,
+ "布": 29802,
+ "ώ": 29803,
+ "͡": 29804,
+ "少": 29805,
+ "파": 29806,
+ "ʃ": 29807,
+ "ム": 29808,
+ "": 29809,
+ "卡": 29810,
+ "ন": 29811,
+ "Μ": 29812,
+ "ɑ": 29813,
+ "😉": 29814,
+ "辑": 29815,
+ "원": 29816,
+ "美": 29817,
+ "产": 29818,
+ "利": 29819,
+ "모": 29820,
+ "联": 29821,
+ "界": 29822,
+ "체": 29823,
+ "种": 29824,
+ "王": 29825,
+ "ľ": 29826,
+ "여": 29827,
+ "메": 29828,
+ "域": 29829,
+ "ვ": 29830,
+ "立": 29831,
+ "록": 29832,
+ "게": 29833,
+ "إ": 29834,
+ "ṭ": 29835,
+ "神": 29836,
+ "ո": 29837,
+ "音": 29838,
+ "☆": 29839,
+ "Ñ": 29840,
+ "조": 29841,
+ "動": 29842,
+ "缓": 29843,
+ "과": 29844,
+ "报": 29845,
+ "ʼ": 29846,
+ "ា": 29847,
+ "되": 29848,
+ "ե": 29849,
+ "视": 29850,
+ "ช": 29851,
+ "详": 29852,
+ "แ": 29853,
+ "¦": 29854,
+ "把": 29855,
+ "க": 29856,
+ "ি": 29857,
+ "출": 29858,
+ "비": 29859,
+ "边": 29860,
+ "框": 29861,
+ "व": 29862,
+ "サ": 29863,
+ "Ι": 29864,
+ "Ο": 29865,
+ "オ": 29866,
+ "¾": 29867,
+ "历": 29868,
+ "ŏ": 29869,
+ "门": 29870,
+ "ข": 29871,
+ "含": 29872,
+ "¬": 29873,
+ "周": 29874,
+ "填": 29875,
+ "待": 29876,
+ "ะ": 29877,
+ "დ": 29878,
+ "Ї": 29879,
+ "额": 29880,
+ "음": 29881,
+ "四": 29882,
+ "だ": 29883,
+ "회": 29884,
+ "止": 29885,
+ "率": 29886,
+ "环": 29887,
+ "パ": 29888,
+ "래": 29889,
+ "闭": 29890,
+ "̀": 29891,
+ "语": 29892,
+ "개": 29893,
+ "身": 29894,
+ "藏": 29895,
+ "य": 29896,
+ "된": 29897,
+ "即": 29898,
+ "拉": 29899,
+ "선": 29900,
+ "변": 29901,
+ "≥": 29902,
+ "ุ": 29903,
+ "些": 29904,
+ "🤷": 29905,
+ "せ": 29906,
+ "左": 29907,
+ "ợ": 29908,
+ "右": 29909,
+ "ể": 29910,
+ "내": 29911,
+ "ּ": 29912,
+ "ז": 29913,
+ "ে": 29914,
+ "告": 29915,
+ "ấ": 29916,
+ "白": 29917,
+ "账": 29918,
+ "费": 29919,
+ "江": 29920,
+ "み": 29921,
+ "‹": 29922,
+ "์": 29923,
+ "": 29924,
+ "造": 29925,
+ "但": 29926,
+ "十": 29927,
+ "它": 29928,
+ "ं": 29929,
+ "ŋ": 29930,
+ "ў": 29931,
+ "セ": 29932,
+ "女": 29933,
+ "⣿": 29934,
+ "ի": 29935,
+ "京": 29936,
+ "触": 29937,
+ "함": 29938,
+ "들": 29939,
+ "Ā": 29940,
+ "": 29941,
+ "石": 29942,
+ "よ": 29943,
+ "田": 29944,
+ "易": 29945,
+ "规": 29946,
+ "展": 29947,
+ "¯": 29948,
+ "做": 29949,
+ "星": 29950,
+ "უ": 29951,
+ "✓": 29952,
+ "თ": 29953,
+ "供": 29954,
+ "명": 29955,
+ "ξ": 29956,
+ "己": 29957,
+ "且": 29958,
+ "插": 29959,
+ "景": 29960,
+ "切": 29961,
+ "ไ": 29962,
+ "없": 29963,
+ "ョ": 29964,
+ "及": 29965,
+ "Ν": 29966,
+ "미": 29967,
+ "ث": 29968,
+ "데": 29969,
+ "价": 29970,
+ "乡": 29971,
+ "ह": 29972,
+ "チ": 29973,
+ "真": 29974,
+ "太": 29975,
+ "ู": 29976,
+ "ダ": 29977,
+ "局": 29978,
+ "♂": 29979,
+ "退": 29980,
+ "ு": 29981,
+ "ক": 29982,
+ "ி": 29983,
+ "何": 29984,
+ "😭": 29985,
+ "¥": 29986,
+ "": 29987,
+ "≈": 29988,
+ "司": 29989,
+ "层": 29990,
+ "실": 29991,
+ "站": 29992,
+ "首": 29993,
+ "款": 29994,
+ "រ": 29995,
+ "間": 29996,
+ "ָ": 29997,
+ "저": 29998,
+ "监": 29999,
+ "ァ": 30000,
+ "册": 30001,
+ "案": 30002,
+ "ो": 30003,
+ "反": 30004,
+ "听": 30005,
+ "族": 30006,
+ "析": 30007,
+ "ื": 30008,
+ "秒": 30009,
+ "공": 30010,
+ "": 30011,
+ "🚀": 30012,
+ "거": 30013,
+ "재": 30014,
+ "": 30015,
+ "場": 30016,
+ "广": 30017,
+ "播": 30018,
+ "║": 30019,
+ "⋅": 30020,
+ "技": 30021,
+ "贴": 30022,
+ "想": 30023,
+ "ʁ": 30024,
+ "ớ": 30025,
+ "ャ": 30026,
+ "중": 30027,
+ "》": 30028,
+ "速": 30029,
+ "频": 30030,
+ "队": 30031,
+ "ำ": 30032,
+ "け": 30033,
+ "ु": 30034,
+ "≤": 30035,
+ "↓": 30036,
+ "须": 30037,
+ "菜": 30038,
+ "̃": 30039,
+ "剪": 30040,
+ "버": 30041,
+ "ェ": 30042,
+ "Λ": 30043,
+ "细": 30044,
+ "選": 30045,
+ "द": 30046,
+ "¹": 30047,
+ "许": 30048,
+ "ầ": 30049,
+ "世": 30050,
+ "ュ": 30051,
+ "ء": 30052,
+ "‡": 30053,
+ "候": 30054,
+ "共": 30055,
+ "크": 30056,
+ "ธ": 30057,
+ "설": 30058,
+ "快": 30059,
+ "友": 30060,
+ "ְ": 30061,
+ "车": 30062,
+ "推": 30063,
+ "花": 30064,
+ "言": 30065,
+ "چ": 30066,
+ "至": 30067,
+ "開": 30068,
+ "校": 30069,
+ "個": 30070,
+ "村": 30071,
+ "つ": 30072,
+ "▌": 30073,
+ "ப": 30074,
+ "결": 30075,
+ "ņ": 30076,
+ "优": 30077,
+ "ន": 30078,
+ "达": 30079,
+ "核": 30080,
+ "ナ": 30081,
+ "场": 30082,
+ "影": 30083,
+ "🏻": 30084,
+ "钮": 30085,
+ "ظ": 30086,
+ "Þ": 30087,
+ "▼": 30088,
+ "お": 30089,
+ "份": 30090,
+ "微": 30091,
+ "ờ": 30092,
+ "识": 30093,
+ "행": 30094,
+ "《": 30095,
+ "ใ": 30096,
+ "ọ": 30097,
+ "预": 30098,
+ "ব": 30099,
+ "த": 30100,
+ "": 30101,
+ "ų": 30102,
+ "마": 30103,
+ "않": 30104,
+ "ɡ": 30105,
+ "계": 30106,
+ "연": 30107,
+ "五": 30108,
+ "Ź": 30109,
+ "め": 30110,
+ "很": 30111,
+ "간": 30112,
+ "無": 30113,
+ "ប": 30114,
+ "社": 30115,
+ "Ê": 30116,
+ "书": 30117,
+ "顶": 30118,
+ "ტ": 30119,
+ "才": 30120,
+ "云": 30121,
+ "└": 30122,
+ "ζ": 30123,
+ "،": 30124,
+ "搜": 30125,
+ "신": 30126,
+ "유": 30127,
+ "": 30128,
+ "✅": 30129,
+ "⭐": 30130,
+ "照": 30131,
+ "短": 30132,
+ "川": 30133,
+ "後": 30134,
+ "范": 30135,
+ "民": 30136,
+ "治": 30137,
+ "章": 30138,
+ "ề": 30139,
+ "바": 30140,
+ "ә": 30141,
+ "⚭": 30142,
+ "河": 30143,
+ "论": 30144,
+ "え": 30145,
+ "Ω": 30146,
+ "√": 30147,
+ "Ă": 30148,
+ "Γ": 30149,
+ "坐": 30150,
+ "적": 30151,
+ "停": 30152,
+ "추": 30153,
+ "受": 30154,
+ "♀": 30155,
+ "ʾ": 30156,
+ "树": 30157,
+ "林": 30158,
+ "치": 30159,
+ "fi": 30160,
+ "▒": 30161,
+ "张": 30162,
+ "着": 30163,
+ "访": 30164,
+ "考": 30165,
+ "教": 30166,
+ "ग": 30167,
+ "准": 30168,
+ "印": 30169,
+ "精": 30170,
+ "窗": 30171,
+ "宝": 30172,
+ "ち": 30173,
+ "围": 30174,
+ "ַ": 30175,
+ "致": 30176,
+ "モ": 30177,
+ "때": 30178,
+ "随": 30179,
+ "储": 30180,
+ "况": 30181,
+ "邮": 30182,
+ "武": 30183,
+ "⛔": 30184,
+ "维": 30185,
+ "ү": 30186,
+ "跳": 30187,
+ "ब": 30188,
+ "投": 30189,
+ "ủ": 30190,
+ "표": 30191,
+ "반": 30192,
+ "英": 30193,
+ "ʰ": 30194,
+ "👍": 30195,
+ "ज": 30196,
+ "带": 30197,
+ "為": 30198,
+ "续": 30199,
+ "ɨ": 30200,
+ "처": 30201,
+ "₂": 30202,
+ "클": 30203,
+ "群": 30204,
+ "현": 30205,
+ "风": 30206,
+ "购": 30207,
+ "ក": 30208,
+ "老": 30209,
+ "留": 30210,
+ "球": 30211,
+ "프": 30212,
+ "▄": 30213,
+ "史": 30214,
+ "Љ": 30215,
+ "⟩": 30216,
+ "분": 30217,
+ "გ": 30218,
+ "店": 30219,
+ "审": 30220,
+ "료": 30221,
+ "목": 30222,
+ "略": 30223,
+ "관": 30224,
+ "ִ": 30225,
+ "科": 30226,
+ "货": 30227,
+ "ம": 30228,
+ "络": 30229,
+ "阳": 30230,
+ "Ḥ": 30231,
+ "資": 30232,
+ "若": 30233,
+ "স": 30234,
+ "ہ": 30235,
+ "宽": 30236,
+ "见": 30237,
+ "ズ": 30238,
+ "游": 30239,
+ "방": 30240,
+ "ồ": 30241,
+ "ɾ": 30242,
+ "열": 30243,
+ "러": 30244,
+ "ך": 30245,
+ "\u001b": 30246,
+ "်": 30247,
+ "余": 30248,
+ "响": 30249,
+ "缩": 30250,
+ "ட": 30251,
+ "评": 30252,
+ "允": 30253,
+ "离": 30254,
+ "🤔": 30255,
+ "Ё": 30256,
+ "ʊ": 30257,
+ "黑": 30258,
+ "马": 30259,
+ "⟨": 30260,
+ "値": 30261,
+ "箱": 30262,
+ "야": 30263,
+ "ម": 30264,
+ "Ő": 30265,
+ "感": 30266,
+ "ツ": 30267,
+ "ụ": 30268,
+ "ポ": 30269,
+ "확": 30270,
+ "声": 30271,
+ "战": 30272,
+ "ѕ": 30273,
+ "変": 30274,
+ "와": 30275,
+ "父": 30276,
+ "ベ": 30277,
+ "助": 30278,
+ "업": 30279,
+ "ʲ": 30280,
+ "ÿ": 30281,
+ "充": 30282,
+ "强": 30283,
+ "博": 30284,
+ "ミ": 30285,
+ "销": 30286,
+ "당": 30287,
+ "記": 30288,
+ "什": 30289,
+ "匹": 30290,
+ "ւ": 30291,
+ "そ": 30292,
+ "코": 30293,
+ "ল": 30294,
+ "ŭ": 30295,
+ "午": 30296,
+ "ニ": 30297,
+ "\u0012": 30298,
+ "ʒ": 30299,
+ "შ": 30300,
+ "某": 30301,
+ "ォ": 30302,
+ "足": 30303,
+ "타": 30304,
+ "Ð": 30305,
+ "ხ": 30306,
+ "름": 30307,
+ "木": 30308,
+ "楼": 30309,
+ "최": 30310,
+ "红": 30311,
+ "¨": 30312,
+ "古": 30313,
+ "\u0006": 30314,
+ "단": 30315,
+ "今": 30316,
+ "ʔ": 30317,
+ "ट": 30318,
+ "ম": 30319,
+ "斯": 30320,
+ "語": 30321,
+ "Ÿ": 30322,
+ "🙄": 30323,
+ "牌": 30324,
+ "안": 30325,
+ "ស": 30326,
+ "颜": 30327,
+ "~": 30328,
+ "克": 30329,
+ "深": 30330,
+ "금": 30331,
+ "會": 30332,
+ "尔": 30333,
+ "释": 30334,
+ "批": 30335,
+ "산": 30336,
+ "野": 30337,
+ "防": 30338,
+ "Η": 30339,
+ "ө": 30340,
+ "ψ": 30341,
+ "ボ": 30342,
+ "": 30343,
+ "各": 30344,
+ "진": 30345,
+ "追": 30346,
+ "句": 30347,
+ "警": 30348,
+ "Φ": 30349,
+ "ѣ": 30350,
+ "ḍ": 30351,
+ "词": 30352,
+ "男": 30353,
+ "글": 30354,
+ "식": 30355,
+ "隐": 30356,
+ "복": 30357,
+ "盘": 30358,
+ "Ì": 30359,
+ "申": 30360,
+ "议": 30361,
+ "ザ": 30362,
+ "近": 30363,
+ "능": 30364,
+ "য": 30365,
+ "東": 30366,
+ "這": 30367,
+ "ர": 30368,
+ "距": 30369,
+ "院": 30370,
+ "德": 30371,
+ "ǐ": 30372,
+ "针": 30373,
+ "▀": 30374,
+ "↔": 30375,
+ "房": 30376,
+ "青": 30377,
+ "政": 30378,
+ "😅": 30379,
+ "递": 30380,
+ "প": 30381,
+ "波": 30382,
+ "ソ": 30383,
+ "绑": 30384,
+ "ビ": 30385,
+ "ễ": 30386,
+ "포": 30387,
+ "\u0010": 30388,
+ "ử": 30389,
+ "등": 30390,
+ "환": 30391,
+ "士": 30392,
+ "ত": 30393,
+ "Θ": 30394,
+ "초": 30395,
+ "境": 30396,
+ "差": 30397,
+ "采": 30398,
+ "디": 30399,
+ "ĩ": 30400,
+ "升": 30401,
+ "背": 30402,
+ "배": 30403,
+ "龙": 30404,
+ "街": 30405,
+ "್": 30406,
+ "ṛ": 30407,
+ "ু": 30408,
+ "弹": 30409,
+ "魔": 30410,
+ "객": 30411,
+ "‰": 30412,
+ "⌁": 30413,
+ "ἐ": 30414,
+ "禁": 30415,
+ "ผ": 30416,
+ "қ": 30417,
+ "島": 30418,
+ "ா": 30419,
+ "♭": 30420,
+ "百": 30421,
+ "ứ": 30422,
+ "ネ": 30423,
+ "专": 30424,
+ "來": 30425,
+ "刷": 30426,
+ "필": 30427,
+ "յ": 30428,
+ "ắ": 30429,
+ "华": 30430,
+ "Β": 30431,
+ "श": 30432,
+ "¸": 30433,
+ "屏": 30434,
+ "死": 30435,
+ "遍": 30436,
+ "검": 30437,
+ "Χ": 30438,
+ "것": 30439,
+ "八": 30440,
+ "览": 30441,
+ "택": 30442,
+ "唯": 30443,
+ "∙": 30444,
+ "¤": 30445,
+ "페": 30446,
+ "让": 30447,
+ "锁": 30448,
+ "무": 30449,
+ "思": 30450,
+ "隔": 30451,
+ "Ô": 30452,
+ "\u0013": 30453,
+ "ṃ": 30454,
+ "ワ": 30455,
+ "低": 30456,
+ "션": 30457,
+ "半": 30458,
+ "较": 30459,
+ "ត": 30460,
+ "享": 30461,
+ "积": 30462,
+ "": 30463,
+ "😊": 30464,
+ "典": 30465,
+ "ǔ": 30466,
+ "六": 30467,
+ "便": 30468,
+ "ɐ": 30469,
+ "简": 30470,
+ "继": 30471,
+ "仅": 30472,
+ "尾": 30473,
+ "": 30474,
+ "வ": 30475,
+ "կ": 30476,
+ "": 30477,
+ "영": 30478,
+ "火": 30479,
+ "湖": 30480,
+ "書": 30481,
+ "발": 30482,
+ "ハ": 30483,
+ "循": 30484,
+ "术": 30485,
+ "結": 30486,
+ "ļ": 30487,
+ "乐": 30488,
+ "滤": 30489,
+ "종": 30490,
+ "ถ": 30491,
+ "ὶ": 30492,
+ "满": 30493,
+ "╝": 30494,
+ "わ": 30495,
+ "ど": 30496,
+ "็": 30497,
+ "형": 30498,
+ "國": 30499,
+ "ự": 30500,
+ "線": 30501,
+ "블": 30502,
+ "封": 30503,
+ "確": 30504,
+ "依": 30505,
+ "ս": 30506,
+ "永": 30507,
+ "색": 30508,
+ "歌": 30509,
+ "數": 30510,
+ "福": 30511,
+ "삭": 30512,
+ "実": 30513,
+ "레": 30514,
+ "ſ": 30515,
+ "千": 30516,
+ "\u000e": 30517,
+ "母": 30518,
+ "더": 30519,
+ "임": 30520,
+ "տ": 30521,
+ "ے": 30522,
+ "几": 30523,
+ "双": 30524,
+ "노": 30525,
+ "ณ": 30526,
+ "掉": 30527,
+ "Ρ": 30528,
+ "ἀ": 30529,
+ "標": 30530,
+ "長": 30531,
+ "档": 30532,
+ "태": 30533,
+ "ペ": 30534,
+ "본": 30535,
+ "": 30536,
+ "底": 30537,
+ "终": 30538,
+ "請": 30539,
+ "კ": 30540,
+ "̯": 30541,
+ "예": 30542,
+ "▬": 30543,
+ "報": 30544,
+ "ピ": 30545,
+ "๏": 30546,
+ "暂": 30547,
+ "李": 30548,
+ "Υ": 30549,
+ "\u0005": 30550,
+ "\u0002": 30551,
+ "替": 30552,
+ "운": 30553,
+ "射": 30554,
+ "\u0018": 30555,
+ "매": 30556,
+ "\u0011": 30557,
+ "🏼": 30558,
+ "票": 30559,
+ "附": 30560,
+ "ノ": 30561,
+ "ũ": 30562,
+ "压": 30563,
+ "阿": 30564,
+ "Ò": 30565,
+ "테": 30566,
+ "∼": 30567,
+ "万": 30568,
+ "մ": 30569,
+ "후": 30570,
+ "普": 30571,
+ "截": 30572,
+ "속": 30573,
+ "括": 30574,
+ "😀": 30575,
+ "ை": 30576,
+ "▶": 30577,
+ "까": 30578,
+ "ট": 30579,
+ "曲": 30580,
+ "师": 30581,
+ "钱": 30582,
+ "栏": 30583,
+ "Ы": 30584,
+ "走": 30585,
+ "ữ": 30586,
+ "": 30587,
+ "归": 30588,
+ "점": 30589,
+ "🔥": 30590,
+ "었": 30591,
+ "連": 30592,
+ "私": 30593,
+ "청": 30594,
+ "刘": 30595,
+ "免": 30596,
+ "": 30597,
+ "奖": 30598,
+ "見": 30599,
+ "ֹ": 30600,
+ "☺": 30601,
+ "ケ": 30602,
+ "역": 30603,
+ "际": 30604,
+ "받": 30605,
+ "望": 30606,
+ "帝": 30607,
+ "减": 30608,
+ "두": 30609,
+ "领": 30610,
+ "": 30611,
+ "钟": 30612,
+ "ガ": 30613,
+ "架": 30614,
+ "든": 30615,
+ "ல": 30616,
+ "松": 30617,
+ "□": 30618,
+ "越": 30619,
+ "答": 30620,
+ "ɕ": 30621,
+ "ῦ": 30622,
+ "染": 30623,
+ "": 30624,
+ "质": 30625,
+ "顺": 30626,
+ "气": 30627,
+ "╗": 30628,
+ "計": 30629,
+ "ქ": 30630,
+ "亮": 30631,
+ "🤦": 30632,
+ "̂": 30633,
+ "ٹ": 30634,
+ "座": 30635,
+ "ˌ": 30636,
+ "均": 30637,
+ "\u000b": 30638,
+ "官": 30639,
+ "适": 30640,
+ "护": 30641,
+ "久": 30642,
+ "春": 30643,
+ "曹": 30644,
+ "皇": 30645,
+ "脚": 30646,
+ "池": 30647,
+ "延": 30648,
+ "키": 30649,
+ "품": 30650,
+ "現": 30651,
+ "檔": 30652,
+ "ば": 30653,
+ "ⴰ": 30654,
+ "希": 30655,
+ "玩": 30656,
+ "固": 30657,
+ "黄": 30658,
+ "": 30659,
+ "☽": 30660,
+ "银": 30661,
+ "\u0003": 30662,
+ "┃": 30663,
+ "👏": 30664,
+ "불": 30665,
+ "攻": 30666,
+ "へ": 30667,
+ "决": 30668,
+ "⊙": 30669,
+ "宁": 30670,
+ "च": 30671,
+ "機": 30672,
+ "義": 30673,
+ "ɲ": 30674,
+ "\u0015": 30675,
+ "했": 30676,
+ "ẩ": 30677,
+ "愛": 30678,
+ "矩": 30679,
+ "패": 30680,
+ "ặ": 30681,
+ "郎": 30682,
+ "Ь": 30683,
+ "绘": 30684,
+ "负": 30685,
+ "ổ": 30686,
+ "ய": 30687,
+ "汉": 30688,
+ "編": 30689,
+ "ێ": 30690,
+ "്": 30691,
+ "じ": 30692,
+ "카": 30693,
+ "似": 30694,
+ "ں": 30695,
+ "や": 30696,
+ "認": 30697,
+ "\u000f": 30698,
+ "過": 30699,
+ "통": 30700,
+ "▪": 30701,
+ "约": 30702,
+ "香": 30703,
+ "买": 30704,
+ "住": 30705,
+ "╚": 30706,
+ "😁": 30707,
+ "扩": 30708,
+ "静": 30709,
+ "려": 30710,
+ "학": 30711,
+ "钥": 30712,
+ "증": 30713,
+ "ỉ": 30714,
+ "她": 30715,
+ "食": 30716,
+ "往": 30717,
+ "點": 30718,
+ "偏": 30719,
+ "康": 30720,
+ "\u0014": 30721,
+ "į": 30722,
+ "준": 30723,
+ "\u0004": 30724,
+ "ฟ": 30725,
+ "♣": 30726,
+ "戏": 30727,
+ "ʂ": 30728,
+ "井": 30729,
+ "军": 30730,
+ "爱": 30731,
+ "ٱ": 30732,
+ "七": 30733,
+ "차": 30734,
+ "币": 30735,
+ "♠": 30736,
+ "哈": 30737,
+ "阅": 30738,
+ "介": 30739,
+ "观": 30740,
+ "區": 30741,
+ "˜": 30742,
+ "ً": 30743,
+ "又": 30744,
+ "冲": 30745,
+ "朝": 30746,
+ "姓": 30747,
+ "课": 30748,
+ "龍": 30749,
+ "각": 30750,
+ "∈": 30751,
+ "米": 30752,
+ "ƒ": 30753,
+ "喜": 30754,
+ "夜": 30755,
+ "团": 30756,
+ "⇒": 30757,
+ "远": 30758,
+ "\u001a": 30759,
+ "ὐ": 30760,
+ "承": 30761,
+ "ಿ": 30762,
+ "室": 30763,
+ "ʀ": 30764,
+ "ង": 30765,
+ "अ": 30766,
+ "罗": 30767,
+ "🙏": 30768,
+ "软": 30769,
+ "🟡": 30770,
+ "건": 30771,
+ "؟": 30772,
+ "း": 30773,
+ "ᴇ": 30774,
+ "ユ": 30775,
+ "토": 30776,
+ "策": 30777,
+ "̄": 30778,
+ "국": 30779,
+ "ֶ": 30780,
+ "协": 30781,
+ "营": 30782,
+ "関": 30783,
+ "吉": 30784,
+ "💀": 30785,
+ "奇": 30786,
+ "滚": 30787,
+ "轴": 30788,
+ "処": 30789,
+ "土": 30790,
+ "划": 30791,
+ "ड": 30792,
+ "临": 30793,
+ "ֵ": 30794,
+ "航": 30795,
+ "浏": 30796,
+ "ゴ": 30797,
+ "別": 30798,
+ "寺": 30799,
+ "於": 30800,
+ "進": 30801,
+ "ὸ": 30802,
+ "風": 30803,
+ "ன": 30804,
+ "班": 30805,
+ "◼": 30806,
+ "九": 30807,
+ "̥": 30808,
+ "號": 30809,
+ "류": 30810,
+ "础": 30811,
+ "般": 30812,
+ "︙": 30813,
+ "̈": 30814,
+ "番": 30815,
+ "✨": 30816,
+ "😎": 30817,
+ "ো": 30818,
+ "😍": 30819,
+ "單": 30820,
+ "帧": 30821,
+ "授": 30822,
+ "赋": 30823,
+ "巴": 30824,
+ "占": 30825,
+ "假": 30826,
+ "ṅ": 30827,
+ "透": 30828,
+ "項": 30829,
+ "ħ": 30830,
+ "馬": 30831,
+ "🟢": 30832,
+ "Ľ": 30833,
+ "լ": 30834,
+ "券": 30835,
+ "같": 30836,
+ "類": 30837,
+ "對": 30838,
+ "월": 30839,
+ "激": 30840,
+ "\u0017": 30841,
+ "戦": 30842,
+ "独": 30843,
+ "訊": 30844,
+ "ិ": 30845,
+ "套": 30846,
+ "ʷ": 30847,
+ "跟": 30848,
+ "ở": 30849,
+ "渲": 30850,
+ "顯": 30851,
+ "降": 30852,
+ "ာ": 30853,
+ "尼": 30854,
+ "血": 30855,
+ "언": 30856,
+ "牛": 30857,
+ "將": 30858,
+ "ศ": 30859,
+ "拍": 30860,
+ "刻": 30861,
+ "ზ": 30862,
+ "╔": 30863,
+ "藤": 30864,
+ "్": 30865,
+ "ῶ": 30866,
+ "🟠": 30867,
+ "良": 30868,
+ "김": 30869,
+ "দ": 30870,
+ "Ṣ": 30871,
+ "録": 30872,
+ "伊": 30873,
+ "落": 30874,
+ "雄": 30875,
+ "雪": 30876,
+ "映": 30877,
+ "著": 30878,
+ "른": 30879,
+ "ფ": 30880,
+ "対": 30881,
+ "智": 30882,
+ "译": 30883,
+ "┬": 30884,
+ "抽": 30885,
+ "ῖ": 30886,
+ "酒": 30887,
+ "Ћ": 30888,
+ "股": 30889,
+ "់": 30890,
+ "순": 30891,
+ "직": 30892,
+ "भ": 30893,
+ "谷": 30894,
+ "물": 30895,
+ "ǒ": 30896,
+ "⠄": 30897,
+ "热": 30898,
+ "終": 30899,
+ "夹": 30900,
+ "干": 30901,
+ "彩": 30902,
+ "敗": 30903,
+ "ќ": 30904,
+ "♯": 30905,
+ "̣": 30906,
+ "վ": 30907,
+ "轮": 30908,
+ "阵": 30909,
+ "夏": 30910,
+ "幕": 30911,
+ "吧": 30912,
+ "港": 30913,
+ "益": 30914,
+ "儿": 30915,
+ "액": 30916,
+ "售": 30917,
+ "兵": 30918,
+ "惠": 30919,
+ "欢": 30920,
+ "": 30921,
+ "零": 30922,
+ "學": 30923,
+ "": 30924,
+ "員": 30925,
+ "ỗ": 30926,
+ "玉": 30927,
+ "逻": 30928,
+ "᥀": 30929,
+ "吗": 30930,
+ "沒": 30931,
+ "≠": 30932,
+ "너": 30933,
+ "ச": 30934,
+ "\u0016": 30935,
+ "夫": 30936,
+ "წ": 30937,
+ "堂": 30938,
+ "電": 30939,
+ "≡": 30940,
+ "陆": 30941,
+ "져": 30942,
+ "研": 30943,
+ "荐": 30944,
+ "健": 30945,
+ "碼": 30946,
+ "练": 30947,
+ "検": 30948,
+ "송": 30949,
+ "ै": 30950,
+ "哪": 30951,
+ "圆": 30952,
+ "Ա": 30953,
+ "↩": 30954,
+ "托": 30955,
+ "̪": 30956,
+ "ू": 30957,
+ "缀": 30958,
+ "네": 30959,
+ "沙": 30960,
+ "兴": 30961,
+ "病": 30962,
+ "\u0007": 30963,
+ "ល": 30964,
+ "ừ": 30965,
+ "Ἀ": 30966,
+ "강": 30967,
+ "항": 30968,
+ "\u0019": 30969,
+ "換": 30970,
+ "温": 30971,
+ "帖": 30972,
+ "ទ": 30973,
+ "込": 30974,
+ "削": 30975,
+ "알": 30976,
+ "征": 30977,
+ "习": 30978,
+ "법": 30979,
+ "栈": 30980,
+ "绝": 30981,
+ "": 30982,
+ "ڕ": 30983,
+ "圖": 30984,
+ "苏": 30985,
+ "発": 30986,
+ "ု": 30987,
+ "町": 30988,
+ "互": 30989,
+ "়": 30990,
+ "ც": 30991,
+ "守": 30992,
+ "새": 30993,
+ "侧": 30994,
+ "草": 30995,
+ "ས": 30996,
+ "扫": 30997,
+ "‒": 30998,
+ "恢": 30999,
+ "ң": 31000,
+ "ण": 31001,
+ "ற": 31002,
+ "째": 31003,
+ "්": 31004,
+ "拟": 31005,
+ "派": 31006,
+ "🏽": 31007,
+ "呼": 31008,
+ "": 31009,
+ "演": 31010,
+ "究": 31011,
+ "교": 31012,
+ "ɣ": 31013,
+ "ए": 31014,
+ "ី": 31015,
+ "ף": 31016,
+ "富": 31017,
+ "駅": 31018,
+ "ず": 31019,
+ "♪": 31020,
+ "😆": 31021,
+ "접": 31022,
+ "ғ": 31023,
+ "▓": 31024,
+ "존": 31025,
+ "ಾ": 31026,
+ "旋": 31027,
+ "ゃ": 31028,
+ "补": 31029,
+ "ץ": 31030,
+ "門": 31031,
+ "ច": 31032,
+ "날": 31033,
+ "ภ": 31034,
+ "ག": 31035,
+ "傳": 31036,
+ "∆": 31037,
+ "": 31038,
+ "ׁ": 31039,
+ "缺": 31040,
+ "頭": 31041,
+ "怪": 31042,
+ "組": 31043,
+ "별": 31044,
+ "Ъ": 31045,
+ "發": 31046,
+ "雷": 31047,
+ "ರ": 31048,
+ "ซ": 31049,
+ "び": 31050,
+ "翻": 31051,
+ "ھ": 31052,
+ "პ": 31053,
+ "題": 31054,
+ "居": 31055,
+ "집": 31056,
+ "🌍": 31057,
+ "˚": 31058,
+ "避": 31059,
+ "줄": 31060,
+ "ុ": 31061,
+ "滑": 31062,
+ "故": 31063,
+ "ญ": 31064,
+ "〜": 31065,
+ "ನ": 31066,
+ "양": 31067,
+ "완": 31068,
+ "ள": 31069,
+ "倍": 31070,
+ "宗": 31071,
+ "択": 31072,
+ "브": 31073,
+ "ɴ": 31074,
+ "効": 31075,
+ "尺": 31076,
+ "視": 31077,
+ "ẽ": 31078,
+ "覆": 31079,
+ "ध": 31080,
+ "骨": 31081,
+ "달": 31082,
+ "ᴛ": 31083,
+ "蓝": 31084,
+ "關": 31085,
+ "額": 31086,
+ "Õ": 31087,
+ "∗": 31088,
+ "卷": 31089,
+ "갑": 31090,
+ "르": 31091,
+ "众": 31092,
+ "ᴀ": 31093,
+ "態": 31094,
+ "ٰ": 31095,
+ "暗": 31096,
+ "君": 31097,
+ "錯": 31098,
+ "ɒ": 31099,
+ "យ": 31100,
+ "ḫ": 31101,
+ "ῆ": 31102,
+ "亚": 31103,
+ "♡": 31104,
+ "割": 31105,
+ "鼠": 31106,
+ "̶": 31107,
+ "Ë": 31108,
+ "読": 31109,
+ "격": 31110,
+ "ゲ": 31111,
+ "眼": 31112,
+ "Ý": 31113,
+ "ژ": 31114,
+ "雨": 31115,
+ "宮": 31116,
+ "쪽": 31117,
+ "ष": 31118,
+ "複": 31119,
+ "剩": 31120,
+ "早": 31121,
+ "杂": 31122,
+ "焦": 31123,
+ "贝": 31124,
+ "突": 31125,
+ "워": 31126,
+ "另": 31127,
+ "摄": 31128,
+ "\b": 31129,
+ "": 31130,
+ "府": 31131,
+ "외": 31132,
+ "盖": 31133,
+ "\u001c": 31134,
+ "ษ": 31135,
+ "佛": 31136,
+ "概": 31137,
+ "與": 31138,
+ "經": 31139,
+ "-": 31140,
+ "һ": 31141,
+ "問": 31142,
+ "ು": 31143,
+ "ἰ": 31144,
+ "話": 31145,
+ "倒": 31146,
+ "葛": 31147,
+ "べ": 31148,
+ "ろ": 31149,
+ "\u001e": 31150,
+ "।": 31151,
+ "ေ": 31152,
+ "ᴏ": 31153,
+ "训": 31154,
+ "體": 31155,
+ "👌": 31156,
+ "內": 31157,
+ "က": 31158,
+ "企": 31159,
+ "약": 31160,
+ "찾": 31161,
+ "ོ": 31162,
+ "破": 31163,
+ "輸": 31164,
+ "림": 31165,
+ "塔": 31166,
+ "턴": 31167,
+ "杀": 31168,
+ "』": 31169,
+ "味": 31170,
+ "浮": 31171,
+ "┆": 31172,
+ "ġ": 31173,
+ "郡": 31174,
+ "┐": 31175,
+ "『": 31176,
+ "阶": 31177,
+ "雅": 31178,
+ "┈": 31179,
+ "园": 31180,
+ ".": 31181,
+ "吃": 31182,
+ "남": 31183,
+ " ": 31184,
+ "ར": 31185,
+ "帮": 31186,
+ "毛": 31187,
+ "耗": 31188,
+ "举": 31189,
+ "ర": 31190,
+ "拿": 31191,
+ "밀": 31192,
+ "ご": 31193,
+ "够": 31194,
+ "礼": 31195,
+ "ព": 31196,
+ "ね": 31197,
+ "": 31198,
+ "兰": 31199,
+ "❌": 31200,
+ "折": 31201,
+ "십": 31202,
+ "💎": 31203,
+ "業": 31204,
+ "诸": 31205,
+ "孙": 31206,
+ "བ": 31207,
+ "😳": 31208,
+ "種": 31209,
+ "Ï": 31210,
+ "ึ": 31211,
+ "": 31212,
+ "医": 31213,
+ "拼": 31214,
+ "↵": 31215,
+ "⅓": 31216,
+ "\u001f": 31217,
+ "မ": 31218,
+ "叫": 31219,
+ "জ": 31220,
+ "予": 31221,
+ "寸": 31222,
+ "梅": 31223,
+ "醒": 31224,
+ "津": 31225,
+ "န": 31226,
+ "ి": 31227,
+ "厂": 31228,
+ "屋": 31229,
+ "ख": 31230,
+ "師": 31231,
+ "👀": 31232,
+ "ỏ": 31233,
+ "ヤ": 31234,
+ "ὰ": 31235,
+ "\u001d": 31236,
+ "◆": 31237,
+ "ដ": 31238,
+ "材": 31239,
+ "ホ": 31240,
+ "張": 31241,
+ "洞": 31242,
+ "餐": 31243,
+ "천": 31244,
+ "হ": 31245,
+ "達": 31246,
+ "們": 31247,
+ "斗": 31248,
+ "横": 31249,
+ "백": 31250,
+ "ំ": 31251,
+ "ۆ": 31252,
+ "말": 31253,
+ "গ": 31254,
+ "佳": 31255,
+ "랜": 31256,
+ "仁": 31257,
+ "陈": 31258,
+ "飞": 31259,
+ "极": 31260,
+ "": 31261,
+ "및": 31262,
+ "仓": 31263,
+ "⬛": 31264,
+ "昌": 31265,
+ "錢": 31266,
+ "殊": 31267,
+ "┴": 31268,
+ "○": 31269,
+ "길": 31270,
+ "泉": 31271,
+ "甲": 31272,
+ "활": 31273,
+ "ひ": 31274,
+ "শ": 31275,
+ "ን": 31276,
+ "Ť": 31277,
+ "ღ": 31278,
+ "皮": 31279,
+ "強": 31280,
+ "赛": 31281,
+ "ా": 31282,
+ "預": 31283,
+ "င": 31284,
+ "튼": 31285,
+ "플": 31286,
+ "ყ": 31287,
+ "⋆": 31288,
+ "ք": 31289,
+ "ા": 31290,
+ "尚": 31291,
+ "또": 31292,
+ "բ": 31293,
+ "┌": 31294,
+ "節": 31295,
+ "森": 31296,
+ "आ": 31297,
+ "办": 31298,
+ "園": 31299,
+ "牙": 31300,
+ "庆": 31301,
+ "隆": 31302,
+ "😔": 31303,
+ "叉": 31304,
+ "գ": 31305,
+ "피": 31306,
+ "ギ": 31307,
+ "啊": 31308,
+ "続": 31309,
+ "灵": 31310,
+ "ヒ": 31311,
+ "忽": 31312,
+ "ʌ": 31313,
+ "량": 31314,
+ "油": 31315,
+ "讯": 31316,
+ "ⵉ": 31317,
+ "릭": 31318,
+ "刚": 31319,
+ "氏": 31320,
+ "ိ": 31321,
+ "Ī": 31322,
+ "誤": 31323,
+ "齐": 31324,
+ "末": 31325,
+ "🙌": 31326,
+ "̞": 31327,
+ "圈": 31328,
+ "念": 31329,
+ "숫": 31330,
+ "毫": 31331,
+ "當": 31332,
+ "規": 31333,
+ "판": 31334,
+ "ు": 31335,
+ "旧": 31336,
+ "卖": 31337,
+ "ฉ": 31338,
+ "幸": 31339,
+ "署": 31340,
+ "근": 31341,
+ "ই": 31342,
+ "岛": 31343,
+ "դ": 31344,
+ "觉": 31345,
+ "害": 31346,
+ "毕": 31347,
+ "ฐ": 31348,
+ "威": 31349,
+ "育": 31350,
+ "呢": 31351,
+ "峰": 31352,
+ "职": 31353,
+ "陽": 31354,
+ "ි": 31355,
+ "亞": 31356,
+ "ұ": 31357,
+ "₃": 31358,
+ "따": 31359,
+ "施": 31360,
+ "泰": 31361,
+ "載": 31362,
+ "
": 31363,
+ "笑": 31364,
+ "華": 31365,
+ "迎": 31366,
+ "됩": 31367,
+ "豆": 31368,
+ "嘉": 31369,
+ "🤡": 31370,
+ "ĕ": 31371,
+ "庄": 31372,
+ "級": 31373,
+ "Ψ": 31374,
+ "ི": 31375,
+ "気": 31376,
+ "责": 31377,
+ "հ": 31378,
+ "អ": 31379,
+ "乱": 31380,
+ "休": 31381,
+ "約": 31382,
+ "ฆ": 31383,
+ "∑": 31384,
+ "察": 31385,
+ "온": 31386,
+ "😬": 31387,
+ "ড": 31388,
+ "乘": 31389,
+ "람": 31390,
+ "इ": 31391,
+ "Ά": 31392,
+ "ந": 31393,
+ "ើ": 31394,
+ "亲": 31395,
+ "េ": 31396,
+ "委": 31397,
+ "赤": 31398,
+ "됨": 31399,
+ "勝": 31400,
+ "怎": 31401,
+ "감": 31402,
+ "宋": 31403,
+ "調": 31404,
+ "짜": 31405,
+ "ী": 31406,
+ "难": 31407,
+ "못": 31408,
+ "티": 31409,
+ "備": 31410,
+ "塞": 31411,
+ "វ": 31412,
+ "险": 31413,
+ "旅": 31414,
+ "虚": 31415,
+ "↳": 31416,
+ "笔": 31417,
+ "馆": 31418,
+ "Қ": 31419,
+ "⚡": 31420,
+ "ೆ": 31421,
+ "※": 31422,
+ "唐": 31423,
+ "律": 31424,
+ "稍": 31425,
+ "散": 31426,
+ "ર": 31427,
+ "ヴ": 31428,
+ "副": 31429,
+ "尽": 31430,
+ "挂": 31431,
+ "県": 31432,
+ "⚠": 31433,
+ "洋": 31434,
+ "鬼": 31435,
+ "암": 31436,
+ "孩": 31437,
+ "℃": 31438,
+ "並": 31439,
+ "ց": 31440,
+ "ូ": 31441,
+ "ℓ": 31442,
+ "ⵏ": 31443,
+ "扣": 31444,
+ "铁": 31445,
+ "闻": 31446,
+ "ˆ": 31447,
+ "戳": 31448,
+ "む": 31449,
+ "秀": 31450,
+ "細": 31451,
+ "ပ": 31452,
+ "御": 31453,
+ "拖": 31454,
+ "좌": 31455,
+ "ؤ": 31456,
+ "绍": 31457,
+ "ỹ": 31458,
+ "참": 31459,
+ "향": 31460,
+ "Ď": 31461,
+ "끝": 31462,
+ "민": 31463,
+ "ძ": 31464,
+ "贵": 31465,
+ "纪": 31466,
+ "秋": 31467,
+ "ಕ": 31468,
+ "ӏ": 31469,
+ "網": 31470,
+ "铺": 31471,
+ "恋": 31472,
+ "fl": 31473,
+ "兼": 31474,
+ "羽": 31475,
+ "창": 31476,
+ "啟": 31477,
+ "弟": 31478,
+ "년": 31479,
+ "慢": 31480,
+ "효": 31481,
+ "許": 31482,
+ "硬": 31483,
+ "잘": 31484,
+ "템": 31485,
+ "્": 31486,
+ "න": 31487,
+ "術": 31488,
+ "ڈ": 31489,
+ "溪": 31490,
+ "": 31491,
+ "暴": 31492,
+ "混": 31493,
+ "夢": 31494,
+ "랑": 31495,
+ "আ": 31496,
+ "還": 31497,
+ "探": 31498,
+ "祖": 31499,
+ "织": 31500,
+ "軍": 31501,
+ "թ": 31502,
+ "務": 31503,
+ "艺": 31504,
+ "ད": 31505,
+ "ት": 31506,
+ "ṁ": 31507,
+ "應": 31508,
+ "擇": 31509,
+ "🥰": 31510,
+ "ķ": 31511,
+ "渡": 31512,
+ "葉": 31513,
+ "령": 31514,
+ "決": 31515,
+ "刀": 31516,
+ "從": 31517,
+ "變": 31518,
+ "올": 31519,
+ "💪": 31520,
+ "灣": 31521,
+ "ር": 31522,
+ "평": 31523,
+ "衣": 31524,
+ "😄": 31525,
+ "ി": 31526,
+ "ჩ": 31527,
+ "ὁ": 31528,
+ "ほ": 31529,
+ "Û": 31530,
+ "চ": 31531,
+ "ර": 31532,
+ "製": 31533,
+ "隊": 31534,
+ "₱": 31535,
+ "纳": 31536,
+ "赖": 31537,
+ "农": 31538,
+ "桥": 31539,
+ "ỳ": 31540,
+ "🏾": 31541,
+ "阻": 31542,
+ "ជ": 31543,
+ "秘": 31544,
+ "박": 31545,
+ "伤": 31546,
+ "稿": 31547,
+ "ం": 31548,
+ "拦": 31549,
+ "넣": 31550,
+ "💕": 31551,
+ "₁": 31552,
+ "宿": 31553,
+ "錄": 31554,
+ "镜": 31555,
+ "채": 31556,
+ "Ə": 31557,
+ "ང": 31558,
+ "⇔": 31559,
+ "☼": 31560,
+ "ུ": 31561,
+ "党": 31562,
+ "급": 31563,
+ "洲": 31564,
+ "ղ": 31565,
+ "說": 31566,
+ "ĭ": 31567,
+ "尝": 31568,
+ "담": 31569,
+ "फ": 31570,
+ "哥": 31571,
+ "圣": 31572,
+ "萨": 31573,
+ "😏": 31574,
+ "ʏ": 31575,
+ "ெ": 31576,
+ "丁": 31577,
+ "虎": 31578,
+ "권": 31579,
+ "善": 31580,
+ "岩": 31581,
+ "커": 31582,
+ "◦": 31583,
+ "抛": 31584,
+ "석": 31585,
+ "Έ": 31586,
+ "宣": 31587,
+ "拳": 31588,
+ "팅": 31589,
+ "枚": 31590,
+ "洛": 31591,
+ "証": 31592,
+ "陵": 31593,
+ "佐": 31594,
+ "館": 31595,
+ "누": 31596,
+ "돌": 31597,
+ "₄": 31598,
+ "稱": 31599,
+ "聊": 31600,
+ "車": 31601,
+ "루": 31602,
+ "״": 31603,
+ "ಠ": 31604,
+ "庫": 31605,
+ "མ": 31606,
+ "統": 31607,
+ "련": 31608,
+ "़": 31609,
+ "ṯ": 31610,
+ "ക": 31611,
+ "旗": 31612,
+ "励": 31613,
+ "紀": 31614,
+ "忠": 31615,
+ "າ": 31616,
+ "杨": 31617,
+ "丹": 31618,
+ "Ù": 31619,
+ "ฝ": 31620,
+ "却": 31621,
+ "舞": 31622,
+ "轉": 31623,
+ "တ": 31624,
+ "丽": 31625,
+ "借": 31626,
+ "ා": 31627,
+ "ょ": 31628,
+ "옵": 31629,
+ "편": 31630,
+ "蒙": 31631,
+ "衡": 31632,
+ "ʋ": 31633,
+ "叶": 31634,
+ "̇": 31635,
+ "⬜": 31636,
+ "🇺": 31637,
+ "Հ": 31638,
+ "谢": 31639,
+ "Ą": 31640,
+ "ே": 31641,
+ "ằ": 31642,
+ "既": 31643,
+ "济": 31644,
+ "≯": 31645,
+ "準": 31646,
+ "답": 31647,
+ "ಲ": 31648,
+ "残": 31649,
+ "虑": 31650,
+ "̆": 31651,
+ "┘": 31652,
+ "急": 31653,
+ "招": 31654,
+ "막": 31655,
+ "≮": 31656,
+ "產": 31657,
+ "Ṭ": 31658,
+ "😢": 31659,
+ "垂": 31660,
+ "親": 31661,
+ "ģ": 31662,
+ "־": 31663,
+ "猫": 31664,
+ "ʟ": 31665,
+ "☃": 31666,
+ "✪": 31667,
+ "刪": 31668,
+ "胡": 31669,
+ "☉": 31670,
+ "晚": 31671,
+ "군": 31672,
+ "승": 31673,
+ "న": 31674,
+ "ὴ": 31675,
+ "曾": 31676,
+ "論": 31677,
+ "ɯ": 31678,
+ "త": 31679,
+ "戰": 31680,
+ "鱼": 31681,
+ "ǧ": 31682,
+ "寶": 31683,
+ "특": 31684,
+ "💯": 31685,
+ "崎": 31686,
+ "甘": 31687,
+ "該": 31688,
+ "링": 31689,
+ "😡": 31690,
+ "उ": 31691,
+ "ែ": 31692,
+ "頁": 31693,
+ "큰": 31694,
+ "➤": 31695,
+ "총": 31696,
+ "💰": 31697,
+ "∂": 31698,
+ "毁": 31699,
+ "聖": 31700,
+ "麻": 31701,
+ "ʐ": 31702,
+ "敏": 31703,
+ "運": 31704,
+ "될": 31705,
+ "쓰": 31706,
+ "ಸ": 31707,
+ "စ": 31708,
+ "✦": 31709,
+ "젝": 31710,
+ "復": 31711,
+ "寻": 31712,
+ "茶": 31713,
+ "ਾ": 31714,
+ "竹": 31715,
+ "遇": 31716,
+ "順": 31717,
+ "며": 31718,
+ "累": 31719,
+ "ĝ": 31720,
+ "ˇ": 31721,
+ "覧": 31722,
+ "এ": 31723,
+ "株": 31724,
+ "취": 31725,
+ "ስ": 31726,
+ "争": 31727,
+ "势": 31728,
+ "宇": 31729,
+ "橋": 31730,
+ "Ӏ": 31731,
+ "堆": 31732,
+ "ⵙ": 31733,
+ "丶": 31734,
+ "棋": 31735,
+ "肉": 31736,
+ "የ": 31737,
+ "": 31738,
+ "❶": 31739,
+ "季": 31740,
+ "ል": 31741,
+ "殿": 31742,
+ "優": 31743,
+ "試": 31744,
+ "첫": 31745,
+ "Ό": 31746,
+ "戶": 31747,
+ "ண": 31748,
+ "羅": 31749,
+ "桃": 31750,
+ "립": 31751,
+ "浪": 31752,
+ "脑": 31753,
+ "😛": 31754,
+ "弃": 31755,
+ "炮": 31756,
+ "轻": 31757,
+ "울": 31758,
+ "": 31759,
+ "ヘ": 31760,
+ "奥": 31761,
+ "💜": 31762,
+ "忘": 31763,
+ "遠": 31764,
+ "飛": 31765,
+ "魏": 31766,
+ "Ē": 31767,
+ "汇": 31768,
+ "央": 31769,
+ "逆": 31770,
+ "露": 31771,
+ "須": 31772,
+ "ѐ": 31773,
+ "ḷ": 31774,
+ "ದ": 31775,
+ "✭": 31776,
+ "寄": 31777,
+ "盟": 31778,
+ "财": 31779,
+ "際": 31780,
+ "ἔ": 31781,
+ "ǫ": 31782,
+ "थ": 31783,
+ "ാ": 31784,
+ "宫": 31785,
+ "巨": 31786,
+ "途": 31787,
+ "ʹ": 31788,
+ "ಗ": 31789,
+ "帐": 31790,
+ "": 31791,
+ "拒": 31792,
+ "药": 31793,
+ "🙃": 31794,
+ "ŕ": 31795,
+ "亡": 31796,
+ "壁": 31797,
+ "ም": 31798,
+ "參": 31799,
+ "😩": 31800,
+ "շ": 31801,
+ "ವ": 31802,
+ "ណ": 31803,
+ "丰": 31804,
+ "獲": 31805,
+ "莉": 31806,
+ "좋": 31807,
+ "ရ": 31808,
+ "₦": 31809,
+ "겠": 31810,
+ "👉": 31811,
+ "吴": 31812,
+ "岡": 31813,
+ "诉": 31814,
+ "읽": 31815,
+ "🥺": 31816,
+ "爆": 31817,
+ "🇸": 31818,
+ "ভ": 31819,
+ "迭": 31820,
+ "엔": 31821,
+ "ἄ": 31822,
+ "捷": 31823,
+ "納": 31824,
+ "邀": 31825,
+ "ಯ": 31826,
+ "爾": 31827,
+ "船": 31828,
+ "赞": 31829,
+ "胜": 31830,
+ "므": 31831,
+ "သ": 31832,
+ "構": 31833,
+ "磁": 31834,
+ "冰": 31835,
+ "딩": 31836,
+ "ે": 31837,
+ "媒": 31838,
+ "繁": 31839,
+ "☠": 31840,
+ "❒": 31841,
+ "仪": 31842,
+ "렬": 31843,
+ "昭": 31844,
+ "珠": 31845,
+ "離": 31846,
+ "ན": 31847,
+ "ల": 31848,
+ "ತ": 31849,
+ "拷": 31850,
+ "粉": 31851,
+ "벤": 31852,
+ "⇽": 31853,
+ "乌": 31854,
+ "拥": 31855,
+ "ҳ": 31856,
+ "ය": 31857,
+ "ེ": 31858,
+ "仙": 31859,
+ "塊": 31860,
+ "幅": 31861,
+ "🎉": 31862,
+ "Մ": 31863,
+ "跨": 31864,
+ "ٔ": 31865,
+ "恩": 31866,
+ "损": 31867,
+ "养": 31868,
+ "奈": 31869,
+ "ǀ": 31870,
+ "严": 31871,
+ "卫": 31872,
+ "迟": 31873,
+ "様": 31874,
+ "裡": 31875,
+ "난": 31876,
+ "았": 31877,
+ "͜": 31878,
+ "Ζ": 31879,
+ "ਰ": 31880,
+ "պ": 31881,
+ "ং": 31882,
+ "丢": 31883,
+ "伝": 31884,
+ "컨": 31885,
+ "ව": 31886,
+ "ြ": 31887,
+ "冷": 31888,
+ "遗": 31889,
+ "銀": 31890,
+ "̌": 31891,
+ "ᴜ": 31892,
+ "瑞": 31893,
+ "ฌ": 31894,
+ "❍": 31895,
+ "ふ": 31896,
+ "聚": 31897,
+ "碎": 31898,
+ "衛": 31899,
+ "অ": 31900,
+ "ញ": 31901,
+ "퍼": 31902,
+ "Ս": 31903,
+ "ນ": 31904,
+ "ẓ": 31905,
+ "✌": 31906,
+ "孝": 31907,
+ "陳": 31908,
+ "히": 31909,
+ "ක": 31910,
+ "黒": 31911,
+ "💖": 31912,
+ "ḩ": 31913,
+ "応": 31914,
+ "饰": 31915,
+ "∪": 31916,
+ "宜": 31917,
+ "樂": 31918,
+ "則": 31919,
+ "勇": 31920,
+ "徐": 31921,
+ "ⵓ": 31922,
+ "權": 31923,
+ "鲁": 31924,
+ "‟": 31925,
+ "庭": 31926,
+ "苗": 31927,
+ "🔴": 31928,
+ "闲": 31929,
+ "독": 31930,
+ "ɹ": 31931,
+ "ҽ": 31932,
+ "ថ": 31933,
+ "宏": 31934,
+ "尊": 31935,
+ "總": 31936,
+ "裝": 31937,
+ "ම": 31938,
+ "▸": 31939,
+ "測": 31940,
+ "ಮ": 31941,
+ "አ": 31942,
+ "轩": 31943,
+ "兄": 31944,
+ "剑": 31945,
+ "ન": 31946,
+ "朱": 31947,
+ "ǝ": 31948,
+ "Ḩ": 31949,
+ "担": 31950,
+ "灰": 31951,
+ "讲": 31952,
+ "롤": 31953,
+ "︎": 31954,
+ "😤": 31955,
+ "ោ": 31956,
+ "애": 31957,
+ "였": 31958,
+ "질": 31959,
+ "振": 31960,
+ "灯": 31961,
+ "ĉ": 31962,
+ "ස": 31963,
+ "閉": 31964,
+ "램": 31965,
+ "ಂ": 31966,
+ "げ": 31967,
+ "̧": 31968,
+ "狂": 31969,
+ "融": 31970,
+ "仍": 31971,
+ "實": 31972,
+ "楽": 31973,
+ "範": 31974,
+ "ٌ": 31975,
+ "వ": 31976,
+ "嵌": 31977,
+ "摩": 31978,
+ "袁": 31979,
+ "ষ": 31980,
+ "乎": 31981,
+ "규": 31982,
+ "岗": 31983,
+ "糊": 31984,
+ "క": 31985,
+ "雲": 31986,
+ "심": 31987,
+ "ई": 31988,
+ "འ": 31989,
+ "ἡ": 31990,
+ "丝": 31991,
+ "Ħ": 31992,
+ "ٍ": 31993,
+ "ٓ": 31994,
+ "အ": 31995,
+ "執": 31996,
+ "벨": 31997,
+ "ゼ": 31998,
+ "梦": 31999
+ },
+ "merges": [
+ [
+ "▁",
+ "t"
+ ],
+ [
+ "i",
+ "n"
+ ],
+ [
+ "e",
+ "r"
+ ],
+ [
+ "▁",
+ "a"
+ ],
+ [
+ "h",
+ "e"
+ ],
+ [
+ "o",
+ "n"
+ ],
+ [
+ "r",
+ "e"
+ ],
+ [
+ "▁",
+ "s"
+ ],
+ [
+ "e",
+ "n"
+ ],
+ [
+ "a",
+ "t"
+ ],
+ [
+ "o",
+ "r"
+ ],
+ [
+ "▁t",
+ "he"
+ ],
+ [
+ "▁th",
+ "e"
+ ],
+ [
+ "▁",
+ "the"
+ ],
+ [
+ "e",
+ "s"
+ ],
+ [
+ "▁",
+ "w"
+ ],
+ [
+ "a",
+ "n"
+ ],
+ [
+ "▁",
+ "c"
+ ],
+ [
+ "i",
+ "s"
+ ],
+ [
+ "i",
+ "t"
+ ],
+ [
+ "o",
+ "u"
+ ],
+ [
+ "▁",
+ "d"
+ ],
+ [
+ "a",
+ "l"
+ ],
+ [
+ "a",
+ "r"
+ ],
+ [
+ "▁",
+ "p"
+ ],
+ [
+ "▁",
+ "f"
+ ],
+ [
+ "e",
+ "d"
+ ],
+ [
+ "▁",
+ "b"
+ ],
+ [
+ "in",
+ "g"
+ ],
+ [
+ "i",
+ "ng"
+ ],
+ [
+ "▁",
+ "o"
+ ],
+ [
+ "▁",
+ "m"
+ ],
+ [
+ "l",
+ "e"
+ ],
+ [
+ "n",
+ "d"
+ ],
+ [
+ "a",
+ "s"
+ ],
+ [
+ "i",
+ "c"
+ ],
+ [
+ "▁",
+ "h"
+ ],
+ [
+ "io",
+ "n"
+ ],
+ [
+ "i",
+ "on"
+ ],
+ [
+ "▁i",
+ "n"
+ ],
+ [
+ "▁",
+ "in"
+ ],
+ [
+ "▁t",
+ "o"
+ ],
+ [
+ "▁",
+ "to"
+ ],
+ [
+ "e",
+ "t"
+ ],
+ [
+ "o",
+ "m"
+ ],
+ [
+ "e",
+ "l"
+ ],
+ [
+ "▁o",
+ "f"
+ ],
+ [
+ "▁",
+ "of"
+ ],
+ [
+ "s",
+ "t"
+ ],
+ [
+ "▁a",
+ "nd"
+ ],
+ [
+ "▁an",
+ "d"
+ ],
+ [
+ "▁",
+ "and"
+ ],
+ [
+ "▁",
+ "l"
+ ],
+ [
+ "▁t",
+ "h"
+ ],
+ [
+ "▁",
+ "th"
+ ],
+ [
+ "▁",
+ "n"
+ ],
+ [
+ "en",
+ "t"
+ ],
+ [
+ "e",
+ "nt"
+ ],
+ [
+ "i",
+ "l"
+ ],
+ [
+ "c",
+ "t"
+ ],
+ [
+ "r",
+ "o"
+ ],
+ [
+ "▁r",
+ "e"
+ ],
+ [
+ "▁",
+ "re"
+ ],
+ [
+ "i",
+ "d"
+ ],
+ [
+ "a",
+ "m"
+ ],
+ [
+ "▁",
+ "I"
+ ],
+ [
+ "a",
+ "d"
+ ],
+ [
+ "▁",
+ "e"
+ ],
+ [
+ "▁",
+ "S"
+ ],
+ [
+ "▁",
+ "g"
+ ],
+ [
+ "▁",
+ "T"
+ ],
+ [
+ "i",
+ "m"
+ ],
+ [
+ "o",
+ "t"
+ ],
+ [
+ "a",
+ "c"
+ ],
+ [
+ "u",
+ "r"
+ ],
+ [
+ "▁",
+ "("
+ ],
+ [
+ "i",
+ "g"
+ ],
+ [
+ "▁",
+ "="
+ ],
+ [
+ "o",
+ "l"
+ ],
+ [
+ "u",
+ "t"
+ ],
+ [
+ "▁",
+ "A"
+ ],
+ [
+ "s",
+ "e"
+ ],
+ [
+ "▁",
+ "u"
+ ],
+ [
+ "v",
+ "e"
+ ],
+ [
+ "▁",
+ "C"
+ ],
+ [
+ "i",
+ "f"
+ ],
+ [
+ "o",
+ "w"
+ ],
+ [
+ "▁",
+ "y"
+ ],
+ [
+ "c",
+ "h"
+ ],
+ [
+ "a",
+ "y"
+ ],
+ [
+ "▁d",
+ "e"
+ ],
+ [
+ "▁",
+ "de"
+ ],
+ [
+ "▁s",
+ "t"
+ ],
+ [
+ "▁",
+ "st"
+ ],
+ [
+ "▁",
+ "|"
+ ],
+ [
+ "ve",
+ "r"
+ ],
+ [
+ "v",
+ "er"
+ ],
+ [
+ ")",
+ ";"
+ ],
+ [
+ "▁",
+ "\""
+ ],
+ [
+ "l",
+ "y"
+ ],
+ [
+ "▁b",
+ "e"
+ ],
+ [
+ "▁",
+ "be"
+ ],
+ [
+ "*",
+ "*"
+ ],
+ [
+ "▁i",
+ "s"
+ ],
+ [
+ "▁",
+ "is"
+ ],
+ [
+ "o",
+ "d"
+ ],
+ [
+ "▁",
+ "M"
+ ],
+ [
+ "at",
+ "ion"
+ ],
+ [
+ "ati",
+ "on"
+ ],
+ [
+ "atio",
+ "n"
+ ],
+ [
+ "u",
+ "l"
+ ],
+ [
+ "▁f",
+ "or"
+ ],
+ [
+ "▁fo",
+ "r"
+ ],
+ [
+ "▁",
+ "for"
+ ],
+ [
+ "▁o",
+ "n"
+ ],
+ [
+ "▁",
+ "on"
+ ],
+ [
+ "a",
+ "g"
+ ],
+ [
+ "c",
+ "e"
+ ],
+ [
+ "te",
+ "r"
+ ],
+ [
+ "t",
+ "er"
+ ],
+ [
+ "i",
+ "r"
+ ],
+ [
+ "t",
+ "h"
+ ],
+ [
+ "▁",
+ "v"
+ ],
+ [
+ "q",
+ "u"
+ ],
+ [
+ "▁",
+ "B"
+ ],
+ [
+ "e",
+ "m"
+ ],
+ [
+ "▁",
+ "P"
+ ],
+ [
+ "▁y",
+ "ou"
+ ],
+ [
+ "▁yo",
+ "u"
+ ],
+ [
+ "▁",
+ "you"
+ ],
+ [
+ "▁t",
+ "hat"
+ ],
+ [
+ "▁th",
+ "at"
+ ],
+ [
+ "▁",
+ "that"
+ ],
+ [
+ "u",
+ "n"
+ ],
+ [
+ "▁",
+ "{"
+ ],
+ [
+ "it",
+ "h"
+ ],
+ [
+ "i",
+ "th"
+ ],
+ [
+ "r",
+ "i"
+ ],
+ [
+ "es",
+ "t"
+ ],
+ [
+ "e",
+ "st"
+ ],
+ [
+ "a",
+ "b"
+ ],
+ [
+ "-",
+ "-"
+ ],
+ [
+ "a",
+ "p"
+ ],
+ [
+ "▁i",
+ "t"
+ ],
+ [
+ "▁",
+ "it"
+ ],
+ [
+ "▁c",
+ "on"
+ ],
+ [
+ "▁co",
+ "n"
+ ],
+ [
+ "▁",
+ "con"
+ ],
+ [
+ "at",
+ "e"
+ ],
+ [
+ "a",
+ "te"
+ ],
+ [
+ "u",
+ "s"
+ ],
+ [
+ "▁",
+ "H"
+ ],
+ [
+ "u",
+ "m"
+ ],
+ [
+ "▁",
+ "D"
+ ],
+ [
+ "o",
+ "s"
+ ],
+ [
+ "p",
+ "e"
+ ],
+ [
+ "▁",
+ "-"
+ ],
+ [
+ "▁w",
+ "h"
+ ],
+ [
+ "▁",
+ "wh"
+ ],
+ [
+ "▁a",
+ "l"
+ ],
+ [
+ "▁",
+ "al"
+ ],
+ [
+ "▁a",
+ "s"
+ ],
+ [
+ "▁",
+ "as"
+ ],
+ [
+ "an",
+ "d"
+ ],
+ [
+ "a",
+ "nd"
+ ],
+ [
+ "is",
+ "t"
+ ],
+ [
+ "i",
+ "st"
+ ],
+ [
+ "▁",
+ "L"
+ ],
+ [
+ "▁",
+ "W"
+ ],
+ [
+ "▁w",
+ "ith"
+ ],
+ [
+ "▁",
+ "with"
+ ],
+ [
+ "▁a",
+ "n"
+ ],
+ [
+ "▁",
+ "an"
+ ],
+ [
+ "er",
+ "e"
+ ],
+ [
+ "e",
+ "re"
+ ],
+ [
+ "▁",
+ "*"
+ ],
+ [
+ "▁",
+ "R"
+ ],
+ [
+ "▁h",
+ "e"
+ ],
+ [
+ "▁",
+ "he"
+ ],
+ [
+ "▁",
+ "F"
+ ],
+ [
+ "o",
+ "c"
+ ],
+ [
+ "▁w",
+ "as"
+ ],
+ [
+ "▁wa",
+ "s"
+ ],
+ [
+ "▁",
+ "was"
+ ],
+ [
+ "er",
+ "s"
+ ],
+ [
+ "e",
+ "rs"
+ ],
+ [
+ "k",
+ "e"
+ ],
+ [
+ "ou",
+ "t"
+ ],
+ [
+ "o",
+ "ut"
+ ],
+ [
+ "h",
+ "t"
+ ],
+ [
+ "▁",
+ "r"
+ ],
+ [
+ "es",
+ "s"
+ ],
+ [
+ "e",
+ "ss"
+ ],
+ [
+ "o",
+ "p"
+ ],
+ [
+ "re",
+ "s"
+ ],
+ [
+ "r",
+ "es"
+ ],
+ [
+ "i",
+ "e"
+ ],
+ [
+ "▁",
+ "E"
+ ],
+ [
+ "▁",
+ "\\"
+ ],
+ [
+ "▁T",
+ "he"
+ ],
+ [
+ "▁Th",
+ "e"
+ ],
+ [
+ "▁",
+ "The"
+ ],
+ [
+ "en",
+ "d"
+ ],
+ [
+ "e",
+ "nd"
+ ],
+ [
+ "l",
+ "d"
+ ],
+ [
+ "▁",
+ "N"
+ ],
+ [
+ "or",
+ "t"
+ ],
+ [
+ "o",
+ "rt"
+ ],
+ [
+ "▁",
+ "G"
+ ],
+ [
+ "/",
+ "/"
+ ],
+ [
+ "▁",
+ "#"
+ ],
+ [
+ "ou",
+ "r"
+ ],
+ [
+ "o",
+ "ur"
+ ],
+ [
+ "t",
+ "e"
+ ],
+ [
+ "il",
+ "l"
+ ],
+ [
+ "i",
+ "ll"
+ ],
+ [
+ "ai",
+ "n"
+ ],
+ [
+ "a",
+ "in"
+ ],
+ [
+ "▁s",
+ "e"
+ ],
+ [
+ "▁",
+ "se"
+ ],
+ [
+ "▁",
+ "$"
+ ],
+ [
+ "▁p",
+ "ro"
+ ],
+ [
+ "▁pr",
+ "o"
+ ],
+ [
+ "▁",
+ "pro"
+ ],
+ [
+ "or",
+ "e"
+ ],
+ [
+ "o",
+ "re"
+ ],
+ [
+ "▁c",
+ "om"
+ ],
+ [
+ "▁co",
+ "m"
+ ],
+ [
+ "▁",
+ "com"
+ ],
+ [
+ "am",
+ "e"
+ ],
+ [
+ "a",
+ "me"
+ ],
+ [
+ "t",
+ "r"
+ ],
+ [
+ "▁n",
+ "e"
+ ],
+ [
+ "▁",
+ "ne"
+ ],
+ [
+ "ro",
+ "m"
+ ],
+ [
+ "r",
+ "om"
+ ],
+ [
+ "u",
+ "b"
+ ],
+ [
+ "▁a",
+ "t"
+ ],
+ [
+ "▁",
+ "at"
+ ],
+ [
+ "▁e",
+ "x"
+ ],
+ [
+ "▁",
+ "ex"
+ ],
+ [
+ "an",
+ "t"
+ ],
+ [
+ "a",
+ "nt"
+ ],
+ [
+ "u",
+ "e"
+ ],
+ [
+ "▁o",
+ "r"
+ ],
+ [
+ "▁",
+ "or"
+ ],
+ [
+ "▁",
+ "}"
+ ],
+ [
+ "ar",
+ "t"
+ ],
+ [
+ "a",
+ "rt"
+ ],
+ [
+ "ct",
+ "ion"
+ ],
+ [
+ "▁",
+ "k"
+ ],
+ [
+ "p",
+ "t"
+ ],
+ [
+ "n",
+ "t"
+ ],
+ [
+ "i",
+ "v"
+ ],
+ [
+ "d",
+ "e"
+ ],
+ [
+ "▁",
+ "O"
+ ],
+ [
+ "p",
+ "l"
+ ],
+ [
+ "ur",
+ "n"
+ ],
+ [
+ "u",
+ "rn"
+ ],
+ [
+ "ig",
+ "ht"
+ ],
+ [
+ "igh",
+ "t"
+ ],
+ [
+ "i",
+ "ght"
+ ],
+ [
+ "al",
+ "l"
+ ],
+ [
+ "a",
+ "ll"
+ ],
+ [
+ "▁t",
+ "his"
+ ],
+ [
+ "▁th",
+ "is"
+ ],
+ [
+ "▁",
+ "this"
+ ],
+ [
+ "se",
+ "r"
+ ],
+ [
+ "s",
+ "er"
+ ],
+ [
+ "av",
+ "e"
+ ],
+ [
+ "a",
+ "ve"
+ ],
+ [
+ "▁n",
+ "ot"
+ ],
+ [
+ "▁no",
+ "t"
+ ],
+ [
+ "▁",
+ "not"
+ ],
+ [
+ "▁a",
+ "re"
+ ],
+ [
+ "▁ar",
+ "e"
+ ],
+ [
+ "▁",
+ "are"
+ ],
+ [
+ "▁",
+ "j"
+ ],
+ [
+ "▁l",
+ "e"
+ ],
+ [
+ "▁",
+ "le"
+ ],
+ [
+ "i",
+ "z"
+ ],
+ [
+ "▁",
+ "'"
+ ],
+ [
+ "ag",
+ "e"
+ ],
+ [
+ "a",
+ "ge"
+ ],
+ [
+ "me",
+ "nt"
+ ],
+ [
+ "men",
+ "t"
+ ],
+ [
+ "m",
+ "ent"
+ ],
+ [
+ "▁t",
+ "r"
+ ],
+ [
+ "▁",
+ "tr"
+ ],
+ [
+ "ac",
+ "k"
+ ],
+ [
+ "a",
+ "ck"
+ ],
+ [
+ "us",
+ "t"
+ ],
+ [
+ "u",
+ "st"
+ ],
+ [
+ "(",
+ ")"
+ ],
+ [
+ "-",
+ ">"
+ ],
+ [
+ "it",
+ "y"
+ ],
+ [
+ "i",
+ "ty"
+ ],
+ [
+ "in",
+ "e"
+ ],
+ [
+ "i",
+ "ne"
+ ],
+ [
+ "ou",
+ "ld"
+ ],
+ [
+ "oul",
+ "d"
+ ],
+ [
+ "o",
+ "uld"
+ ],
+ [
+ "▁",
+ "J"
+ ],
+ [
+ "o",
+ "g"
+ ],
+ [
+ "▁f",
+ "rom"
+ ],
+ [
+ "▁fr",
+ "om"
+ ],
+ [
+ "▁fro",
+ "m"
+ ],
+ [
+ "▁",
+ "from"
+ ],
+ [
+ "▁w",
+ "e"
+ ],
+ [
+ "▁",
+ "we"
+ ],
+ [
+ "el",
+ "l"
+ ],
+ [
+ "e",
+ "ll"
+ ],
+ [
+ "▁s",
+ "h"
+ ],
+ [
+ "▁",
+ "sh"
+ ],
+ [
+ "▁e",
+ "n"
+ ],
+ [
+ "▁",
+ "en"
+ ],
+ [
+ "ur",
+ "e"
+ ],
+ [
+ "u",
+ "re"
+ ],
+ [
+ "por",
+ "t"
+ ],
+ [
+ "po",
+ "rt"
+ ],
+ [
+ "p",
+ "ort"
+ ],
+ [
+ "▁c",
+ "h"
+ ],
+ [
+ "▁",
+ "ch"
+ ],
+ [
+ "n",
+ "e"
+ ],
+ [
+ "▁b",
+ "y"
+ ],
+ [
+ "▁",
+ "by"
+ ],
+ [
+ "pe",
+ "r"
+ ],
+ [
+ "p",
+ "er"
+ ],
+ [
+ "ar",
+ "d"
+ ],
+ [
+ "a",
+ "rd"
+ ],
+ [
+ "as",
+ "s"
+ ],
+ [
+ "a",
+ "ss"
+ ],
+ [
+ "g",
+ "e"
+ ],
+ [
+ "a",
+ "k"
+ ],
+ [
+ "ar",
+ "e"
+ ],
+ [
+ "a",
+ "re"
+ ],
+ [
+ "o",
+ "k"
+ ],
+ [
+ "a",
+ "v"
+ ],
+ [
+ "iv",
+ "e"
+ ],
+ [
+ "i",
+ "ve"
+ ],
+ [
+ "f",
+ "f"
+ ],
+ [
+ "ie",
+ "s"
+ ],
+ [
+ "i",
+ "es"
+ ],
+ [
+ "at",
+ "h"
+ ],
+ [
+ "a",
+ "th"
+ ],
+ [
+ "tu",
+ "rn"
+ ],
+ [
+ "t",
+ "urn"
+ ],
+ [
+ "▁",
+ "U"
+ ],
+ [
+ "in",
+ "t"
+ ],
+ [
+ "i",
+ "nt"
+ ],
+ [
+ "--",
+ "--"
+ ],
+ [
+ "---",
+ "-"
+ ],
+ [
+ "-",
+ "---"
+ ],
+ [
+ "▁i",
+ "m"
+ ],
+ [
+ "▁",
+ "im"
+ ],
+ [
+ "os",
+ "t"
+ ],
+ [
+ "o",
+ "st"
+ ],
+ [
+ "ia",
+ "l"
+ ],
+ [
+ "i",
+ "al"
+ ],
+ [
+ "▁h",
+ "ave"
+ ],
+ [
+ "▁ha",
+ "ve"
+ ],
+ [
+ "▁hav",
+ "e"
+ ],
+ [
+ "▁",
+ "have"
+ ],
+ [
+ "in",
+ "d"
+ ],
+ [
+ "i",
+ "nd"
+ ],
+ [
+ "i",
+ "p"
+ ],
+ [
+ "an",
+ "s"
+ ],
+ [
+ "a",
+ "ns"
+ ],
+ [
+ "x",
+ "t"
+ ],
+ [
+ "▁d",
+ "o"
+ ],
+ [
+ "▁",
+ "do"
+ ],
+ [
+ "c",
+ "l"
+ ],
+ [
+ "▁i",
+ "f"
+ ],
+ [
+ "▁",
+ "if"
+ ],
+ [
+ "co",
+ "n"
+ ],
+ [
+ "c",
+ "on"
+ ],
+ [
+ "i",
+ "a"
+ ],
+ [
+ "▁h",
+ "is"
+ ],
+ [
+ "▁hi",
+ "s"
+ ],
+ [
+ "▁",
+ "his"
+ ],
+ [
+ "ul",
+ "t"
+ ],
+ [
+ "u",
+ "lt"
+ ],
+ [
+ "ro",
+ "u"
+ ],
+ [
+ "r",
+ "ou"
+ ],
+ [
+ "▁s",
+ "u"
+ ],
+ [
+ "▁",
+ "su"
+ ],
+ [
+ "r",
+ "a"
+ ],
+ [
+ "▁u",
+ "n"
+ ],
+ [
+ "▁",
+ "un"
+ ],
+ [
+ "ab",
+ "le"
+ ],
+ [
+ "abl",
+ "e"
+ ],
+ [
+ "a",
+ "ble"
+ ],
+ [
+ "▁",
+ "<"
+ ],
+ [
+ "▁",
+ "K"
+ ],
+ [
+ "om",
+ "e"
+ ],
+ [
+ "o",
+ "me"
+ ],
+ [
+ "▁q",
+ "u"
+ ],
+ [
+ "▁",
+ "qu"
+ ],
+ [
+ "ge",
+ "t"
+ ],
+ [
+ "g",
+ "et"
+ ],
+ [
+ "▁m",
+ "e"
+ ],
+ [
+ "▁",
+ "me"
+ ],
+ [
+ "as",
+ "t"
+ ],
+ [
+ "a",
+ "st"
+ ],
+ [
+ "ec",
+ "t"
+ ],
+ [
+ "e",
+ "ct"
+ ],
+ [
+ "▁#",
+ "#"
+ ],
+ [
+ "▁",
+ "##"
+ ],
+ [
+ "t",
+ "o"
+ ],
+ [
+ "▁c",
+ "l"
+ ],
+ [
+ "▁",
+ "cl"
+ ],
+ [
+ "▁a",
+ "b"
+ ],
+ [
+ "▁",
+ "ab"
+ ],
+ [
+ "ic",
+ "e"
+ ],
+ [
+ "i",
+ "ce"
+ ],
+ [
+ "ir",
+ "e"
+ ],
+ [
+ "i",
+ "re"
+ ],
+ [
+ "be",
+ "r"
+ ],
+ [
+ "b",
+ "er"
+ ],
+ [
+ "on",
+ "e"
+ ],
+ [
+ "o",
+ "ne"
+ ],
+ [
+ "ic",
+ "h"
+ ],
+ [
+ "i",
+ "ch"
+ ],
+ [
+ "he",
+ "n"
+ ],
+ [
+ "h",
+ "en"
+ ],
+ [
+ "▁c",
+ "an"
+ ],
+ [
+ "▁ca",
+ "n"
+ ],
+ [
+ "▁",
+ "can"
+ ],
+ [
+ "▁T",
+ "h"
+ ],
+ [
+ "▁",
+ "Th"
+ ],
+ [
+ "▁l",
+ "a"
+ ],
+ [
+ "▁",
+ "la"
+ ],
+ [
+ "▁a",
+ "ll"
+ ],
+ [
+ "▁al",
+ "l"
+ ],
+ [
+ "▁",
+ "all"
+ ],
+ [
+ "im",
+ "e"
+ ],
+ [
+ "i",
+ "me"
+ ],
+ [
+ "il",
+ "e"
+ ],
+ [
+ "i",
+ "le"
+ ],
+ [
+ "id",
+ "e"
+ ],
+ [
+ "i",
+ "de"
+ ],
+ [
+ "\"",
+ ","
+ ],
+ [
+ "▁p",
+ "l"
+ ],
+ [
+ "▁",
+ "pl"
+ ],
+ [
+ "▁",
+ "V"
+ ],
+ [
+ "r",
+ "u"
+ ],
+ [
+ "or",
+ "m"
+ ],
+ [
+ "o",
+ "rm"
+ ],
+ [
+ "▁h",
+ "ad"
+ ],
+ [
+ "▁ha",
+ "d"
+ ],
+ [
+ "▁",
+ "had"
+ ],
+ [
+ "u",
+ "d"
+ ],
+ [
+ "as",
+ "e"
+ ],
+ [
+ "a",
+ "se"
+ ],
+ [
+ "or",
+ "d"
+ ],
+ [
+ "o",
+ "rd"
+ ],
+ [
+ ")",
+ ","
+ ],
+ [
+ "▁h",
+ "er"
+ ],
+ [
+ "▁he",
+ "r"
+ ],
+ [
+ "▁",
+ "her"
+ ],
+ [
+ "▁I",
+ "n"
+ ],
+ [
+ "▁",
+ "In"
+ ],
+ [
+ "ac",
+ "e"
+ ],
+ [
+ "a",
+ "ce"
+ ],
+ [
+ "▁b",
+ "ut"
+ ],
+ [
+ "▁bu",
+ "t"
+ ],
+ [
+ "▁",
+ "but"
+ ],
+ [
+ "at",
+ "a"
+ ],
+ [
+ "a",
+ "ta"
+ ],
+ [
+ ":",
+ ":"
+ ],
+ [
+ "**",
+ "**"
+ ],
+ [
+ "***",
+ "*"
+ ],
+ [
+ "*",
+ "***"
+ ],
+ [
+ "on",
+ "g"
+ ],
+ [
+ "o",
+ "ng"
+ ],
+ [
+ "▁",
+ "&"
+ ],
+ [
+ ".",
+ "."
+ ],
+ [
+ "it",
+ "e"
+ ],
+ [
+ "i",
+ "te"
+ ],
+ [
+ "yp",
+ "e"
+ ],
+ [
+ "y",
+ "pe"
+ ],
+ [
+ "ac",
+ "t"
+ ],
+ [
+ "a",
+ "ct"
+ ],
+ [
+ "od",
+ "e"
+ ],
+ [
+ "o",
+ "de"
+ ],
+ [
+ "▁y",
+ "our"
+ ],
+ [
+ "▁you",
+ "r"
+ ],
+ [
+ "▁yo",
+ "ur"
+ ],
+ [
+ "▁",
+ "your"
+ ],
+ [
+ "▁o",
+ "ut"
+ ],
+ [
+ "▁ou",
+ "t"
+ ],
+ [
+ "▁",
+ "out"
+ ],
+ [
+ "▁g",
+ "o"
+ ],
+ [
+ "▁",
+ "go"
+ ],
+ [
+ "li",
+ "c"
+ ],
+ [
+ "l",
+ "ic"
+ ],
+ [
+ "al",
+ "ly"
+ ],
+ [
+ "all",
+ "y"
+ ],
+ [
+ "▁s",
+ "o"
+ ],
+ [
+ "▁",
+ "so"
+ ],
+ [
+ "or",
+ "k"
+ ],
+ [
+ "a",
+ "u"
+ ],
+ [
+ "▁u",
+ "p"
+ ],
+ [
+ "▁",
+ "up"
+ ],
+ [
+ "▁",
+ "_"
+ ],
+ [
+ "l",
+ "l"
+ ],
+ [
+ "=",
+ "="
+ ],
+ [
+ "▁m",
+ "y"
+ ],
+ [
+ "▁",
+ "my"
+ ],
+ [
+ "p",
+ "p"
+ ],
+ [
+ "c",
+ "c"
+ ],
+ [
+ "▁/",
+ "/"
+ ],
+ [
+ "▁",
+ "//"
+ ],
+ [
+ "▁the",
+ "y"
+ ],
+ [
+ "▁th",
+ "ey"
+ ],
+ [
+ "▁",
+ "they"
+ ],
+ [
+ "g",
+ "h"
+ ],
+ [
+ "▁u",
+ "s"
+ ],
+ [
+ "▁",
+ "us"
+ ],
+ [
+ "i",
+ "b"
+ ],
+ [
+ "ion",
+ "s"
+ ],
+ [
+ "io",
+ "ns"
+ ],
+ [
+ "i",
+ "ons"
+ ],
+ [
+ "ac",
+ "h"
+ ],
+ [
+ "a",
+ "ch"
+ ],
+ [
+ "en",
+ "s"
+ ],
+ [
+ "e",
+ "ns"
+ ],
+ [
+ "▁a",
+ "r"
+ ],
+ [
+ "▁",
+ "ar"
+ ],
+ [
+ "o",
+ "b"
+ ],
+ [
+ "el",
+ "f"
+ ],
+ [
+ "oo",
+ "k"
+ ],
+ [
+ "o",
+ "ok"
+ ],
+ [
+ "at",
+ "ed"
+ ],
+ [
+ "ate",
+ "d"
+ ],
+ [
+ "a",
+ "ted"
+ ],
+ [
+ "an",
+ "g"
+ ],
+ [
+ "a",
+ "ng"
+ ],
+ [
+ "ig",
+ "n"
+ ],
+ [
+ "i",
+ "gn"
+ ],
+ [
+ "▁re",
+ "turn"
+ ],
+ [
+ "▁r",
+ "eturn"
+ ],
+ [
+ "▁ret",
+ "urn"
+ ],
+ [
+ "▁",
+ "return"
+ ],
+ [
+ "▁re",
+ "s"
+ ],
+ [
+ "▁r",
+ "es"
+ ],
+ [
+ "▁",
+ "res"
+ ],
+ [
+ "c",
+ "k"
+ ],
+ [
+ "ou",
+ "s"
+ ],
+ [
+ "o",
+ "us"
+ ],
+ [
+ "с",
+ "т"
+ ],
+ [
+ ")",
+ "."
+ ],
+ [
+ "▁",
+ "п"
+ ],
+ [
+ ".",
+ "\""
+ ],
+ [
+ "н",
+ "а"
+ ],
+ [
+ "▁",
+ "i"
+ ],
+ [
+ "ai",
+ "l"
+ ],
+ [
+ "a",
+ "il"
+ ],
+ [
+ "e",
+ "p"
+ ],
+ [
+ "▁a",
+ "d"
+ ],
+ [
+ "▁",
+ "ad"
+ ],
+ [
+ "an",
+ "ce"
+ ],
+ [
+ "anc",
+ "e"
+ ],
+ [
+ "(",
+ "\""
+ ],
+ [
+ "▁*",
+ "*"
+ ],
+ [
+ "▁",
+ "**"
+ ],
+ [
+ "th",
+ "er"
+ ],
+ [
+ "the",
+ "r"
+ ],
+ [
+ "t",
+ "her"
+ ],
+ [
+ "ak",
+ "e"
+ ],
+ [
+ "a",
+ "ke"
+ ],
+ [
+ "▁w",
+ "ill"
+ ],
+ [
+ "▁",
+ "will"
+ ],
+ [
+ "▁c",
+ "omp"
+ ],
+ [
+ "▁com",
+ "p"
+ ],
+ [
+ "▁co",
+ "mp"
+ ],
+ [
+ "▁",
+ "comp"
+ ],
+ [
+ "▁o",
+ "ne"
+ ],
+ [
+ "▁on",
+ "e"
+ ],
+ [
+ "▁",
+ "one"
+ ],
+ [
+ "▁g",
+ "et"
+ ],
+ [
+ "▁ge",
+ "t"
+ ],
+ [
+ "▁",
+ "get"
+ ],
+ [
+ "o",
+ "v"
+ ],
+ [
+ "▁",
+ "Y"
+ ],
+ [
+ "ar",
+ "y"
+ ],
+ [
+ "a",
+ "ry"
+ ],
+ [
+ "oc",
+ "k"
+ ],
+ [
+ "o",
+ "ck"
+ ],
+ [
+ "▁s",
+ "he"
+ ],
+ [
+ "▁sh",
+ "e"
+ ],
+ [
+ "▁",
+ "she"
+ ],
+ [
+ "ch",
+ "e"
+ ],
+ [
+ "c",
+ "he"
+ ],
+ [
+ "f",
+ "t"
+ ],
+ [
+ "▁n",
+ "ew"
+ ],
+ [
+ "▁ne",
+ "w"
+ ],
+ [
+ "▁",
+ "new"
+ ],
+ [
+ "▁d",
+ "es"
+ ],
+ [
+ "▁de",
+ "s"
+ ],
+ [
+ "▁",
+ "des"
+ ],
+ [
+ "▁l",
+ "i"
+ ],
+ [
+ "▁",
+ "li"
+ ],
+ [
+ "en",
+ "ce"
+ ],
+ [
+ "enc",
+ "e"
+ ],
+ [
+ "▁s",
+ "a"
+ ],
+ [
+ "▁",
+ "sa"
+ ],
+ [
+ "re",
+ "ss"
+ ],
+ [
+ "res",
+ "s"
+ ],
+ [
+ "r",
+ "ess"
+ ],
+ [
+ "▁e",
+ "l"
+ ],
+ [
+ "▁",
+ "el"
+ ],
+ [
+ "▁u",
+ "nd"
+ ],
+ [
+ "▁un",
+ "d"
+ ],
+ [
+ "▁",
+ "und"
+ ],
+ [
+ "e",
+ "g"
+ ],
+ [
+ "fe",
+ "r"
+ ],
+ [
+ "f",
+ "er"
+ ],
+ [
+ "r",
+ "y"
+ ],
+ [
+ "ea",
+ "r"
+ ],
+ [
+ "e",
+ "ar"
+ ],
+ [
+ "os",
+ "e"
+ ],
+ [
+ "o",
+ "se"
+ ],
+ [
+ "ve",
+ "ry"
+ ],
+ [
+ "ver",
+ "y"
+ ],
+ [
+ "v",
+ "ery"
+ ],
+ [
+ "'",
+ ","
+ ],
+ [
+ "▁",
+ "+"
+ ],
+ [
+ "▁",
+ "в"
+ ],
+ [
+ "▁H",
+ "e"
+ ],
+ [
+ "▁",
+ "He"
+ ],
+ [
+ "ub",
+ "lic"
+ ],
+ [
+ "ubl",
+ "ic"
+ ],
+ [
+ "u",
+ "blic"
+ ],
+ [
+ "▁the",
+ "ir"
+ ],
+ [
+ "iz",
+ "e"
+ ],
+ [
+ "i",
+ "ze"
+ ],
+ [
+ "▁w",
+ "ere"
+ ],
+ [
+ "▁we",
+ "re"
+ ],
+ [
+ "▁wer",
+ "e"
+ ],
+ [
+ "▁",
+ "were"
+ ],
+ [
+ "in",
+ "k"
+ ],
+ [
+ "ow",
+ "n"
+ ],
+ [
+ "o",
+ "wn"
+ ],
+ [
+ "I",
+ "n"
+ ],
+ [
+ "{",
+ "\\"
+ ],
+ [
+ "▁h",
+ "as"
+ ],
+ [
+ "▁ha",
+ "s"
+ ],
+ [
+ "▁",
+ "has"
+ ],
+ [
+ "▁p",
+ "er"
+ ],
+ [
+ "▁pe",
+ "r"
+ ],
+ [
+ "▁",
+ "per"
+ ],
+ [
+ "▁I",
+ "t"
+ ],
+ [
+ "▁",
+ "It"
+ ],
+ [
+ "▁S",
+ "t"
+ ],
+ [
+ "▁",
+ "St"
+ ],
+ [
+ "he",
+ "r"
+ ],
+ [
+ "h",
+ "er"
+ ],
+ [
+ "je",
+ "ct"
+ ],
+ [
+ "j",
+ "ect"
+ ],
+ [
+ "р",
+ "а"
+ ],
+ [
+ "il",
+ "d"
+ ],
+ [
+ "i",
+ "ld"
+ ],
+ [
+ "s",
+ "o"
+ ],
+ [
+ "▁s",
+ "p"
+ ],
+ [
+ "▁",
+ "sp"
+ ],
+ [
+ "н",
+ "и"
+ ],
+ [
+ "d",
+ "u"
+ ],
+ [
+ "ro",
+ "w"
+ ],
+ [
+ "r",
+ "ow"
+ ],
+ [
+ "al",
+ "ue"
+ ],
+ [
+ "alu",
+ "e"
+ ],
+ [
+ "se",
+ "t"
+ ],
+ [
+ "s",
+ "et"
+ ],
+ [
+ "fo",
+ "rm"
+ ],
+ [
+ "for",
+ "m"
+ ],
+ [
+ "f",
+ "orm"
+ ],
+ [
+ "co",
+ "m"
+ ],
+ [
+ "c",
+ "om"
+ ],
+ [
+ "▁m",
+ "an"
+ ],
+ [
+ "▁ma",
+ "n"
+ ],
+ [
+ "▁",
+ "man"
+ ],
+ [
+ "on",
+ "t"
+ ],
+ [
+ "o",
+ "nt"
+ ],
+ [
+ "ul",
+ "l"
+ ],
+ [
+ "u",
+ "ll"
+ ],
+ [
+ "▁c",
+ "ont"
+ ],
+ [
+ "▁con",
+ "t"
+ ],
+ [
+ "▁co",
+ "nt"
+ ],
+ [
+ "▁",
+ "cont"
+ ],
+ [
+ "▁m",
+ "ore"
+ ],
+ [
+ "▁mor",
+ "e"
+ ],
+ [
+ "▁mo",
+ "re"
+ ],
+ [
+ "▁",
+ "more"
+ ],
+ [
+ "ic",
+ "k"
+ ],
+ [
+ "i",
+ "ck"
+ ],
+ [
+ "▁w",
+ "ould"
+ ],
+ [
+ "▁wo",
+ "uld"
+ ],
+ [
+ "▁e",
+ "v"
+ ],
+ [
+ "▁",
+ "ev"
+ ],
+ [
+ "▁ab",
+ "out"
+ ],
+ [
+ "▁",
+ "about"
+ ],
+ [
+ "it",
+ "ion"
+ ],
+ [
+ "iti",
+ "on"
+ ],
+ [
+ "▁",
+ "z"
+ ],
+ [
+ "ou",
+ "nd"
+ ],
+ [
+ "oun",
+ "d"
+ ],
+ [
+ "o",
+ "und"
+ ],
+ [
+ "re",
+ "e"
+ ],
+ [
+ "r",
+ "ee"
+ ],
+ [
+ "▁C",
+ "h"
+ ],
+ [
+ "▁",
+ "Ch"
+ ],
+ [
+ "▁wh",
+ "ich"
+ ],
+ [
+ "▁",
+ "which"
+ ],
+ [
+ "i",
+ "o"
+ ],
+ [
+ "()",
+ ";"
+ ],
+ [
+ "(",
+ ");"
+ ],
+ [
+ "▁w",
+ "ho"
+ ],
+ [
+ "▁wh",
+ "o"
+ ],
+ [
+ "▁",
+ "who"
+ ],
+ [
+ "er",
+ "r"
+ ],
+ [
+ "e",
+ "rr"
+ ],
+ [
+ "or",
+ "y"
+ ],
+ [
+ "o",
+ "ry"
+ ],
+ [
+ "ou",
+ "nt"
+ ],
+ [
+ "oun",
+ "t"
+ ],
+ [
+ "o",
+ "unt"
+ ],
+ [
+ "at",
+ "ions"
+ ],
+ [
+ "ation",
+ "s"
+ ],
+ [
+ "ati",
+ "ons"
+ ],
+ [
+ "atio",
+ "ns"
+ ],
+ [
+ "▁",
+ "с"
+ ],
+ [
+ "ri",
+ "ng"
+ ],
+ [
+ "rin",
+ "g"
+ ],
+ [
+ "r",
+ "ing"
+ ],
+ [
+ "<",
+ "/"
+ ],
+ [
+ "▁f",
+ "e"
+ ],
+ [
+ "▁",
+ "fe"
+ ],
+ [
+ "к",
+ "о"
+ ],
+ [
+ "н",
+ "о"
+ ],
+ [
+ "▁d",
+ "is"
+ ],
+ [
+ "▁di",
+ "s"
+ ],
+ [
+ "▁",
+ "dis"
+ ],
+ [
+ "m",
+ "a"
+ ],
+ [
+ "▁t",
+ "hem"
+ ],
+ [
+ "▁the",
+ "m"
+ ],
+ [
+ "▁th",
+ "em"
+ ],
+ [
+ "▁a",
+ "ny"
+ ],
+ [
+ "▁an",
+ "y"
+ ],
+ [
+ "▁",
+ "any"
+ ],
+ [
+ "▁n",
+ "o"
+ ],
+ [
+ "▁",
+ "no"
+ ],
+ [
+ "--",
+ "------"
+ ],
+ [
+ "----",
+ "----"
+ ],
+ [
+ "---",
+ "-----"
+ ],
+ [
+ "-----",
+ "---"
+ ],
+ [
+ "------",
+ "--"
+ ],
+ [
+ "-------",
+ "-"
+ ],
+ [
+ "-",
+ "-------"
+ ],
+ [
+ "▁p",
+ "re"
+ ],
+ [
+ "▁pr",
+ "e"
+ ],
+ [
+ "▁",
+ "pre"
+ ],
+ [
+ "▁t",
+ "e"
+ ],
+ [
+ "▁",
+ "te"
+ ],
+ [
+ "▁r",
+ "o"
+ ],
+ [
+ "▁",
+ "ro"
+ ],
+ [
+ "▁h",
+ "im"
+ ],
+ [
+ "▁hi",
+ "m"
+ ],
+ [
+ "▁",
+ "him"
+ ],
+ [
+ "▁",
+ ":"
+ ],
+ [
+ "u",
+ "p"
+ ],
+ [
+ "▁in",
+ "t"
+ ],
+ [
+ "▁i",
+ "nt"
+ ],
+ [
+ "▁",
+ "int"
+ ],
+ [
+ "▁a",
+ "g"
+ ],
+ [
+ "▁",
+ "ag"
+ ],
+ [
+ "S",
+ "t"
+ ],
+ [
+ "ar",
+ "k"
+ ],
+ [
+ "e",
+ "x"
+ ],
+ [
+ "p",
+ "h"
+ ],
+ [
+ "ie",
+ "nt"
+ ],
+ [
+ "ien",
+ "t"
+ ],
+ [
+ "i",
+ "ent"
+ ],
+ [
+ "el",
+ "y"
+ ],
+ [
+ "e",
+ "ly"
+ ],
+ [
+ "▁p",
+ "r"
+ ],
+ [
+ "▁",
+ "pr"
+ ],
+ [
+ "E",
+ "R"
+ ],
+ [
+ "▁im",
+ "port"
+ ],
+ [
+ "▁imp",
+ "ort"
+ ],
+ [
+ "▁",
+ "import"
+ ],
+ [
+ "▁t",
+ "ime"
+ ],
+ [
+ "▁tim",
+ "e"
+ ],
+ [
+ "▁ti",
+ "me"
+ ],
+ [
+ "▁",
+ "time"
+ ],
+ [
+ "р",
+ "о"
+ ],
+ [
+ "pr",
+ "o"
+ ],
+ [
+ "p",
+ "ro"
+ ],
+ [
+ "Us",
+ "er"
+ ],
+ [
+ "Use",
+ "r"
+ ],
+ [
+ "U",
+ "ser"
+ ],
+ [
+ "l",
+ "o"
+ ],
+ [
+ "▁",
+ "/"
+ ],
+ [
+ "▁",
+ "["
+ ],
+ [
+ "or",
+ "s"
+ ],
+ [
+ "o",
+ "rs"
+ ],
+ [
+ "=",
+ "\""
+ ],
+ [
+ "▁t",
+ "here"
+ ],
+ [
+ "▁the",
+ "re"
+ ],
+ [
+ "▁th",
+ "ere"
+ ],
+ [
+ "▁ther",
+ "e"
+ ],
+ [
+ "▁",
+ "there"
+ ],
+ [
+ "▁l",
+ "ike"
+ ],
+ [
+ "▁li",
+ "ke"
+ ],
+ [
+ "▁lik",
+ "e"
+ ],
+ [
+ "▁",
+ "like"
+ ],
+ [
+ "ol",
+ "d"
+ ],
+ [
+ "o",
+ "ld"
+ ],
+ [
+ "▁w",
+ "hen"
+ ],
+ [
+ "▁wh",
+ "en"
+ ],
+ [
+ "▁whe",
+ "n"
+ ],
+ [
+ "▁",
+ "when"
+ ],
+ [
+ "ve",
+ "rs"
+ ],
+ [
+ "ver",
+ "s"
+ ],
+ [
+ "v",
+ "ers"
+ ],
+ [
+ "▁s",
+ "ome"
+ ],
+ [
+ "▁so",
+ "me"
+ ],
+ [
+ "▁som",
+ "e"
+ ],
+ [
+ "▁",
+ "some"
+ ],
+ [
+ "in",
+ "gs"
+ ],
+ [
+ "ing",
+ "s"
+ ],
+ [
+ ")",
+ ")"
+ ],
+ [
+ "▁p",
+ "art"
+ ],
+ [
+ "▁par",
+ "t"
+ ],
+ [
+ "▁pa",
+ "rt"
+ ],
+ [
+ "▁",
+ "part"
+ ],
+ [
+ "ic",
+ "al"
+ ],
+ [
+ "ica",
+ "l"
+ ],
+ [
+ "i",
+ "cal"
+ ],
+ [
+ "▁f",
+ "un"
+ ],
+ [
+ "▁fu",
+ "n"
+ ],
+ [
+ "▁",
+ "fun"
+ ],
+ [
+ "▁k",
+ "n"
+ ],
+ [
+ "▁",
+ "kn"
+ ],
+ [
+ "ay",
+ "s"
+ ],
+ [
+ "a",
+ "ys"
+ ],
+ [
+ "ie",
+ "r"
+ ],
+ [
+ "i",
+ "er"
+ ],
+ [
+ "▁b",
+ "een"
+ ],
+ [
+ "▁be",
+ "en"
+ ],
+ [
+ "ov",
+ "e"
+ ],
+ [
+ "o",
+ "ve"
+ ],
+ [
+ "▁s",
+ "c"
+ ],
+ [
+ "▁",
+ "sc"
+ ],
+ [
+ "ia",
+ "n"
+ ],
+ [
+ "i",
+ "an"
+ ],
+ [
+ "▁o",
+ "ver"
+ ],
+ [
+ "▁ov",
+ "er"
+ ],
+ [
+ "▁",
+ "over"
+ ],
+ [
+ "ie",
+ "l"
+ ],
+ [
+ "i",
+ "el"
+ ],
+ [
+ "▁p",
+ "e"
+ ],
+ [
+ "▁",
+ "pe"
+ ],
+ [
+ "ri",
+ "b"
+ ],
+ [
+ "r",
+ "ib"
+ ],
+ [
+ "pu",
+ "t"
+ ],
+ [
+ "p",
+ "ut"
+ ],
+ [
+ "e",
+ "c"
+ ],
+ [
+ "et",
+ "h"
+ ],
+ [
+ "e",
+ "th"
+ ],
+ [
+ "ar",
+ "am"
+ ],
+ [
+ "ara",
+ "m"
+ ],
+ [
+ "a",
+ "ram"
+ ],
+ [
+ "ap",
+ "p"
+ ],
+ [
+ "a",
+ "pp"
+ ],
+ [
+ "▁",
+ "–"
+ ],
+ [
+ "▁s",
+ "tat"
+ ],
+ [
+ "▁st",
+ "at"
+ ],
+ [
+ "▁sta",
+ "t"
+ ],
+ [
+ "▁",
+ "stat"
+ ],
+ [
+ "po",
+ "n"
+ ],
+ [
+ "p",
+ "on"
+ ],
+ [
+ "▁w",
+ "hat"
+ ],
+ [
+ "▁wh",
+ "at"
+ ],
+ [
+ "▁",
+ "what"
+ ],
+ [
+ "pt",
+ "ion"
+ ],
+ [
+ "w",
+ "e"
+ ],
+ [
+ "ad",
+ "e"
+ ],
+ [
+ "a",
+ "de"
+ ],
+ [
+ "▁w",
+ "ork"
+ ],
+ [
+ "▁wor",
+ "k"
+ ],
+ [
+ "▁",
+ "work"
+ ],
+ [
+ "te",
+ "xt"
+ ],
+ [
+ "tex",
+ "t"
+ ],
+ [
+ "t",
+ "ext"
+ ],
+ [
+ "▁s",
+ "aid"
+ ],
+ [
+ "▁sa",
+ "id"
+ ],
+ [
+ "▁#",
+ "##"
+ ],
+ [
+ "▁##",
+ "#"
+ ],
+ [
+ "▁",
+ "###"
+ ],
+ [
+ "I",
+ "N"
+ ],
+ [
+ "▁j",
+ "ust"
+ ],
+ [
+ "▁ju",
+ "st"
+ ],
+ [
+ "▁",
+ "just"
+ ],
+ [
+ "ir",
+ "st"
+ ],
+ [
+ "irs",
+ "t"
+ ],
+ [
+ "▁in",
+ "to"
+ ],
+ [
+ "▁int",
+ "o"
+ ],
+ [
+ "▁",
+ "into"
+ ],
+ [
+ "▁con",
+ "st"
+ ],
+ [
+ "▁cons",
+ "t"
+ ],
+ [
+ "▁",
+ "const"
+ ],
+ [
+ "our",
+ "ce"
+ ],
+ [
+ "t",
+ "t"
+ ],
+ [
+ "p",
+ "s"
+ ],
+ [
+ "p",
+ "r"
+ ],
+ [
+ "er",
+ "v"
+ ],
+ [
+ "e",
+ "rv"
+ ],
+ [
+ "it",
+ "t"
+ ],
+ [
+ "i",
+ "tt"
+ ],
+ [
+ "u",
+ "g"
+ ],
+ [
+ "_",
+ "{"
+ ],
+ [
+ "en",
+ "ts"
+ ],
+ [
+ "ent",
+ "s"
+ ],
+ [
+ "is",
+ "h"
+ ],
+ [
+ "i",
+ "sh"
+ ],
+ [
+ "en",
+ "er"
+ ],
+ [
+ "ene",
+ "r"
+ ],
+ [
+ "e",
+ "ner"
+ ],
+ [
+ "▁in",
+ "ter"
+ ],
+ [
+ "▁int",
+ "er"
+ ],
+ [
+ "▁inte",
+ "r"
+ ],
+ [
+ "▁",
+ "inter"
+ ],
+ [
+ "pl",
+ "e"
+ ],
+ [
+ "p",
+ "le"
+ ],
+ [
+ "ol",
+ "l"
+ ],
+ [
+ "o",
+ "ll"
+ ],
+ [
+ "me",
+ "r"
+ ],
+ [
+ "m",
+ "er"
+ ],
+ [
+ "at",
+ "er"
+ ],
+ [
+ "ate",
+ "r"
+ ],
+ [
+ "a",
+ "ter"
+ ],
+ [
+ "oo",
+ "l"
+ ],
+ [
+ "o",
+ "ol"
+ ],
+ [
+ "e",
+ "f"
+ ],
+ [
+ "▁p",
+ "ublic"
+ ],
+ [
+ "▁pub",
+ "lic"
+ ],
+ [
+ "▁pu",
+ "blic"
+ ],
+ [
+ "▁publi",
+ "c"
+ ],
+ [
+ "▁",
+ "public"
+ ],
+ [
+ "▁o",
+ "ther"
+ ],
+ [
+ "▁ot",
+ "her"
+ ],
+ [
+ "▁",
+ "other"
+ ],
+ [
+ "р",
+ "е"
+ ],
+ [
+ "▁d",
+ "ef"
+ ],
+ [
+ "▁de",
+ "f"
+ ],
+ [
+ "▁",
+ "def"
+ ],
+ [
+ "▁",
+ "@"
+ ],
+ [
+ "г",
+ "о"
+ ],
+ [
+ "oin",
+ "t"
+ ],
+ [
+ "oi",
+ "nt"
+ ],
+ [
+ "o",
+ "int"
+ ],
+ [
+ "▁o",
+ "ff"
+ ],
+ [
+ "▁of",
+ "f"
+ ],
+ [
+ "▁",
+ "off"
+ ],
+ [
+ "oi",
+ "d"
+ ],
+ [
+ "o",
+ "id"
+ ],
+ [
+ "re",
+ "turn"
+ ],
+ [
+ "ret",
+ "urn"
+ ],
+ [
+ "r",
+ "eturn"
+ ],
+ [
+ "▁s",
+ "et"
+ ],
+ [
+ "▁se",
+ "t"
+ ],
+ [
+ "▁",
+ "set"
+ ],
+ [
+ "w",
+ "o"
+ ],
+ [
+ "ft",
+ "er"
+ ],
+ [
+ "fte",
+ "r"
+ ],
+ [
+ "f",
+ "ter"
+ ],
+ [
+ "s",
+ "h"
+ ],
+ [
+ "**",
+ "******"
+ ],
+ [
+ "****",
+ "****"
+ ],
+ [
+ "******",
+ "**"
+ ],
+ [
+ "▁o",
+ "ur"
+ ],
+ [
+ "▁ou",
+ "r"
+ ],
+ [
+ "▁",
+ "our"
+ ],
+ [
+ "ri",
+ "v"
+ ],
+ [
+ "r",
+ "iv"
+ ],
+ [
+ "is",
+ "s"
+ ],
+ [
+ "i",
+ "ss"
+ ],
+ [
+ "▁W",
+ "e"
+ ],
+ [
+ "▁",
+ "We"
+ ],
+ [
+ "n",
+ "g"
+ ],
+ [
+ "▁o",
+ "b"
+ ],
+ [
+ "▁",
+ "ob"
+ ],
+ [
+ "s",
+ "s"
+ ],
+ [
+ "g",
+ "r"
+ ],
+ [
+ "▁t",
+ "han"
+ ],
+ [
+ "▁th",
+ "an"
+ ],
+ [
+ "▁",
+ "than"
+ ],
+ [
+ "pe",
+ "ct"
+ ],
+ [
+ "pec",
+ "t"
+ ],
+ [
+ "p",
+ "ect"
+ ],
+ [
+ "ie",
+ "d"
+ ],
+ [
+ "i",
+ "ed"
+ ],
+ [
+ "s",
+ "c"
+ ],
+ [
+ "ie",
+ "w"
+ ],
+ [
+ "i",
+ "ew"
+ ],
+ [
+ "de",
+ "r"
+ ],
+ [
+ "d",
+ "er"
+ ],
+ [
+ "ys",
+ "t"
+ ],
+ [
+ "y",
+ "st"
+ ],
+ [
+ "e",
+ "v"
+ ],
+ [
+ "▁c",
+ "ould"
+ ],
+ [
+ "▁co",
+ "uld"
+ ],
+ [
+ "▁cou",
+ "ld"
+ ],
+ [
+ "▁",
+ "could"
+ ],
+ [
+ "an",
+ "n"
+ ],
+ [
+ "a",
+ "nn"
+ ],
+ [
+ "en",
+ "c"
+ ],
+ [
+ "e",
+ "nc"
+ ],
+ [
+ "O",
+ "N"
+ ],
+ [
+ "i",
+ "x"
+ ],
+ [
+ "an",
+ "c"
+ ],
+ [
+ "a",
+ "nc"
+ ],
+ [
+ "▁al",
+ "so"
+ ],
+ [
+ "▁als",
+ "o"
+ ],
+ [
+ "▁",
+ "also"
+ ],
+ [
+ "re",
+ "at"
+ ],
+ [
+ "rea",
+ "t"
+ ],
+ [
+ "▁a",
+ "m"
+ ],
+ [
+ "▁",
+ "am"
+ ],
+ [
+ "▁b",
+ "ec"
+ ],
+ [
+ "▁be",
+ "c"
+ ],
+ [
+ "▁",
+ "bec"
+ ],
+ [
+ "▁",
+ "и"
+ ],
+ [
+ "ua",
+ "l"
+ ],
+ [
+ "u",
+ "al"
+ ],
+ [
+ "pe",
+ "c"
+ ],
+ [
+ "p",
+ "ec"
+ ],
+ [
+ "▁",
+ "."
+ ],
+ [
+ "▁b",
+ "l"
+ ],
+ [
+ "▁",
+ "bl"
+ ],
+ [
+ "le",
+ "ct"
+ ],
+ [
+ "l",
+ "ect"
+ ],
+ [
+ "op",
+ "le"
+ ],
+ [
+ "opl",
+ "e"
+ ],
+ [
+ "o",
+ "ple"
+ ],
+ [
+ "y",
+ "s"
+ ],
+ [
+ "▁g",
+ "r"
+ ],
+ [
+ "▁",
+ "gr"
+ ],
+ [
+ "ic",
+ "t"
+ ],
+ [
+ "i",
+ "ct"
+ ],
+ [
+ "i",
+ "k"
+ ],
+ [
+ "tr",
+ "ing"
+ ],
+ [
+ "tri",
+ "ng"
+ ],
+ [
+ "t",
+ "ring"
+ ],
+ [
+ "▁T",
+ "his"
+ ],
+ [
+ "▁Th",
+ "is"
+ ],
+ [
+ "▁",
+ "This"
+ ],
+ [
+ "▁b",
+ "ack"
+ ],
+ [
+ "▁ba",
+ "ck"
+ ],
+ [
+ "▁",
+ "back"
+ ],
+ [
+ "▁",
+ "о"
+ ],
+ [
+ "▁f",
+ "in"
+ ],
+ [
+ "▁fi",
+ "n"
+ ],
+ [
+ "▁",
+ "fin"
+ ],
+ [
+ "at",
+ "ch"
+ ],
+ [
+ "Co",
+ "n"
+ ],
+ [
+ "C",
+ "on"
+ ],
+ [
+ "(",
+ "'"
+ ],
+ [
+ "er",
+ "m"
+ ],
+ [
+ "e",
+ "rm"
+ ],
+ [
+ "▁=",
+ "="
+ ],
+ [
+ "▁",
+ "=="
+ ],
+ [
+ "_",
+ "_"
+ ],
+ [
+ "na",
+ "me"
+ ],
+ [
+ "nam",
+ "e"
+ ],
+ [
+ "n",
+ "ame"
+ ],
+ [
+ ",",
+ "\""
+ ],
+ [
+ "▁d",
+ "id"
+ ],
+ [
+ "▁di",
+ "d"
+ ],
+ [
+ "▁",
+ "did"
+ ],
+ [
+ "is",
+ "e"
+ ],
+ [
+ "i",
+ "se"
+ ],
+ [
+ "▁on",
+ "ly"
+ ],
+ [
+ "▁",
+ "only"
+ ],
+ [
+ "ru",
+ "ct"
+ ],
+ [
+ "r",
+ "uct"
+ ],
+ [
+ "le",
+ "s"
+ ],
+ [
+ "l",
+ "es"
+ ],
+ [
+ "▁t",
+ "hen"
+ ],
+ [
+ "▁the",
+ "n"
+ ],
+ [
+ "▁th",
+ "en"
+ ],
+ [
+ "▁",
+ "then"
+ ],
+ [
+ "au",
+ "se"
+ ],
+ [
+ "aus",
+ "e"
+ ],
+ [
+ "a",
+ "use"
+ ],
+ [
+ "в",
+ "а"
+ ],
+ [
+ "▁it",
+ "s"
+ ],
+ [
+ "▁i",
+ "ts"
+ ],
+ [
+ "▁",
+ "its"
+ ],
+ [
+ "ri",
+ "t"
+ ],
+ [
+ "r",
+ "it"
+ ],
+ [
+ "▁k",
+ "now"
+ ],
+ [
+ "▁kn",
+ "ow"
+ ],
+ [
+ "▁",
+ "know"
+ ],
+ [
+ "ie",
+ "ld"
+ ],
+ [
+ "iel",
+ "d"
+ ],
+ [
+ "i",
+ "eld"
+ ],
+ [
+ "▁c",
+ "lass"
+ ],
+ [
+ "▁cl",
+ "ass"
+ ],
+ [
+ "▁clas",
+ "s"
+ ],
+ [
+ "▁",
+ "class"
+ ],
+ [
+ "▁",
+ ">"
+ ],
+ [
+ "▁e",
+ "m"
+ ],
+ [
+ "▁",
+ "em"
+ ],
+ [
+ "▁$",
+ "\\"
+ ],
+ [
+ "▁",
+ "$\\"
+ ],
+ [
+ "▁y",
+ "ear"
+ ],
+ [
+ "▁ye",
+ "ar"
+ ],
+ [
+ "▁",
+ "year"
+ ],
+ [
+ "w",
+ "n"
+ ],
+ [
+ "}",
+ ","
+ ],
+ [
+ "▁d",
+ "el"
+ ],
+ [
+ "▁de",
+ "l"
+ ],
+ [
+ "▁",
+ "del"
+ ],
+ [
+ "al",
+ "e"
+ ],
+ [
+ "a",
+ "le"
+ ],
+ [
+ "t",
+ "y"
+ ],
+ [
+ "fi",
+ "g"
+ ],
+ [
+ "f",
+ "ig"
+ ],
+ [
+ "s",
+ "p"
+ ],
+ [
+ "he",
+ "d"
+ ],
+ [
+ "h",
+ "ed"
+ ],
+ [
+ "ro",
+ "und"
+ ],
+ [
+ "rou",
+ "nd"
+ ],
+ [
+ "r",
+ "ound"
+ ],
+ [
+ "e",
+ "w"
+ ],
+ [
+ "▁d",
+ "i"
+ ],
+ [
+ "▁",
+ "di"
+ ],
+ [
+ "▁d",
+ "er"
+ ],
+ [
+ "▁de",
+ "r"
+ ],
+ [
+ "▁",
+ "der"
+ ],
+ [
+ "р",
+ "и"
+ ],
+ [
+ "re",
+ "d"
+ ],
+ [
+ "r",
+ "ed"
+ ],
+ [
+ "th",
+ "is"
+ ],
+ [
+ "t",
+ "his"
+ ],
+ [
+ "le",
+ "t"
+ ],
+ [
+ "l",
+ "et"
+ ],
+ [
+ "R",
+ "E"
+ ],
+ [
+ "a",
+ "x"
+ ],
+ [
+ "f",
+ "r"
+ ],
+ [
+ "ess",
+ "age"
+ ],
+ [
+ "essa",
+ "ge"
+ ],
+ [
+ "ou",
+ "gh"
+ ],
+ [
+ "o",
+ "ugh"
+ ],
+ [
+ "▁c",
+ "omm"
+ ],
+ [
+ "▁com",
+ "m"
+ ],
+ [
+ "▁co",
+ "mm"
+ ],
+ [
+ "▁",
+ "comm"
+ ],
+ [
+ "f",
+ "o"
+ ],
+ [
+ "uc",
+ "h"
+ ],
+ [
+ "u",
+ "ch"
+ ],
+ [
+ "o",
+ "y"
+ ],
+ [
+ "▁pe",
+ "ople"
+ ],
+ [
+ "▁",
+ "people"
+ ],
+ [
+ "yst",
+ "em"
+ ],
+ [
+ "ys",
+ "tem"
+ ],
+ [
+ "▁f",
+ "irst"
+ ],
+ [
+ "▁fir",
+ "st"
+ ],
+ [
+ "▁",
+ "first"
+ ],
+ [
+ "▁f",
+ "unction"
+ ],
+ [
+ "▁fun",
+ "ction"
+ ],
+ [
+ "▁",
+ "function"
+ ],
+ [
+ "an",
+ "ge"
+ ],
+ [
+ "ang",
+ "e"
+ ],
+ [
+ "▁h",
+ "ow"
+ ],
+ [
+ "▁ho",
+ "w"
+ ],
+ [
+ "▁",
+ "how"
+ ],
+ [
+ "▁e",
+ "t"
+ ],
+ [
+ "▁",
+ "et"
+ ],
+ [
+ "a",
+ "h"
+ ],
+ [
+ "▁l",
+ "ook"
+ ],
+ [
+ "▁lo",
+ "ok"
+ ],
+ [
+ "▁",
+ "look"
+ ],
+ [
+ "т",
+ "о"
+ ],
+ [
+ "un",
+ "d"
+ ],
+ [
+ "u",
+ "nd"
+ ],
+ [
+ "▁u",
+ "nder"
+ ],
+ [
+ "▁un",
+ "der"
+ ],
+ [
+ "▁und",
+ "er"
+ ],
+ [
+ "▁",
+ "under"
+ ],
+ [
+ "к",
+ "а"
+ ],
+ [
+ "▁",
+ "!"
+ ],
+ [
+ "ra",
+ "y"
+ ],
+ [
+ "r",
+ "ay"
+ ],
+ [
+ "S",
+ "T"
+ ],
+ [
+ "if",
+ "ic"
+ ],
+ [
+ "ifi",
+ "c"
+ ],
+ [
+ "i",
+ "fic"
+ ],
+ [
+ "л",
+ "и"
+ ],
+ [
+ "re",
+ "ad"
+ ],
+ [
+ "rea",
+ "d"
+ ],
+ [
+ "r",
+ "ead"
+ ],
+ [
+ "▁b",
+ "et"
+ ],
+ [
+ "▁be",
+ "t"
+ ],
+ [
+ "▁",
+ "bet"
+ ],
+ [
+ "io",
+ "us"
+ ],
+ [
+ "i",
+ "ous"
+ ],
+ [
+ "ar",
+ "g"
+ ],
+ [
+ "a",
+ "rg"
+ ],
+ [
+ "▁n",
+ "eed"
+ ],
+ [
+ "▁ne",
+ "ed"
+ ],
+ [
+ "▁",
+ "need"
+ ],
+ [
+ "ma",
+ "th"
+ ],
+ [
+ "mat",
+ "h"
+ ],
+ [
+ "m",
+ "ath"
+ ],
+ [
+ "▁н",
+ "а"
+ ],
+ [
+ "▁",
+ "на"
+ ],
+ [
+ "er",
+ "t"
+ ],
+ [
+ "e",
+ "rt"
+ ],
+ [
+ "▁o",
+ "p"
+ ],
+ [
+ "▁",
+ "op"
+ ],
+ [
+ "▁a",
+ "cc"
+ ],
+ [
+ "▁ac",
+ "c"
+ ],
+ [
+ "▁",
+ "acc"
+ ],
+ [
+ "Pr",
+ "o"
+ ],
+ [
+ "P",
+ "ro"
+ ],
+ [
+ "▁e",
+ "st"
+ ],
+ [
+ "▁es",
+ "t"
+ ],
+ [
+ "▁",
+ "est"
+ ],
+ [
+ "▁U",
+ "n"
+ ],
+ [
+ "▁",
+ "Un"
+ ],
+ [
+ "▁e",
+ "nt"
+ ],
+ [
+ "▁en",
+ "t"
+ ],
+ [
+ "▁",
+ "ent"
+ ],
+ [
+ "▁re",
+ "c"
+ ],
+ [
+ "▁r",
+ "ec"
+ ],
+ [
+ "▁",
+ "rec"
+ ],
+ [
+ "▁u",
+ "se"
+ ],
+ [
+ "▁us",
+ "e"
+ ],
+ [
+ "▁",
+ "use"
+ ],
+ [
+ "е",
+ "н"
+ ],
+ [
+ "▁p",
+ "ar"
+ ],
+ [
+ "▁pa",
+ "r"
+ ],
+ [
+ "▁",
+ "par"
+ ],
+ [
+ "a",
+ "z"
+ ],
+ [
+ "▁",
+ "д"
+ ],
+ [
+ "▁W",
+ "h"
+ ],
+ [
+ "▁",
+ "Wh"
+ ],
+ [
+ "sel",
+ "f"
+ ],
+ [
+ "s",
+ "elf"
+ ],
+ [
+ "▁k",
+ "e"
+ ],
+ [
+ "▁",
+ "ke"
+ ],
+ [
+ "т",
+ "а"
+ ],
+ [
+ "▁w",
+ "ant"
+ ],
+ [
+ "▁wa",
+ "nt"
+ ],
+ [
+ "▁",
+ "want"
+ ],
+ [
+ "▁e",
+ "nd"
+ ],
+ [
+ "▁en",
+ "d"
+ ],
+ [
+ "▁",
+ "end"
+ ],
+ [
+ "▁d",
+ "on"
+ ],
+ [
+ "▁do",
+ "n"
+ ],
+ [
+ "▁",
+ "don"
+ ],
+ [
+ "e",
+ "k"
+ ],
+ [
+ "re",
+ "n"
+ ],
+ [
+ "r",
+ "en"
+ ],
+ [
+ "Na",
+ "me"
+ ],
+ [
+ "N",
+ "ame"
+ ],
+ [
+ "▁=",
+ ">"
+ ],
+ [
+ "▁",
+ "=>"
+ ],
+ [
+ "▁a",
+ "pp"
+ ],
+ [
+ "▁ap",
+ "p"
+ ],
+ [
+ "▁",
+ "app"
+ ],
+ [
+ "▁qu",
+ "e"
+ ],
+ [
+ "▁q",
+ "ue"
+ ],
+ [
+ "▁",
+ "que"
+ ],
+ [
+ "ig",
+ "h"
+ ],
+ [
+ "i",
+ "gh"
+ ],
+ [
+ "▁b",
+ "u"
+ ],
+ [
+ "▁",
+ "bu"
+ ],
+ [
+ "eq",
+ "u"
+ ],
+ [
+ "e",
+ "qu"
+ ],
+ [
+ "ve",
+ "l"
+ ],
+ [
+ "v",
+ "el"
+ ],
+ [
+ "▁a",
+ "ct"
+ ],
+ [
+ "▁ac",
+ "t"
+ ],
+ [
+ "▁",
+ "act"
+ ],
+ [
+ "cr",
+ "e"
+ ],
+ [
+ "c",
+ "re"
+ ],
+ [
+ "A",
+ "T"
+ ],
+ [
+ "▁v",
+ "ar"
+ ],
+ [
+ "▁va",
+ "r"
+ ],
+ [
+ "▁",
+ "var"
+ ],
+ [
+ "ce",
+ "ss"
+ ],
+ [
+ "ces",
+ "s"
+ ],
+ [
+ "c",
+ "ess"
+ ],
+ [
+ "==",
+ "=="
+ ],
+ [
+ "===",
+ "="
+ ],
+ [
+ "=",
+ "==="
+ ],
+ [
+ "E",
+ "x"
+ ],
+ [
+ "▁a",
+ "dd"
+ ],
+ [
+ "▁ad",
+ "d"
+ ],
+ [
+ "▁",
+ "add"
+ ],
+ [
+ "▁m",
+ "od"
+ ],
+ [
+ "▁mo",
+ "d"
+ ],
+ [
+ "▁",
+ "mod"
+ ],
+ [
+ "un",
+ "g"
+ ],
+ [
+ "u",
+ "ng"
+ ],
+ [
+ "▁w",
+ "here"
+ ],
+ [
+ "▁wh",
+ "ere"
+ ],
+ [
+ "▁whe",
+ "re"
+ ],
+ [
+ "▁",
+ "where"
+ ],
+ [
+ "ni",
+ "ng"
+ ],
+ [
+ "n",
+ "ing"
+ ],
+ [
+ "▁f",
+ "l"
+ ],
+ [
+ "▁",
+ "fl"
+ ],
+ [
+ "al",
+ "s"
+ ],
+ [
+ "a",
+ "ls"
+ ],
+ [
+ "ter",
+ "n"
+ ],
+ [
+ "te",
+ "rn"
+ ],
+ [
+ "t",
+ "ern"
+ ],
+ [
+ "}",
+ "}"
+ ],
+ [
+ "▁A",
+ "l"
+ ],
+ [
+ "▁",
+ "Al"
+ ],
+ [
+ "▁p",
+ "os"
+ ],
+ [
+ "▁po",
+ "s"
+ ],
+ [
+ "▁",
+ "pos"
+ ],
+ [
+ "an",
+ "k"
+ ],
+ [
+ "▁a",
+ "p"
+ ],
+ [
+ "▁",
+ "ap"
+ ],
+ [
+ "en",
+ "g"
+ ],
+ [
+ "e",
+ "ng"
+ ],
+ [
+ "▁",
+ "“"
+ ],
+ [
+ "bl",
+ "e"
+ ],
+ [
+ "b",
+ "le"
+ ],
+ [
+ "▁re",
+ "g"
+ ],
+ [
+ "▁r",
+ "eg"
+ ],
+ [
+ "▁",
+ "reg"
+ ],
+ [
+ "^",
+ "{"
+ ],
+ [
+ "▁S",
+ "he"
+ ],
+ [
+ "▁Sh",
+ "e"
+ ],
+ [
+ "▁",
+ "She"
+ ],
+ [
+ "▁*",
+ "/"
+ ],
+ [
+ "▁",
+ "*/"
+ ],
+ [
+ "ud",
+ "e"
+ ],
+ [
+ "u",
+ "de"
+ ],
+ [
+ "ad",
+ "d"
+ ],
+ [
+ "a",
+ "dd"
+ ],
+ [
+ "▁t",
+ "wo"
+ ],
+ [
+ "▁tw",
+ "o"
+ ],
+ [
+ "▁",
+ "two"
+ ],
+ [
+ "▁c",
+ "ol"
+ ],
+ [
+ "▁co",
+ "l"
+ ],
+ [
+ "▁",
+ "col"
+ ],
+ [
+ "▁s",
+ "m"
+ ],
+ [
+ "▁",
+ "sm"
+ ],
+ [
+ "ai",
+ "r"
+ ],
+ [
+ "a",
+ "ir"
+ ],
+ [
+ "▁m",
+ "ay"
+ ],
+ [
+ "▁ma",
+ "y"
+ ],
+ [
+ "▁",
+ "may"
+ ],
+ [
+ "fo",
+ "re"
+ ],
+ [
+ "for",
+ "e"
+ ],
+ [
+ "f",
+ "ore"
+ ],
+ [
+ "▁Y",
+ "ou"
+ ],
+ [
+ "▁",
+ "You"
+ ],
+ [
+ "ro",
+ "ugh"
+ ],
+ [
+ "rou",
+ "gh"
+ ],
+ [
+ "r",
+ "ough"
+ ],
+ [
+ "▁c",
+ "he"
+ ],
+ [
+ "▁ch",
+ "e"
+ ],
+ [
+ "▁",
+ "che"
+ ],
+ [
+ "▁a",
+ "tt"
+ ],
+ [
+ "▁at",
+ "t"
+ ],
+ [
+ "▁",
+ "att"
+ ],
+ [
+ "ot",
+ "h"
+ ],
+ [
+ "o",
+ "th"
+ ],
+ [
+ "л",
+ "а"
+ ],
+ [
+ "▁c",
+ "o"
+ ],
+ [
+ "▁",
+ "co"
+ ],
+ [
+ "at",
+ "es"
+ ],
+ [
+ "ate",
+ "s"
+ ],
+ [
+ "a",
+ "tes"
+ ],
+ [
+ "▁re",
+ "m"
+ ],
+ [
+ "▁r",
+ "em"
+ ],
+ [
+ "▁",
+ "rem"
+ ],
+ [
+ "oo",
+ "d"
+ ],
+ [
+ "o",
+ "od"
+ ],
+ [
+ "Ty",
+ "pe"
+ ],
+ [
+ "Typ",
+ "e"
+ ],
+ [
+ "T",
+ "ype"
+ ],
+ [
+ "le",
+ "d"
+ ],
+ [
+ "l",
+ "ed"
+ ],
+ [
+ "fu",
+ "l"
+ ],
+ [
+ "f",
+ "ul"
+ ],
+ [
+ "▁s",
+ "elf"
+ ],
+ [
+ "▁sel",
+ "f"
+ ],
+ [
+ "▁",
+ "self"
+ ],
+ [
+ "o",
+ "f"
+ ],
+ [
+ "▁A",
+ "r"
+ ],
+ [
+ "▁",
+ "Ar"
+ ],
+ [
+ "qu",
+ "e"
+ ],
+ [
+ "q",
+ "ue"
+ ],
+ [
+ "▁e",
+ "very"
+ ],
+ [
+ "▁ev",
+ "ery"
+ ],
+ [
+ "▁ever",
+ "y"
+ ],
+ [
+ "▁",
+ "every"
+ ],
+ [
+ "re",
+ "f"
+ ],
+ [
+ "r",
+ "ef"
+ ],
+ [
+ "Th",
+ "e"
+ ],
+ [
+ "T",
+ "he"
+ ],
+ [
+ "▁A",
+ "nd"
+ ],
+ [
+ "▁An",
+ "d"
+ ],
+ [
+ "▁",
+ "And"
+ ],
+ [
+ "▁re",
+ "l"
+ ],
+ [
+ "▁r",
+ "el"
+ ],
+ [
+ "▁",
+ "rel"
+ ],
+ [
+ "O",
+ "R"
+ ],
+ [
+ "I",
+ "d"
+ ],
+ [
+ "▁e",
+ "ven"
+ ],
+ [
+ "▁ev",
+ "en"
+ ],
+ [
+ "▁",
+ "even"
+ ],
+ [
+ "E",
+ "N"
+ ],
+ [
+ "▁h",
+ "and"
+ ],
+ [
+ "▁ha",
+ "nd"
+ ],
+ [
+ "▁han",
+ "d"
+ ],
+ [
+ "▁",
+ "hand"
+ ],
+ [
+ "ai",
+ "t"
+ ],
+ [
+ "a",
+ "it"
+ ],
+ [
+ "▁sh",
+ "ould"
+ ],
+ [
+ "▁",
+ "should"
+ ],
+ [
+ "▁a",
+ "fter"
+ ],
+ [
+ "▁af",
+ "ter"
+ ],
+ [
+ "▁",
+ "after"
+ ],
+ [
+ "▁d",
+ "if"
+ ],
+ [
+ "▁di",
+ "f"
+ ],
+ [
+ "gh",
+ "t"
+ ],
+ [
+ "g",
+ "ht"
+ ],
+ [
+ "if",
+ "e"
+ ],
+ [
+ "i",
+ "fe"
+ ],
+ [
+ "at",
+ "or"
+ ],
+ [
+ "ato",
+ "r"
+ ],
+ [
+ "a",
+ "tor"
+ ],
+ [
+ "as",
+ "h"
+ ],
+ [
+ "a",
+ "sh"
+ ],
+ [
+ "ri",
+ "but"
+ ],
+ [
+ "rib",
+ "ut"
+ ],
+ [
+ "ribu",
+ "t"
+ ],
+ [
+ "um",
+ "ber"
+ ],
+ [
+ "umb",
+ "er"
+ ],
+ [
+ "u",
+ "mber"
+ ],
+ [
+ "▁s",
+ "ee"
+ ],
+ [
+ "▁se",
+ "e"
+ ],
+ [
+ "▁",
+ "see"
+ ],
+ [
+ "m",
+ "s"
+ ],
+ [
+ "▁c",
+ "all"
+ ],
+ [
+ "▁cal",
+ "l"
+ ],
+ [
+ "▁ca",
+ "ll"
+ ],
+ [
+ "▁",
+ "call"
+ ],
+ [
+ "y",
+ "n"
+ ],
+ [
+ "d",
+ "d"
+ ],
+ [
+ "▁e",
+ "s"
+ ],
+ [
+ "▁",
+ "es"
+ ],
+ [
+ "▁m",
+ "ake"
+ ],
+ [
+ "▁ma",
+ "ke"
+ ],
+ [
+ "▁",
+ "make"
+ ],
+ [
+ "ot",
+ "her"
+ ],
+ [
+ "oth",
+ "er"
+ ],
+ [
+ "othe",
+ "r"
+ ],
+ [
+ "o",
+ "ther"
+ ],
+ [
+ "▁",
+ "—"
+ ],
+ [
+ "\")",
+ ";"
+ ],
+ [
+ "\"",
+ ");"
+ ],
+ [
+ "st",
+ "r"
+ ],
+ [
+ "s",
+ "tr"
+ ],
+ [
+ "▁l",
+ "ong"
+ ],
+ [
+ "▁lo",
+ "ng"
+ ],
+ [
+ "▁lon",
+ "g"
+ ],
+ [
+ "▁",
+ "long"
+ ],
+ [
+ "le",
+ "ment"
+ ],
+ [
+ "lem",
+ "ent"
+ ],
+ [
+ "l",
+ "ement"
+ ],
+ [
+ "▁w",
+ "or"
+ ],
+ [
+ "▁wo",
+ "r"
+ ],
+ [
+ "▁",
+ "wor"
+ ],
+ [
+ "it",
+ "s"
+ ],
+ [
+ "i",
+ "ts"
+ ],
+ [
+ "▁I",
+ "f"
+ ],
+ [
+ "▁",
+ "If"
+ ],
+ [
+ "al",
+ "se"
+ ],
+ [
+ "als",
+ "e"
+ ],
+ [
+ "л",
+ "ь"
+ ],
+ [
+ "wa",
+ "rd"
+ ],
+ [
+ "war",
+ "d"
+ ],
+ [
+ "w",
+ "ard"
+ ],
+ [
+ "▁п",
+ "о"
+ ],
+ [
+ "▁",
+ "по"
+ ],
+ [
+ "va",
+ "l"
+ ],
+ [
+ "v",
+ "al"
+ ],
+ [
+ "on",
+ "s"
+ ],
+ [
+ "o",
+ "ns"
+ ],
+ [
+ "▁",
+ "Z"
+ ],
+ [
+ "▁n",
+ "ow"
+ ],
+ [
+ "▁no",
+ "w"
+ ],
+ [
+ "▁",
+ "now"
+ ],
+ [
+ "da",
+ "ta"
+ ],
+ [
+ "dat",
+ "a"
+ ],
+ [
+ "d",
+ "ata"
+ ],
+ [
+ "am",
+ "p"
+ ],
+ [
+ "a",
+ "mp"
+ ],
+ [
+ "en",
+ "se"
+ ],
+ [
+ "ens",
+ "e"
+ ],
+ [
+ "▁th",
+ "rough"
+ ],
+ [
+ "▁thr",
+ "ough"
+ ],
+ [
+ "▁thro",
+ "ugh"
+ ],
+ [
+ "▁",
+ "through"
+ ],
+ [
+ "▁d",
+ "own"
+ ],
+ [
+ "▁do",
+ "wn"
+ ],
+ [
+ "▁dow",
+ "n"
+ ],
+ [
+ "▁",
+ "down"
+ ],
+ [
+ "at",
+ "t"
+ ],
+ [
+ "a",
+ "tt"
+ ],
+ [
+ "▁st",
+ "atic"
+ ],
+ [
+ "▁stat",
+ "ic"
+ ],
+ [
+ "▁",
+ "static"
+ ],
+ [
+ "ic",
+ "s"
+ ],
+ [
+ "i",
+ "cs"
+ ],
+ [
+ "#",
+ "#"
+ ],
+ [
+ "po",
+ "s"
+ ],
+ [
+ "p",
+ "os"
+ ],
+ [
+ "▁v",
+ "oid"
+ ],
+ [
+ "▁vo",
+ "id"
+ ],
+ [
+ "▁",
+ "void"
+ ],
+ [
+ "a",
+ "w"
+ ],
+ [
+ "ou",
+ "n"
+ ],
+ [
+ "o",
+ "un"
+ ],
+ [
+ "▁w",
+ "ay"
+ ],
+ [
+ "▁wa",
+ "y"
+ ],
+ [
+ "▁",
+ "way"
+ ],
+ [
+ "ib",
+ "le"
+ ],
+ [
+ "i",
+ "ble"
+ ],
+ [
+ "ve",
+ "nt"
+ ],
+ [
+ "ven",
+ "t"
+ ],
+ [
+ "v",
+ "ent"
+ ],
+ [
+ "ow",
+ "er"
+ ],
+ [
+ "owe",
+ "r"
+ ],
+ [
+ "o",
+ "wer"
+ ],
+ [
+ "▁th",
+ "ink"
+ ],
+ [
+ "▁thin",
+ "k"
+ ],
+ [
+ "▁",
+ "think"
+ ],
+ [
+ "t",
+ "s"
+ ],
+ [
+ "*",
+ "/"
+ ],
+ [
+ "▁a",
+ "gain"
+ ],
+ [
+ "▁ag",
+ "ain"
+ ],
+ [
+ "▁",
+ "again"
+ ],
+ [
+ "at",
+ "ing"
+ ],
+ [
+ "ati",
+ "ng"
+ ],
+ [
+ "atin",
+ "g"
+ ],
+ [
+ "a",
+ "ting"
+ ],
+ [
+ "т",
+ "е"
+ ],
+ [
+ "ne",
+ "r"
+ ],
+ [
+ "n",
+ "er"
+ ],
+ [
+ "▁m",
+ "ost"
+ ],
+ [
+ "▁mo",
+ "st"
+ ],
+ [
+ "▁mos",
+ "t"
+ ],
+ [
+ "▁",
+ "most"
+ ],
+ [
+ "li",
+ "ne"
+ ],
+ [
+ "lin",
+ "e"
+ ],
+ [
+ "l",
+ "ine"
+ ],
+ [
+ "y",
+ "m"
+ ],
+ [
+ "▁s",
+ "ub"
+ ],
+ [
+ "▁su",
+ "b"
+ ],
+ [
+ "▁",
+ "sub"
+ ],
+ [
+ "er",
+ "son"
+ ],
+ [
+ "ers",
+ "on"
+ ],
+ [
+ "▁re",
+ "qu"
+ ],
+ [
+ "▁r",
+ "equ"
+ ],
+ [
+ "▁req",
+ "u"
+ ],
+ [
+ "▁",
+ "requ"
+ ],
+ [
+ "A",
+ "L"
+ ],
+ [
+ "A",
+ "R"
+ ],
+ [
+ "ab",
+ "el"
+ ],
+ [
+ "abe",
+ "l"
+ ],
+ [
+ "a",
+ "bel"
+ ],
+ [
+ "on",
+ "d"
+ ],
+ [
+ "o",
+ "nd"
+ ],
+ [
+ "))",
+ ";"
+ ],
+ [
+ ")",
+ ");"
+ ],
+ [
+ "▁S",
+ "e"
+ ],
+ [
+ "▁",
+ "Se"
+ ],
+ [
+ "▁B",
+ "ut"
+ ],
+ [
+ "▁Bu",
+ "t"
+ ],
+ [
+ "▁",
+ "But"
+ ],
+ [
+ "al",
+ "k"
+ ],
+ [
+ "▁A",
+ "n"
+ ],
+ [
+ "▁",
+ "An"
+ ],
+ [
+ "ne",
+ "w"
+ ],
+ [
+ "n",
+ "ew"
+ ],
+ [
+ "▁b",
+ "ecause"
+ ],
+ [
+ "▁bec",
+ "ause"
+ ],
+ [
+ "▁",
+ "because"
+ ],
+ [
+ "ge",
+ "r"
+ ],
+ [
+ "g",
+ "er"
+ ],
+ [
+ "ul",
+ "ar"
+ ],
+ [
+ "ula",
+ "r"
+ ],
+ [
+ "u",
+ "lar"
+ ],
+ [
+ "ro",
+ "up"
+ ],
+ [
+ "rou",
+ "p"
+ ],
+ [
+ "r",
+ "oup"
+ ],
+ [
+ "t",
+ "a"
+ ],
+ [
+ "..",
+ "."
+ ],
+ [
+ ".",
+ ".."
+ ],
+ [
+ "▁c",
+ "ons"
+ ],
+ [
+ "▁con",
+ "s"
+ ],
+ [
+ "▁co",
+ "ns"
+ ],
+ [
+ "▁",
+ "cons"
+ ],
+ [
+ "▁r",
+ "ight"
+ ],
+ [
+ "▁ri",
+ "ght"
+ ],
+ [
+ "▁rig",
+ "ht"
+ ],
+ [
+ "▁",
+ "right"
+ ],
+ [
+ "▁f",
+ "r"
+ ],
+ [
+ "▁",
+ "fr"
+ ],
+ [
+ "b",
+ "e"
+ ],
+ [
+ "il",
+ "y"
+ ],
+ [
+ "i",
+ "ly"
+ ],
+ [
+ "к",
+ "и"
+ ],
+ [
+ "▁p",
+ "h"
+ ],
+ [
+ "▁",
+ "ph"
+ ],
+ [
+ "ea",
+ "d"
+ ],
+ [
+ "e",
+ "ad"
+ ],
+ [
+ "?",
+ "\""
+ ],
+ [
+ "▁g",
+ "u"
+ ],
+ [
+ "▁",
+ "gu"
+ ],
+ [
+ "▁el",
+ "se"
+ ],
+ [
+ "▁els",
+ "e"
+ ],
+ [
+ "▁",
+ "else"
+ ],
+ [
+ "▁s",
+ "om"
+ ],
+ [
+ "▁so",
+ "m"
+ ],
+ [
+ "▁",
+ "som"
+ ],
+ [
+ "re",
+ "nt"
+ ],
+ [
+ "ren",
+ "t"
+ ],
+ [
+ "r",
+ "ent"
+ ],
+ [
+ "c",
+ "o"
+ ],
+ [
+ "em",
+ "ent"
+ ],
+ [
+ "eme",
+ "nt"
+ ],
+ [
+ "emen",
+ "t"
+ ],
+ [
+ "e",
+ "ment"
+ ],
+ [
+ "▁s",
+ "tr"
+ ],
+ [
+ "▁st",
+ "r"
+ ],
+ [
+ "▁",
+ "str"
+ ],
+ [
+ "au",
+ "lt"
+ ],
+ [
+ "aul",
+ "t"
+ ],
+ [
+ "a",
+ "ult"
+ ],
+ [
+ "▁",
+ "з"
+ ],
+ [
+ "л",
+ "о"
+ ],
+ [
+ "se",
+ "rt"
+ ],
+ [
+ "ser",
+ "t"
+ ],
+ [
+ "s",
+ "ert"
+ ],
+ [
+ "va",
+ "r"
+ ],
+ [
+ "v",
+ "ar"
+ ],
+ [
+ "ty",
+ "pe"
+ ],
+ [
+ "typ",
+ "e"
+ ],
+ [
+ "t",
+ "ype"
+ ],
+ [
+ "▁C",
+ "om"
+ ],
+ [
+ "▁Co",
+ "m"
+ ],
+ [
+ "▁",
+ "Com"
+ ],
+ [
+ "л",
+ "е"
+ ],
+ [
+ "in",
+ "s"
+ ],
+ [
+ "i",
+ "ns"
+ ],
+ [
+ "m",
+ "e"
+ ],
+ [
+ "wa",
+ "y"
+ ],
+ [
+ "w",
+ "ay"
+ ],
+ [
+ "id",
+ "ent"
+ ],
+ [
+ "ide",
+ "nt"
+ ],
+ [
+ "iden",
+ "t"
+ ],
+ [
+ "▁p",
+ "rov"
+ ],
+ [
+ "▁pro",
+ "v"
+ ],
+ [
+ "▁pr",
+ "ov"
+ ],
+ [
+ "▁",
+ "prov"
+ ],
+ [
+ "▁",
+ "м"
+ ],
+ [
+ "▁tr",
+ "ue"
+ ],
+ [
+ "▁",
+ "true"
+ ],
+ [
+ "▁P",
+ "ro"
+ ],
+ [
+ "▁Pr",
+ "o"
+ ],
+ [
+ "▁",
+ "Pro"
+ ],
+ [
+ "f",
+ "l"
+ ],
+ [
+ "▁s",
+ "l"
+ ],
+ [
+ "▁",
+ "sl"
+ ],
+ [
+ "▁A",
+ "s"
+ ],
+ [
+ "▁",
+ "As"
+ ],
+ [
+ "}",
+ "\\"
+ ],
+ [
+ "I",
+ "D"
+ ],
+ [
+ "ue",
+ "s"
+ ],
+ [
+ "u",
+ "es"
+ ],
+ [
+ "▁in",
+ "st"
+ ],
+ [
+ "▁ins",
+ "t"
+ ],
+ [
+ "▁",
+ "inst"
+ ],
+ [
+ "▁n",
+ "ame"
+ ],
+ [
+ "▁na",
+ "me"
+ ],
+ [
+ "▁nam",
+ "e"
+ ],
+ [
+ "▁",
+ "name"
+ ],
+ [
+ "o",
+ "x"
+ ],
+ [
+ "▁",
+ ")"
+ ],
+ [
+ "l",
+ "i"
+ ],
+ [
+ "am",
+ "es"
+ ],
+ [
+ "ame",
+ "s"
+ ],
+ [
+ "a",
+ "mes"
+ ],
+ [
+ "Re",
+ "s"
+ ],
+ [
+ "R",
+ "es"
+ ],
+ [
+ "▁s",
+ "ur"
+ ],
+ [
+ "▁su",
+ "r"
+ ],
+ [
+ "▁",
+ "sur"
+ ],
+ [
+ "par",
+ "am"
+ ],
+ [
+ "pa",
+ "ram"
+ ],
+ [
+ "para",
+ "m"
+ ],
+ [
+ "p",
+ "aram"
+ ],
+ [
+ "▁st",
+ "art"
+ ],
+ [
+ "▁star",
+ "t"
+ ],
+ [
+ "▁sta",
+ "rt"
+ ],
+ [
+ "▁",
+ "start"
+ ],
+ [
+ "a",
+ "j"
+ ],
+ [
+ "S",
+ "E"
+ ],
+ [
+ "as",
+ "k"
+ ],
+ [
+ "a",
+ "sk"
+ ],
+ [
+ "I",
+ "T"
+ ],
+ [
+ "St",
+ "ring"
+ ],
+ [
+ "Str",
+ "ing"
+ ],
+ [
+ "S",
+ "tring"
+ ],
+ [
+ "▁a",
+ "ss"
+ ],
+ [
+ "▁as",
+ "s"
+ ],
+ [
+ "▁",
+ "ass"
+ ],
+ [
+ "▁p",
+ "lay"
+ ],
+ [
+ "▁pl",
+ "ay"
+ ],
+ [
+ "▁",
+ "play"
+ ],
+ [
+ "ti",
+ "ng"
+ ],
+ [
+ "t",
+ "ing"
+ ],
+ [
+ "to",
+ "n"
+ ],
+ [
+ "t",
+ "on"
+ ],
+ [
+ "▁b",
+ "efore"
+ ],
+ [
+ "▁be",
+ "fore"
+ ],
+ [
+ "▁bef",
+ "ore"
+ ],
+ [
+ "▁",
+ "before"
+ ],
+ [
+ "▁p",
+ "ol"
+ ],
+ [
+ "▁po",
+ "l"
+ ],
+ [
+ "▁",
+ "pol"
+ ],
+ [
+ "ar",
+ "ch"
+ ],
+ [
+ "arc",
+ "h"
+ ],
+ [
+ "▁w",
+ "ell"
+ ],
+ [
+ "▁we",
+ "ll"
+ ],
+ [
+ "▁wel",
+ "l"
+ ],
+ [
+ "▁",
+ "well"
+ ],
+ [
+ "Co",
+ "m"
+ ],
+ [
+ "C",
+ "om"
+ ],
+ [
+ "an",
+ "y"
+ ],
+ [
+ "a",
+ "ny"
+ ],
+ [
+ "ol",
+ "og"
+ ],
+ [
+ "olo",
+ "g"
+ ],
+ [
+ "o",
+ "log"
+ ],
+ [
+ "▁e",
+ "rr"
+ ],
+ [
+ "▁er",
+ "r"
+ ],
+ [
+ "▁",
+ "err"
+ ],
+ [
+ "▁the",
+ "se"
+ ],
+ [
+ "▁th",
+ "ese"
+ ],
+ [
+ "ar",
+ "s"
+ ],
+ [
+ "a",
+ "rs"
+ ],
+ [
+ "e",
+ "b"
+ ],
+ [
+ "▁b",
+ "r"
+ ],
+ [
+ "▁",
+ "br"
+ ],
+ [
+ "▁in",
+ "cl"
+ ],
+ [
+ "▁inc",
+ "l"
+ ],
+ [
+ "▁",
+ "incl"
+ ],
+ [
+ "▁h",
+ "el"
+ ],
+ [
+ "▁he",
+ "l"
+ ],
+ [
+ "▁",
+ "hel"
+ ],
+ [
+ "er",
+ "n"
+ ],
+ [
+ "e",
+ "rn"
+ ],
+ [
+ "od",
+ "y"
+ ],
+ [
+ "o",
+ "dy"
+ ],
+ [
+ "в",
+ "о"
+ ],
+ [
+ "▁in",
+ "d"
+ ],
+ [
+ "▁i",
+ "nd"
+ ],
+ [
+ "▁",
+ "ind"
+ ],
+ [
+ "--",
+ "--------------"
+ ],
+ [
+ "----",
+ "------------"
+ ],
+ [
+ "--------",
+ "--------"
+ ],
+ [
+ "---",
+ "-------------"
+ ],
+ [
+ "------------",
+ "----"
+ ],
+ [
+ "-----",
+ "-----------"
+ ],
+ [
+ "----------",
+ "------"
+ ],
+ [
+ "------",
+ "----------"
+ ],
+ [
+ "-------------",
+ "---"
+ ],
+ [
+ "--------------",
+ "--"
+ ],
+ [
+ "---------",
+ "-------"
+ ],
+ [
+ "-------",
+ "---------"
+ ],
+ [
+ "-----------",
+ "-----"
+ ],
+ [
+ "▁d",
+ "ata"
+ ],
+ [
+ "▁da",
+ "ta"
+ ],
+ [
+ "▁dat",
+ "a"
+ ],
+ [
+ "▁",
+ "data"
+ ],
+ [
+ "▁g",
+ "ood"
+ ],
+ [
+ "▁go",
+ "od"
+ ],
+ [
+ "▁",
+ "good"
+ ],
+ [
+ "L",
+ "E"
+ ],
+ [
+ "]",
+ ","
+ ],
+ [
+ "▁a",
+ "v"
+ ],
+ [
+ "▁",
+ "av"
+ ],
+ [
+ "▁a",
+ "c"
+ ],
+ [
+ "▁",
+ "ac"
+ ],
+ [
+ "id",
+ "er"
+ ],
+ [
+ "ide",
+ "r"
+ ],
+ [
+ "i",
+ "der"
+ ],
+ [
+ "н",
+ "е"
+ ],
+ [
+ "▁",
+ "Q"
+ ],
+ [
+ "▁m",
+ "in"
+ ],
+ [
+ "▁mi",
+ "n"
+ ],
+ [
+ "▁",
+ "min"
+ ],
+ [
+ "▁m",
+ "uch"
+ ],
+ [
+ "▁mu",
+ "ch"
+ ],
+ [
+ "c",
+ "i"
+ ],
+ [
+ "el",
+ "s"
+ ],
+ [
+ "e",
+ "ls"
+ ],
+ [
+ "▁c",
+ "ur"
+ ],
+ [
+ "▁cu",
+ "r"
+ ],
+ [
+ "▁",
+ "cur"
+ ],
+ [
+ "▁v",
+ "alue"
+ ],
+ [
+ "▁val",
+ "ue"
+ ],
+ [
+ "▁",
+ "value"
+ ],
+ [
+ "er",
+ "y"
+ ],
+ [
+ "e",
+ "ry"
+ ],
+ [
+ "u",
+ "f"
+ ],
+ [
+ "▁l",
+ "oc"
+ ],
+ [
+ "▁lo",
+ "c"
+ ],
+ [
+ "▁",
+ "loc"
+ ],
+ [
+ "re",
+ "ak"
+ ],
+ [
+ "rea",
+ "k"
+ ],
+ [
+ "at",
+ "ive"
+ ],
+ [
+ "ati",
+ "ve"
+ ],
+ [
+ "ativ",
+ "e"
+ ],
+ [
+ "im",
+ "es"
+ ],
+ [
+ "ime",
+ "s"
+ ],
+ [
+ "i",
+ "mes"
+ ],
+ [
+ "C",
+ "l"
+ ],
+ [
+ "▁",
+ ","
+ ],
+ [
+ "▁s",
+ "er"
+ ],
+ [
+ "▁se",
+ "r"
+ ],
+ [
+ "▁",
+ "ser"
+ ],
+ [
+ "▁d",
+ "ie"
+ ],
+ [
+ "▁di",
+ "e"
+ ],
+ [
+ "▁",
+ "die"
+ ],
+ [
+ "▁tr",
+ "ans"
+ ],
+ [
+ "▁tra",
+ "ns"
+ ],
+ [
+ "▁",
+ "trans"
+ ],
+ [
+ "▁res",
+ "ult"
+ ],
+ [
+ "▁",
+ "result"
+ ],
+ [
+ "ex",
+ "t"
+ ],
+ [
+ "e",
+ "xt"
+ ],
+ [
+ "▁a",
+ "ut"
+ ],
+ [
+ "▁au",
+ "t"
+ ],
+ [
+ "▁",
+ "aut"
+ ],
+ [
+ "la",
+ "nd"
+ ],
+ [
+ "lan",
+ "d"
+ ],
+ [
+ "l",
+ "and"
+ ],
+ [
+ "▁&",
+ "&"
+ ],
+ [
+ "▁",
+ "&&"
+ ],
+ [
+ "C",
+ "h"
+ ],
+ [
+ "te",
+ "n"
+ ],
+ [
+ "t",
+ "en"
+ ],
+ [
+ "}",
+ "$"
+ ],
+ [
+ "▁t",
+ "ype"
+ ],
+ [
+ "▁typ",
+ "e"
+ ],
+ [
+ "▁ty",
+ "pe"
+ ],
+ [
+ "▁",
+ "type"
+ ],
+ [
+ "con",
+ "d"
+ ],
+ [
+ "co",
+ "nd"
+ ],
+ [
+ "c",
+ "ond"
+ ],
+ [
+ "ic",
+ "es"
+ ],
+ [
+ "ice",
+ "s"
+ ],
+ [
+ "i",
+ "ces"
+ ],
+ [
+ "▁v",
+ "ery"
+ ],
+ [
+ "▁ver",
+ "y"
+ ],
+ [
+ "▁ve",
+ "ry"
+ ],
+ [
+ "▁",
+ "very"
+ ],
+ [
+ "▁o",
+ "wn"
+ ],
+ [
+ "▁",
+ "own"
+ ],
+ [
+ "▁f",
+ "il"
+ ],
+ [
+ "▁fi",
+ "l"
+ ],
+ [
+ "▁",
+ "fil"
+ ],
+ [
+ "it",
+ "ies"
+ ],
+ [
+ "iti",
+ "es"
+ ],
+ [
+ "i",
+ "ties"
+ ],
+ [
+ "▁p",
+ "rodu"
+ ],
+ [
+ "▁pro",
+ "du"
+ ],
+ [
+ "▁prod",
+ "u"
+ ],
+ [
+ "▁",
+ "produ"
+ ],
+ [
+ "▁re",
+ "ad"
+ ],
+ [
+ "▁r",
+ "ead"
+ ],
+ [
+ "▁",
+ "read"
+ ],
+ [
+ "▁f",
+ "orm"
+ ],
+ [
+ "▁for",
+ "m"
+ ],
+ [
+ "▁fo",
+ "rm"
+ ],
+ [
+ "▁",
+ "form"
+ ],
+ [
+ "▁c",
+ "ase"
+ ],
+ [
+ "▁cas",
+ "e"
+ ],
+ [
+ "▁ca",
+ "se"
+ ],
+ [
+ "▁",
+ "case"
+ ],
+ [
+ "at",
+ "her"
+ ],
+ [
+ "ath",
+ "er"
+ ],
+ [
+ "a",
+ "ther"
+ ],
+ [
+ "т",
+ "и"
+ ],
+ [
+ "д",
+ "а"
+ ],
+ [
+ "е",
+ "р"
+ ],
+ [
+ "T",
+ "h"
+ ],
+ [
+ "au",
+ "t"
+ ],
+ [
+ "a",
+ "ut"
+ ],
+ [
+ "▁s",
+ "pec"
+ ],
+ [
+ "▁sp",
+ "ec"
+ ],
+ [
+ "▁spe",
+ "c"
+ ],
+ [
+ "▁",
+ "spec"
+ ],
+ [
+ "i",
+ "j"
+ ],
+ [
+ "b",
+ "l"
+ ],
+ [
+ "il",
+ "ity"
+ ],
+ [
+ "ili",
+ "ty"
+ ],
+ [
+ "▁",
+ "é"
+ ],
+ [
+ "▁e",
+ "r"
+ ],
+ [
+ "▁",
+ "er"
+ ],
+ [
+ "▁d",
+ "oes"
+ ],
+ [
+ "▁do",
+ "es"
+ ],
+ [
+ "▁",
+ "does"
+ ],
+ [
+ "▁h",
+ "ere"
+ ],
+ [
+ "▁he",
+ "re"
+ ],
+ [
+ "▁her",
+ "e"
+ ],
+ [
+ "▁",
+ "here"
+ ],
+ [
+ "th",
+ "e"
+ ],
+ [
+ "t",
+ "he"
+ ],
+ [
+ "ur",
+ "es"
+ ],
+ [
+ "ure",
+ "s"
+ ],
+ [
+ "u",
+ "res"
+ ],
+ [
+ "▁",
+ "%"
+ ],
+ [
+ "mi",
+ "n"
+ ],
+ [
+ "m",
+ "in"
+ ],
+ [
+ "▁n",
+ "ull"
+ ],
+ [
+ "▁nu",
+ "ll"
+ ],
+ [
+ "▁",
+ "null"
+ ],
+ [
+ "ra",
+ "p"
+ ],
+ [
+ "r",
+ "ap"
+ ],
+ [
+ "\"",
+ ")"
+ ],
+ [
+ "r",
+ "r"
+ ],
+ [
+ "Li",
+ "st"
+ ],
+ [
+ "L",
+ "ist"
+ ],
+ [
+ "ri",
+ "ght"
+ ],
+ [
+ "rig",
+ "ht"
+ ],
+ [
+ "r",
+ "ight"
+ ],
+ [
+ "▁U",
+ "ser"
+ ],
+ [
+ "▁Us",
+ "er"
+ ],
+ [
+ "▁Use",
+ "r"
+ ],
+ [
+ "▁",
+ "User"
+ ],
+ [
+ "U",
+ "L"
+ ],
+ [
+ "at",
+ "ional"
+ ],
+ [
+ "ation",
+ "al"
+ ],
+ [
+ "ati",
+ "onal"
+ ],
+ [
+ "atio",
+ "nal"
+ ],
+ [
+ "▁b",
+ "eing"
+ ],
+ [
+ "▁be",
+ "ing"
+ ],
+ [
+ "▁bei",
+ "ng"
+ ],
+ [
+ "▁",
+ "being"
+ ],
+ [
+ "A",
+ "N"
+ ],
+ [
+ "s",
+ "k"
+ ],
+ [
+ "▁c",
+ "ar"
+ ],
+ [
+ "▁ca",
+ "r"
+ ],
+ [
+ "▁",
+ "car"
+ ],
+ [
+ "ol",
+ "e"
+ ],
+ [
+ "o",
+ "le"
+ ],
+ [
+ "▁d",
+ "ist"
+ ],
+ [
+ "▁dis",
+ "t"
+ ],
+ [
+ "▁di",
+ "st"
+ ],
+ [
+ "▁",
+ "dist"
+ ],
+ [
+ "pl",
+ "ic"
+ ],
+ [
+ "p",
+ "lic"
+ ],
+ [
+ "ol",
+ "low"
+ ],
+ [
+ "oll",
+ "ow"
+ ],
+ [
+ "▁p",
+ "res"
+ ],
+ [
+ "▁pre",
+ "s"
+ ],
+ [
+ "▁pr",
+ "es"
+ ],
+ [
+ "▁",
+ "pres"
+ ],
+ [
+ "▁s",
+ "uch"
+ ],
+ [
+ "▁su",
+ "ch"
+ ],
+ [
+ "▁suc",
+ "h"
+ ],
+ [
+ "▁",
+ "such"
+ ],
+ [
+ "re",
+ "am"
+ ],
+ [
+ "rea",
+ "m"
+ ],
+ [
+ "in",
+ "ce"
+ ],
+ [
+ "inc",
+ "e"
+ ],
+ [
+ "ga",
+ "n"
+ ],
+ [
+ "g",
+ "an"
+ ],
+ [
+ "▁F",
+ "or"
+ ],
+ [
+ "▁Fo",
+ "r"
+ ],
+ [
+ "▁",
+ "For"
+ ],
+ [
+ "\"",
+ ":"
+ ],
+ [
+ "so",
+ "n"
+ ],
+ [
+ "s",
+ "on"
+ ],
+ [
+ "riv",
+ "ate"
+ ],
+ [
+ "▁y",
+ "ears"
+ ],
+ [
+ "▁year",
+ "s"
+ ],
+ [
+ "▁ye",
+ "ars"
+ ],
+ [
+ "▁s",
+ "erv"
+ ],
+ [
+ "▁se",
+ "rv"
+ ],
+ [
+ "▁ser",
+ "v"
+ ],
+ [
+ "▁",
+ "serv"
+ ],
+ [
+ "▁m",
+ "ade"
+ ],
+ [
+ "▁ma",
+ "de"
+ ],
+ [
+ "▁mad",
+ "e"
+ ],
+ [
+ "▁",
+ "made"
+ ],
+ [
+ "de",
+ "f"
+ ],
+ [
+ "d",
+ "ef"
+ ],
+ [
+ ";",
+ "\r"
+ ],
+ [
+ "▁g",
+ "l"
+ ],
+ [
+ "▁",
+ "gl"
+ ],
+ [
+ "▁b",
+ "el"
+ ],
+ [
+ "▁be",
+ "l"
+ ],
+ [
+ "▁",
+ "bel"
+ ],
+ [
+ "▁l",
+ "ist"
+ ],
+ [
+ "▁li",
+ "st"
+ ],
+ [
+ "▁",
+ "list"
+ ],
+ [
+ "▁c",
+ "or"
+ ],
+ [
+ "▁co",
+ "r"
+ ],
+ [
+ "▁",
+ "cor"
+ ],
+ [
+ "▁d",
+ "et"
+ ],
+ [
+ "▁de",
+ "t"
+ ],
+ [
+ "▁",
+ "det"
+ ],
+ [
+ "ce",
+ "ption"
+ ],
+ [
+ "cept",
+ "ion"
+ ],
+ [
+ "eg",
+ "in"
+ ],
+ [
+ "e",
+ "gin"
+ ],
+ [
+ "▁",
+ "б"
+ ],
+ [
+ "▁c",
+ "har"
+ ],
+ [
+ "▁ch",
+ "ar"
+ ],
+ [
+ "▁cha",
+ "r"
+ ],
+ [
+ "▁",
+ "char"
+ ],
+ [
+ "tr",
+ "ans"
+ ],
+ [
+ "tra",
+ "ns"
+ ],
+ [
+ "▁f",
+ "am"
+ ],
+ [
+ "▁fa",
+ "m"
+ ],
+ [
+ "▁!",
+ "="
+ ],
+ [
+ "▁",
+ "!="
+ ],
+ [
+ "ou",
+ "se"
+ ],
+ [
+ "ous",
+ "e"
+ ],
+ [
+ "o",
+ "use"
+ ],
+ [
+ "▁d",
+ "ec"
+ ],
+ [
+ "▁de",
+ "c"
+ ],
+ [
+ "▁",
+ "dec"
+ ],
+ [
+ "ic",
+ "a"
+ ],
+ [
+ "i",
+ "ca"
+ ],
+ [
+ "▁m",
+ "any"
+ ],
+ [
+ "▁man",
+ "y"
+ ],
+ [
+ "▁ma",
+ "ny"
+ ],
+ [
+ "▁",
+ "many"
+ ],
+ [
+ "ak",
+ "ing"
+ ],
+ [
+ "aki",
+ "ng"
+ ],
+ [
+ "a",
+ "king"
+ ],
+ [
+ "▁",
+ "à"
+ ],
+ [
+ "▁s",
+ "im"
+ ],
+ [
+ "▁si",
+ "m"
+ ],
+ [
+ "▁",
+ "sim"
+ ],
+ [
+ "ag",
+ "es"
+ ],
+ [
+ "age",
+ "s"
+ ],
+ [
+ "a",
+ "ges"
+ ],
+ [
+ "uf",
+ "f"
+ ],
+ [
+ "u",
+ "ff"
+ ],
+ [
+ "as",
+ "ed"
+ ],
+ [
+ "ase",
+ "d"
+ ],
+ [
+ "a",
+ "sed"
+ ],
+ [
+ "ma",
+ "n"
+ ],
+ [
+ "m",
+ "an"
+ ],
+ [
+ "▁S",
+ "h"
+ ],
+ [
+ "▁",
+ "Sh"
+ ],
+ [
+ "ie",
+ "t"
+ ],
+ [
+ "i",
+ "et"
+ ],
+ [
+ "ir",
+ "ect"
+ ],
+ [
+ "ire",
+ "ct"
+ ],
+ [
+ "i",
+ "rect"
+ ],
+ [
+ "▁R",
+ "e"
+ ],
+ [
+ "▁",
+ "Re"
+ ],
+ [
+ "▁d",
+ "iffer"
+ ],
+ [
+ "▁dif",
+ "fer"
+ ],
+ [
+ "▁diff",
+ "er"
+ ],
+ [
+ "▁f",
+ "ind"
+ ],
+ [
+ "▁fin",
+ "d"
+ ],
+ [
+ "▁fi",
+ "nd"
+ ],
+ [
+ "▁",
+ "find"
+ ],
+ [
+ "eth",
+ "od"
+ ],
+ [
+ "▁",
+ "\r"
+ ],
+ [
+ "in",
+ "es"
+ ],
+ [
+ "ine",
+ "s"
+ ],
+ [
+ "i",
+ "nes"
+ ],
+ [
+ "▁in",
+ "v"
+ ],
+ [
+ "▁i",
+ "nv"
+ ],
+ [
+ "▁",
+ "inv"
+ ],
+ [
+ "▁p",
+ "oint"
+ ],
+ [
+ "▁po",
+ "int"
+ ],
+ [
+ "▁poi",
+ "nt"
+ ],
+ [
+ "▁",
+ "point"
+ ],
+ [
+ "▁The",
+ "y"
+ ],
+ [
+ "▁Th",
+ "ey"
+ ],
+ [
+ "▁",
+ "They"
+ ],
+ [
+ "▁u",
+ "sed"
+ ],
+ [
+ "▁us",
+ "ed"
+ ],
+ [
+ "▁use",
+ "d"
+ ],
+ [
+ "▁",
+ "used"
+ ],
+ [
+ "ct",
+ "ions"
+ ],
+ [
+ "ction",
+ "s"
+ ],
+ [
+ "▁st",
+ "ill"
+ ],
+ [
+ "i",
+ "ó"
+ ],
+ [
+ "in",
+ "ed"
+ ],
+ [
+ "ine",
+ "d"
+ ],
+ [
+ "i",
+ "ned"
+ ],
+ [
+ "▁wh",
+ "ile"
+ ],
+ [
+ "▁",
+ "while"
+ ],
+ [
+ "I",
+ "t"
+ ],
+ [
+ "em",
+ "ber"
+ ],
+ [
+ "emb",
+ "er"
+ ],
+ [
+ "e",
+ "mber"
+ ],
+ [
+ "▁s",
+ "ay"
+ ],
+ [
+ "▁sa",
+ "y"
+ ],
+ [
+ "▁",
+ "say"
+ ],
+ [
+ "▁he",
+ "lp"
+ ],
+ [
+ "▁hel",
+ "p"
+ ],
+ [
+ "▁",
+ "help"
+ ],
+ [
+ "▁c",
+ "re"
+ ],
+ [
+ "▁cr",
+ "e"
+ ],
+ [
+ "▁",
+ "cre"
+ ],
+ [
+ "▁",
+ "x"
+ ],
+ [
+ "▁T",
+ "r"
+ ],
+ [
+ "▁",
+ "Tr"
+ ],
+ [
+ "um",
+ "ent"
+ ],
+ [
+ "ume",
+ "nt"
+ ],
+ [
+ "umen",
+ "t"
+ ],
+ [
+ "u",
+ "ment"
+ ],
+ [
+ "▁s",
+ "k"
+ ],
+ [
+ "▁",
+ "sk"
+ ],
+ [
+ "ou",
+ "ght"
+ ],
+ [
+ "ough",
+ "t"
+ ],
+ [
+ "ual",
+ "ly"
+ ],
+ [
+ "u",
+ "ally"
+ ],
+ [
+ "m",
+ "essage"
+ ],
+ [
+ "▁C",
+ "on"
+ ],
+ [
+ "▁Co",
+ "n"
+ ],
+ [
+ "▁",
+ "Con"
+ ],
+ [
+ "▁m",
+ "on"
+ ],
+ [
+ "▁mo",
+ "n"
+ ],
+ [
+ "▁",
+ "mon"
+ ],
+ [
+ "ar",
+ "ed"
+ ],
+ [
+ "are",
+ "d"
+ ],
+ [
+ "a",
+ "red"
+ ],
+ [
+ "wor",
+ "k"
+ ],
+ [
+ "w",
+ "ork"
+ ],
+ [
+ ")",
+ ":"
+ ],
+ [
+ "is",
+ "ter"
+ ],
+ [
+ "ist",
+ "er"
+ ],
+ [
+ "iste",
+ "r"
+ ],
+ [
+ "i",
+ "ster"
+ ],
+ [
+ "ar",
+ "n"
+ ],
+ [
+ "a",
+ "rn"
+ ],
+ [
+ "iz",
+ "ed"
+ ],
+ [
+ "ize",
+ "d"
+ ],
+ [
+ "i",
+ "zed"
+ ],
+ [
+ "Dat",
+ "a"
+ ],
+ [
+ "Da",
+ "ta"
+ ],
+ [
+ "D",
+ "ata"
+ ],
+ [
+ "or",
+ "n"
+ ],
+ [
+ "o",
+ "rn"
+ ],
+ [
+ "▁h",
+ "ead"
+ ],
+ [
+ "▁he",
+ "ad"
+ ],
+ [
+ "▁",
+ "head"
+ ],
+ [
+ "D",
+ "E"
+ ],
+ [
+ "▁L",
+ "e"
+ ],
+ [
+ "▁",
+ "Le"
+ ],
+ [
+ "▁p",
+ "erson"
+ ],
+ [
+ "▁per",
+ "son"
+ ],
+ [
+ "▁pers",
+ "on"
+ ],
+ [
+ "▁",
+ "person"
+ ],
+ [
+ "ment",
+ "s"
+ ],
+ [
+ "men",
+ "ts"
+ ],
+ [
+ "m",
+ "ents"
+ ],
+ [
+ "eng",
+ "th"
+ ],
+ [
+ "e",
+ "ngth"
+ ],
+ [
+ "▁f",
+ "alse"
+ ],
+ [
+ "▁fal",
+ "se"
+ ],
+ [
+ "▁fals",
+ "e"
+ ],
+ [
+ "▁",
+ "false"
+ ],
+ [
+ "▁m",
+ "ed"
+ ],
+ [
+ "▁me",
+ "d"
+ ],
+ [
+ "▁",
+ "med"
+ ],
+ [
+ "▁D",
+ "e"
+ ],
+ [
+ "▁",
+ "De"
+ ],
+ [
+ "ac",
+ "he"
+ ],
+ [
+ "ach",
+ "e"
+ ],
+ [
+ "a",
+ "che"
+ ],
+ [
+ "it",
+ "ed"
+ ],
+ [
+ "ite",
+ "d"
+ ],
+ [
+ "i",
+ "ted"
+ ],
+ [
+ "▁l",
+ "et"
+ ],
+ [
+ "▁le",
+ "t"
+ ],
+ [
+ "▁",
+ "let"
+ ],
+ [
+ "▁s",
+ "how"
+ ],
+ [
+ "▁sh",
+ "ow"
+ ],
+ [
+ "▁",
+ "show"
+ ],
+ [
+ "▁s",
+ "ame"
+ ],
+ [
+ "▁sa",
+ "me"
+ ],
+ [
+ "▁sam",
+ "e"
+ ],
+ [
+ "▁",
+ "same"
+ ],
+ [
+ "us",
+ "s"
+ ],
+ [
+ "u",
+ "ss"
+ ],
+ [
+ "▁g",
+ "ener"
+ ],
+ [
+ "▁gen",
+ "er"
+ ],
+ [
+ "▁ge",
+ "ner"
+ ],
+ [
+ "▁gene",
+ "r"
+ ],
+ [
+ "▁",
+ "gener"
+ ],
+ [
+ "▁",
+ "у"
+ ],
+ [
+ "cu",
+ "r"
+ ],
+ [
+ "c",
+ "ur"
+ ],
+ [
+ "▁re",
+ "al"
+ ],
+ [
+ "▁",
+ "real"
+ ],
+ [
+ "ce",
+ "d"
+ ],
+ [
+ "c",
+ "ed"
+ ],
+ [
+ "\"",
+ ">"
+ ],
+ [
+ "st",
+ "ruct"
+ ],
+ [
+ "str",
+ "uct"
+ ],
+ [
+ "stru",
+ "ct"
+ ],
+ [
+ "be",
+ "gin"
+ ],
+ [
+ "b",
+ "egin"
+ ],
+ [
+ "ce",
+ "pt"
+ ],
+ [
+ "cep",
+ "t"
+ ],
+ [
+ "▁b",
+ "o"
+ ],
+ [
+ "▁",
+ "bo"
+ ],
+ [
+ "ir",
+ "ed"
+ ],
+ [
+ "ire",
+ "d"
+ ],
+ [
+ "i",
+ "red"
+ ],
+ [
+ "▁F",
+ "r"
+ ],
+ [
+ "▁",
+ "Fr"
+ ],
+ [
+ "▁st",
+ "ud"
+ ],
+ [
+ "▁",
+ "stud"
+ ],
+ [
+ "de",
+ "v"
+ ],
+ [
+ "d",
+ "ev"
+ ],
+ [
+ "A",
+ "r"
+ ],
+ [
+ "(",
+ "\\"
+ ],
+ [
+ "▁C",
+ "l"
+ ],
+ [
+ "▁",
+ "Cl"
+ ],
+ [
+ "we",
+ "en"
+ ],
+ [
+ "w",
+ "een"
+ ],
+ [
+ "▁t",
+ "oo"
+ ],
+ [
+ "▁to",
+ "o"
+ ],
+ [
+ "▁",
+ "too"
+ ],
+ [
+ "▁t",
+ "est"
+ ],
+ [
+ "▁te",
+ "st"
+ ],
+ [
+ "▁",
+ "test"
+ ],
+ [
+ "▁d",
+ "ay"
+ ],
+ [
+ "▁da",
+ "y"
+ ],
+ [
+ "▁",
+ "day"
+ ],
+ [
+ "o",
+ "h"
+ ],
+ [
+ "▁f",
+ "ollow"
+ ],
+ [
+ "▁fol",
+ "low"
+ ],
+ [
+ "▁",
+ "follow"
+ ],
+ [
+ "at",
+ "ure"
+ ],
+ [
+ "atur",
+ "e"
+ ],
+ [
+ "atu",
+ "re"
+ ],
+ [
+ "z",
+ "e"
+ ],
+ [
+ "ie",
+ "n"
+ ],
+ [
+ "i",
+ "en"
+ ],
+ [
+ "re",
+ "g"
+ ],
+ [
+ "r",
+ "eg"
+ ],
+ [
+ "ce",
+ "s"
+ ],
+ [
+ "c",
+ "es"
+ ],
+ [
+ "ur",
+ "ing"
+ ],
+ [
+ "uri",
+ "ng"
+ ],
+ [
+ "u",
+ "ring"
+ ],
+ [
+ "am",
+ "b"
+ ],
+ [
+ "a",
+ "mb"
+ ],
+ [
+ "in",
+ "a"
+ ],
+ [
+ "i",
+ "na"
+ ],
+ [
+ "cr",
+ "i"
+ ],
+ [
+ "c",
+ "ri"
+ ],
+ [
+ "▁e",
+ "d"
+ ],
+ [
+ "▁",
+ "ed"
+ ],
+ [
+ "S",
+ "S"
+ ],
+ [
+ "uc",
+ "k"
+ ],
+ [
+ "u",
+ "ck"
+ ],
+ [
+ "▁/",
+ "*"
+ ],
+ [
+ "▁",
+ "/*"
+ ],
+ [
+ "C",
+ "T"
+ ],
+ [
+ "▁T",
+ "here"
+ ],
+ [
+ "▁The",
+ "re"
+ ],
+ [
+ "▁Th",
+ "ere"
+ ],
+ [
+ "▁Ther",
+ "e"
+ ],
+ [
+ "▁",
+ "There"
+ ],
+ [
+ "▁t",
+ "ake"
+ ],
+ [
+ "▁tak",
+ "e"
+ ],
+ [
+ "▁ta",
+ "ke"
+ ],
+ [
+ "▁",
+ "take"
+ ],
+ [
+ "pa",
+ "r"
+ ],
+ [
+ "p",
+ "ar"
+ ],
+ [
+ "ul",
+ "e"
+ ],
+ [
+ "u",
+ "le"
+ ],
+ [
+ "ca",
+ "l"
+ ],
+ [
+ "c",
+ "al"
+ ],
+ [
+ "fo",
+ "r"
+ ],
+ [
+ "f",
+ "or"
+ ],
+ [
+ "**",
+ "**************"
+ ],
+ [
+ "****",
+ "************"
+ ],
+ [
+ "********",
+ "********"
+ ],
+ [
+ "************",
+ "****"
+ ],
+ [
+ "**************",
+ "**"
+ ],
+ [
+ "s",
+ "ource"
+ ],
+ [
+ "▁th",
+ "ose"
+ ],
+ [
+ "co",
+ "l"
+ ],
+ [
+ "c",
+ "ol"
+ ],
+ [
+ "▁e",
+ "ff"
+ ],
+ [
+ "▁",
+ "eff"
+ ],
+ [
+ "mo",
+ "d"
+ ],
+ [
+ "m",
+ "od"
+ ],
+ [
+ "con",
+ "t"
+ ],
+ [
+ "co",
+ "nt"
+ ],
+ [
+ "c",
+ "ont"
+ ],
+ [
+ "}",
+ "{"
+ ],
+ [
+ "▁a",
+ "round"
+ ],
+ [
+ "▁ar",
+ "ound"
+ ],
+ [
+ "▁",
+ "around"
+ ],
+ [
+ "pr",
+ "ess"
+ ],
+ [
+ "pre",
+ "ss"
+ ],
+ [
+ "pres",
+ "s"
+ ],
+ [
+ "p",
+ "ress"
+ ],
+ [
+ "b",
+ "y"
+ ],
+ [
+ "▁go",
+ "ing"
+ ],
+ [
+ "▁",
+ "going"
+ ],
+ [
+ "pon",
+ "se"
+ ],
+ [
+ "pons",
+ "e"
+ ],
+ [
+ "▁",
+ "С"
+ ],
+ [
+ "▁l",
+ "ine"
+ ],
+ [
+ "▁li",
+ "ne"
+ ],
+ [
+ "▁lin",
+ "e"
+ ],
+ [
+ "▁",
+ "line"
+ ],
+ [
+ "da",
+ "te"
+ ],
+ [
+ "dat",
+ "e"
+ ],
+ [
+ "d",
+ "ate"
+ ],
+ [
+ "co",
+ "de"
+ ],
+ [
+ "cod",
+ "e"
+ ],
+ [
+ "c",
+ "ode"
+ ],
+ [
+ "[",
+ "'"
+ ],
+ [
+ "▁l",
+ "ife"
+ ],
+ [
+ "▁li",
+ "fe"
+ ],
+ [
+ "▁lif",
+ "e"
+ ],
+ [
+ "▁",
+ "life"
+ ],
+ [
+ "as",
+ "on"
+ ],
+ [
+ "a",
+ "son"
+ ],
+ [
+ "▁u",
+ "sing"
+ ],
+ [
+ "▁us",
+ "ing"
+ ],
+ [
+ "▁",
+ "using"
+ ],
+ [
+ "▁v",
+ "al"
+ ],
+ [
+ "▁va",
+ "l"
+ ],
+ [
+ "▁",
+ "val"
+ ],
+ [
+ "▁d",
+ "u"
+ ],
+ [
+ "▁",
+ "du"
+ ],
+ [
+ "y",
+ "p"
+ ],
+ [
+ "▁O",
+ "n"
+ ],
+ [
+ "▁",
+ "On"
+ ],
+ [
+ "▁f",
+ "ound"
+ ],
+ [
+ "▁fo",
+ "und"
+ ],
+ [
+ "▁fou",
+ "nd"
+ ],
+ [
+ "▁",
+ "found"
+ ],
+ [
+ "ol",
+ "ut"
+ ],
+ [
+ "olu",
+ "t"
+ ],
+ [
+ "'",
+ "]"
+ ],
+ [
+ "ar",
+ "ent"
+ ],
+ [
+ "are",
+ "nt"
+ ],
+ [
+ "aren",
+ "t"
+ ],
+ [
+ "a",
+ "rent"
+ ],
+ [
+ "▁s",
+ "tring"
+ ],
+ [
+ "▁st",
+ "ring"
+ ],
+ [
+ "▁str",
+ "ing"
+ ],
+ [
+ "▁stri",
+ "ng"
+ ],
+ [
+ "▁",
+ "string"
+ ],
+ [
+ "▁m",
+ "et"
+ ],
+ [
+ "▁me",
+ "t"
+ ],
+ [
+ "▁",
+ "met"
+ ],
+ [
+ "▁w",
+ "r"
+ ],
+ [
+ "▁",
+ "wr"
+ ],
+ [
+ "us",
+ "h"
+ ],
+ [
+ "u",
+ "sh"
+ ],
+ [
+ "st",
+ "ring"
+ ],
+ [
+ "str",
+ "ing"
+ ],
+ [
+ "stri",
+ "ng"
+ ],
+ [
+ "s",
+ "tring"
+ ],
+ [
+ "si",
+ "ze"
+ ],
+ [
+ "s",
+ "ize"
+ ],
+ [
+ "▁v",
+ "er"
+ ],
+ [
+ "▁ve",
+ "r"
+ ],
+ [
+ "▁",
+ "ver"
+ ],
+ [
+ "▁e",
+ "ach"
+ ],
+ [
+ "▁",
+ "each"
+ ],
+ [
+ "val",
+ "ue"
+ ],
+ [
+ "v",
+ "alue"
+ ],
+ [
+ "▁l",
+ "ast"
+ ],
+ [
+ "▁la",
+ "st"
+ ],
+ [
+ "▁las",
+ "t"
+ ],
+ [
+ "▁",
+ "last"
+ ],
+ [
+ "▁g",
+ "ot"
+ ],
+ [
+ "▁go",
+ "t"
+ ],
+ [
+ "▁",
+ "got"
+ ],
+ [
+ "ve",
+ "n"
+ ],
+ [
+ "v",
+ "en"
+ ],
+ [
+ "ba",
+ "ck"
+ ],
+ [
+ "b",
+ "ack"
+ ],
+ [
+ "Se",
+ "t"
+ ],
+ [
+ "S",
+ "et"
+ ],
+ [
+ "e",
+ "y"
+ ],
+ [
+ "ro",
+ "l"
+ ],
+ [
+ "r",
+ "ol"
+ ],
+ [
+ "▁c",
+ "r"
+ ],
+ [
+ "▁",
+ "cr"
+ ],
+ [
+ "th",
+ "ing"
+ ],
+ [
+ "t",
+ "hing"
+ ],
+ [
+ "re",
+ "t"
+ ],
+ [
+ "r",
+ "et"
+ ],
+ [
+ "é",
+ "s"
+ ],
+ [
+ "is",
+ "m"
+ ],
+ [
+ "i",
+ "sm"
+ ],
+ [
+ "▁bet",
+ "ween"
+ ],
+ [
+ "▁",
+ "between"
+ ],
+ [
+ "O",
+ "b"
+ ],
+ [
+ "et",
+ "hing"
+ ],
+ [
+ "eth",
+ "ing"
+ ],
+ [
+ "e",
+ "thing"
+ ],
+ [
+ "m",
+ "p"
+ ],
+ [
+ "▁l",
+ "o"
+ ],
+ [
+ "▁",
+ "lo"
+ ],
+ [
+ "at",
+ "s"
+ ],
+ [
+ "a",
+ "ts"
+ ],
+ [
+ "▁N",
+ "ew"
+ ],
+ [
+ "▁Ne",
+ "w"
+ ],
+ [
+ "▁",
+ "New"
+ ],
+ [
+ "в",
+ "и"
+ ],
+ [
+ "ad",
+ "o"
+ ],
+ [
+ "a",
+ "do"
+ ],
+ [
+ "de",
+ "x"
+ ],
+ [
+ "d",
+ "ex"
+ ],
+ [
+ "д",
+ "и"
+ ],
+ [
+ "▁p",
+ "ass"
+ ],
+ [
+ "▁pas",
+ "s"
+ ],
+ [
+ "▁pa",
+ "ss"
+ ],
+ [
+ "▁",
+ "pass"
+ ],
+ [
+ "w",
+ "h"
+ ],
+ [
+ "▁d",
+ "en"
+ ],
+ [
+ "▁de",
+ "n"
+ ],
+ [
+ "▁",
+ "den"
+ ],
+ [
+ "Ge",
+ "t"
+ ],
+ [
+ "G",
+ "et"
+ ],
+ [
+ "ap",
+ "t"
+ ],
+ [
+ "a",
+ "pt"
+ ],
+ [
+ "▁a",
+ "sk"
+ ],
+ [
+ "▁as",
+ "k"
+ ],
+ [
+ "▁",
+ "ask"
+ ],
+ [
+ "▁s",
+ "up"
+ ],
+ [
+ "▁su",
+ "p"
+ ],
+ [
+ "▁",
+ "sup"
+ ],
+ [
+ "Val",
+ "ue"
+ ],
+ [
+ "V",
+ "alue"
+ ],
+ [
+ "н",
+ "ы"
+ ],
+ [
+ "▁t",
+ "ry"
+ ],
+ [
+ "▁tr",
+ "y"
+ ],
+ [
+ "▁",
+ "try"
+ ],
+ [
+ "lat",
+ "ion"
+ ],
+ [
+ "l",
+ "ation"
+ ],
+ [
+ "da",
+ "y"
+ ],
+ [
+ "d",
+ "ay"
+ ],
+ [
+ "ne",
+ "ss"
+ ],
+ [
+ "nes",
+ "s"
+ ],
+ [
+ "n",
+ "ess"
+ ],
+ [
+ "et",
+ "s"
+ ],
+ [
+ "e",
+ "ts"
+ ],
+ [
+ "▁ex",
+ "per"
+ ],
+ [
+ "▁exp",
+ "er"
+ ],
+ [
+ "▁",
+ "exper"
+ ],
+ [
+ "T",
+ "r"
+ ],
+ [
+ "▁M",
+ "ar"
+ ],
+ [
+ "▁Ma",
+ "r"
+ ],
+ [
+ "▁",
+ "Mar"
+ ],
+ [
+ "se",
+ "rv"
+ ],
+ [
+ "ser",
+ "v"
+ ],
+ [
+ "s",
+ "erv"
+ ],
+ [
+ "b",
+ "r"
+ ],
+ [
+ "▁n",
+ "umber"
+ ],
+ [
+ "▁num",
+ "ber"
+ ],
+ [
+ "▁nu",
+ "mber"
+ ],
+ [
+ "▁",
+ "number"
+ ],
+ [
+ "in",
+ "al"
+ ],
+ [
+ "ina",
+ "l"
+ ],
+ [
+ "i",
+ "nal"
+ ],
+ [
+ "ce",
+ "nt"
+ ],
+ [
+ "cen",
+ "t"
+ ],
+ [
+ "c",
+ "ent"
+ ],
+ [
+ "/",
+ "*"
+ ],
+ [
+ "no",
+ "t"
+ ],
+ [
+ "n",
+ "ot"
+ ],
+ [
+ "ion",
+ "al"
+ ],
+ [
+ "io",
+ "nal"
+ ],
+ [
+ "iona",
+ "l"
+ ],
+ [
+ "i",
+ "onal"
+ ],
+ [
+ "▁f",
+ "inal"
+ ],
+ [
+ "▁fin",
+ "al"
+ ],
+ [
+ "▁fi",
+ "nal"
+ ],
+ [
+ "▁",
+ "final"
+ ],
+ [
+ "'",
+ ")"
+ ],
+ [
+ "▁r",
+ "un"
+ ],
+ [
+ "▁ru",
+ "n"
+ ],
+ [
+ "▁",
+ "run"
+ ],
+ [
+ "ov",
+ "er"
+ ],
+ [
+ "ove",
+ "r"
+ ],
+ [
+ "o",
+ "ver"
+ ],
+ [
+ "▁n",
+ "ever"
+ ],
+ [
+ "▁ne",
+ "ver"
+ ],
+ [
+ "▁",
+ "never"
+ ],
+ [
+ "u",
+ "c"
+ ],
+ [
+ "▁h",
+ "igh"
+ ],
+ [
+ "▁hig",
+ "h"
+ ],
+ [
+ "▁hi",
+ "gh"
+ ],
+ [
+ "▁",
+ "high"
+ ],
+ [
+ "yl",
+ "e"
+ ],
+ [
+ "y",
+ "le"
+ ],
+ [
+ "▁in",
+ "s"
+ ],
+ [
+ "▁i",
+ "ns"
+ ],
+ [
+ "▁",
+ "ins"
+ ],
+ [
+ "▁b",
+ "est"
+ ],
+ [
+ "▁be",
+ "st"
+ ],
+ [
+ "▁bes",
+ "t"
+ ],
+ [
+ "▁",
+ "best"
+ ],
+ [
+ "it",
+ "tle"
+ ],
+ [
+ "itt",
+ "le"
+ ],
+ [
+ "ri",
+ "c"
+ ],
+ [
+ "r",
+ "ic"
+ ],
+ [
+ "▁s",
+ "ign"
+ ],
+ [
+ "▁si",
+ "gn"
+ ],
+ [
+ "▁sig",
+ "n"
+ ],
+ [
+ "▁",
+ "sign"
+ ],
+ [
+ "▁d",
+ "em"
+ ],
+ [
+ "▁de",
+ "m"
+ ],
+ [
+ "▁",
+ "dem"
+ ],
+ [
+ "in",
+ "ess"
+ ],
+ [
+ "ine",
+ "ss"
+ ],
+ [
+ "ines",
+ "s"
+ ],
+ [
+ "i",
+ "ness"
+ ],
+ [
+ "g",
+ "y"
+ ],
+ [
+ "▁w",
+ "ar"
+ ],
+ [
+ "▁wa",
+ "r"
+ ],
+ [
+ "▁",
+ "war"
+ ],
+ [
+ "is",
+ "hed"
+ ],
+ [
+ "ish",
+ "ed"
+ ],
+ [
+ "▁g",
+ "iv"
+ ],
+ [
+ "▁gi",
+ "v"
+ ],
+ [
+ "ke",
+ "y"
+ ],
+ [
+ "k",
+ "ey"
+ ],
+ [
+ "▁",
+ "X"
+ ],
+ [
+ "(",
+ "$"
+ ],
+ [
+ "▁ch",
+ "ild"
+ ],
+ [
+ "▁chi",
+ "ld"
+ ],
+ [
+ "▁",
+ "child"
+ ],
+ [
+ "le",
+ "ss"
+ ],
+ [
+ "les",
+ "s"
+ ],
+ [
+ "l",
+ "ess"
+ ],
+ [
+ "way",
+ "s"
+ ],
+ [
+ "wa",
+ "ys"
+ ],
+ [
+ "w",
+ "ays"
+ ],
+ [
+ "in",
+ "cl"
+ ],
+ [
+ "inc",
+ "l"
+ ],
+ [
+ "ro",
+ "p"
+ ],
+ [
+ "r",
+ "op"
+ ],
+ [
+ "ra",
+ "w"
+ ],
+ [
+ "r",
+ "aw"
+ ],
+ [
+ ":",
+ "//"
+ ],
+ [
+ "▁",
+ "«"
+ ],
+ [
+ "n",
+ "o"
+ ],
+ [
+ "ind",
+ "ow"
+ ],
+ [
+ "indo",
+ "w"
+ ],
+ [
+ "f",
+ "e"
+ ],
+ [
+ "ri",
+ "end"
+ ],
+ [
+ "rie",
+ "nd"
+ ],
+ [
+ "rien",
+ "d"
+ ],
+ [
+ "▁l",
+ "es"
+ ],
+ [
+ "▁le",
+ "s"
+ ],
+ [
+ "▁",
+ "les"
+ ],
+ [
+ "▁l",
+ "os"
+ ],
+ [
+ "▁lo",
+ "s"
+ ],
+ [
+ "▁",
+ "los"
+ ],
+ [
+ "fil",
+ "e"
+ ],
+ [
+ "fi",
+ "le"
+ ],
+ [
+ "f",
+ "ile"
+ ],
+ [
+ "form",
+ "ation"
+ ],
+ [
+ "format",
+ "ion"
+ ],
+ [
+ "cc",
+ "ess"
+ ],
+ [
+ "c",
+ "cess"
+ ],
+ [
+ "▁",
+ "В"
+ ],
+ [
+ "n",
+ "a"
+ ],
+ [
+ "▁i",
+ "l"
+ ],
+ [
+ "▁",
+ "il"
+ ],
+ [
+ "is",
+ "ion"
+ ],
+ [
+ "isi",
+ "on"
+ ],
+ [
+ "le",
+ "r"
+ ],
+ [
+ "l",
+ "er"
+ ],
+ [
+ "▁a",
+ "rt"
+ ],
+ [
+ "▁ar",
+ "t"
+ ],
+ [
+ "▁",
+ "art"
+ ],
+ [
+ "Con",
+ "t"
+ ],
+ [
+ "Co",
+ "nt"
+ ],
+ [
+ "C",
+ "ont"
+ ],
+ [
+ "▁w",
+ "orld"
+ ],
+ [
+ "▁wor",
+ "ld"
+ ],
+ [
+ "▁",
+ "world"
+ ],
+ [
+ "▁t",
+ "urn"
+ ],
+ [
+ "▁tu",
+ "rn"
+ ],
+ [
+ "▁tur",
+ "n"
+ ],
+ [
+ "▁",
+ "turn"
+ ],
+ [
+ "▁re",
+ "ally"
+ ],
+ [
+ "▁real",
+ "ly"
+ ],
+ [
+ "▁E",
+ "x"
+ ],
+ [
+ "▁",
+ "Ex"
+ ],
+ [
+ "м",
+ "а"
+ ],
+ [
+ "▁",
+ "П"
+ ],
+ [
+ "ter",
+ "s"
+ ],
+ [
+ "te",
+ "rs"
+ ],
+ [
+ "t",
+ "ers"
+ ],
+ [
+ "ar",
+ "get"
+ ],
+ [
+ "arg",
+ "et"
+ ],
+ [
+ "arge",
+ "t"
+ ],
+ [
+ "Er",
+ "r"
+ ],
+ [
+ "E",
+ "rr"
+ ],
+ [
+ "▁h",
+ "app"
+ ],
+ [
+ "▁ha",
+ "pp"
+ ],
+ [
+ "ti",
+ "me"
+ ],
+ [
+ "tim",
+ "e"
+ ],
+ [
+ "t",
+ "ime"
+ ],
+ [
+ "▁S",
+ "o"
+ ],
+ [
+ "▁",
+ "So"
+ ],
+ [
+ "di",
+ "v"
+ ],
+ [
+ "d",
+ "iv"
+ ],
+ [
+ "▁did",
+ "n"
+ ],
+ [
+ "▁di",
+ "dn"
+ ],
+ [
+ "ad",
+ "a"
+ ],
+ [
+ "a",
+ "da"
+ ],
+ [
+ "oo",
+ "t"
+ ],
+ [
+ "o",
+ "ot"
+ ],
+ [
+ "}",
+ ")"
+ ],
+ [
+ "▁s",
+ "ch"
+ ],
+ [
+ "▁sc",
+ "h"
+ ],
+ [
+ "▁",
+ "sch"
+ ],
+ [
+ "▁c",
+ "le"
+ ],
+ [
+ "▁cl",
+ "e"
+ ],
+ [
+ "▁",
+ "cle"
+ ],
+ [
+ "▁some",
+ "thing"
+ ],
+ [
+ "▁som",
+ "ething"
+ ],
+ [
+ "▁somet",
+ "hing"
+ ],
+ [
+ "▁",
+ "something"
+ ],
+ [
+ "()",
+ "."
+ ],
+ [
+ "(",
+ ")."
+ ],
+ [
+ "▁c",
+ "our"
+ ],
+ [
+ "▁co",
+ "ur"
+ ],
+ [
+ "▁cou",
+ "r"
+ ],
+ [
+ "ev",
+ "er"
+ ],
+ [
+ "eve",
+ "r"
+ ],
+ [
+ "e",
+ "ver"
+ ],
+ [
+ "an",
+ "ts"
+ ],
+ [
+ "ant",
+ "s"
+ ],
+ [
+ "▁",
+ "?"
+ ],
+ [
+ "T",
+ "o"
+ ],
+ [
+ "▁",
+ "`"
+ ],
+ [
+ "tr",
+ "y"
+ ],
+ [
+ "t",
+ "ry"
+ ],
+ [
+ "u",
+ "x"
+ ],
+ [
+ "ai",
+ "s"
+ ],
+ [
+ "a",
+ "is"
+ ],
+ [
+ "ro",
+ "ss"
+ ],
+ [
+ "ros",
+ "s"
+ ],
+ [
+ "r",
+ "oss"
+ ],
+ [
+ "hi",
+ "p"
+ ],
+ [
+ "h",
+ "ip"
+ ],
+ [
+ "▁re",
+ "p"
+ ],
+ [
+ "▁r",
+ "ep"
+ ],
+ [
+ "▁",
+ "rep"
+ ],
+ [
+ "la",
+ "bel"
+ ],
+ [
+ "lab",
+ "el"
+ ],
+ [
+ "l",
+ "abel"
+ ],
+ [
+ "▁b",
+ "oth"
+ ],
+ [
+ "▁bo",
+ "th"
+ ],
+ [
+ "▁bot",
+ "h"
+ ],
+ [
+ "▁",
+ "both"
+ ],
+ [
+ "*",
+ ","
+ ],
+ [
+ "ot",
+ "t"
+ ],
+ [
+ "o",
+ "tt"
+ ],
+ [
+ "м",
+ "и"
+ ],
+ [
+ "an",
+ "e"
+ ],
+ [
+ "a",
+ "ne"
+ ],
+ [
+ "▁o",
+ "pen"
+ ],
+ [
+ "▁op",
+ "en"
+ ],
+ [
+ "▁",
+ "open"
+ ],
+ [
+ "w",
+ "w"
+ ],
+ [
+ "▁c",
+ "ome"
+ ],
+ [
+ "▁com",
+ "e"
+ ],
+ [
+ "▁co",
+ "me"
+ ],
+ [
+ "▁",
+ "come"
+ ],
+ [
+ "▁e",
+ "xt"
+ ],
+ [
+ "▁ex",
+ "t"
+ ],
+ [
+ "▁",
+ "ext"
+ ],
+ [
+ "re",
+ "m"
+ ],
+ [
+ "r",
+ "em"
+ ],
+ [
+ "_{",
+ "\\"
+ ],
+ [
+ "_",
+ "{\\"
+ ],
+ [
+ "▁o",
+ "ld"
+ ],
+ [
+ "▁ol",
+ "d"
+ ],
+ [
+ "▁",
+ "old"
+ ],
+ [
+ "ch",
+ "ed"
+ ],
+ [
+ "che",
+ "d"
+ ],
+ [
+ "c",
+ "hed"
+ ],
+ [
+ ".",
+ "_"
+ ],
+ [
+ "M",
+ "E"
+ ],
+ [
+ "if",
+ "y"
+ ],
+ [
+ "i",
+ "fy"
+ ],
+ [
+ "g",
+ "g"
+ ],
+ [
+ "Co",
+ "l"
+ ],
+ [
+ "C",
+ "ol"
+ ],
+ [
+ "vi",
+ "ew"
+ ],
+ [
+ "v",
+ "iew"
+ ],
+ [
+ "▁b",
+ "us"
+ ],
+ [
+ "▁bu",
+ "s"
+ ],
+ [
+ "▁",
+ "bus"
+ ],
+ [
+ "▁m",
+ "ust"
+ ],
+ [
+ "▁mus",
+ "t"
+ ],
+ [
+ "▁mu",
+ "st"
+ ],
+ [
+ "▁",
+ "must"
+ ],
+ [
+ "▁d",
+ "ifferent"
+ ],
+ [
+ "▁differ",
+ "ent"
+ ],
+ [
+ "lo",
+ "g"
+ ],
+ [
+ "l",
+ "og"
+ ],
+ [
+ "is",
+ "ts"
+ ],
+ [
+ "ist",
+ "s"
+ ],
+ [
+ "i",
+ "sts"
+ ],
+ [
+ "ro",
+ "ll"
+ ],
+ [
+ "rol",
+ "l"
+ ],
+ [
+ "r",
+ "oll"
+ ],
+ [
+ "a",
+ "i"
+ ],
+ [
+ "▁з",
+ "а"
+ ],
+ [
+ "▁",
+ "за"
+ ],
+ [
+ "▁s",
+ "ystem"
+ ],
+ [
+ "▁sys",
+ "tem"
+ ],
+ [
+ "▁syst",
+ "em"
+ ],
+ [
+ "▁",
+ "system"
+ ],
+ [
+ "iv",
+ "ers"
+ ],
+ [
+ "ive",
+ "rs"
+ ],
+ [
+ "iver",
+ "s"
+ ],
+ [
+ "i",
+ "vers"
+ ],
+ [
+ "at",
+ "us"
+ ],
+ [
+ "atu",
+ "s"
+ ],
+ [
+ "ot",
+ "e"
+ ],
+ [
+ "o",
+ "te"
+ ],
+ [
+ "me",
+ "d"
+ ],
+ [
+ "m",
+ "ed"
+ ],
+ [
+ "]",
+ "."
+ ],
+ [
+ "ak",
+ "es"
+ ],
+ [
+ "ake",
+ "s"
+ ],
+ [
+ "a",
+ "kes"
+ ],
+ [
+ "R",
+ "O"
+ ],
+ [
+ "▁c",
+ "ent"
+ ],
+ [
+ "▁ce",
+ "nt"
+ ],
+ [
+ "▁",
+ "cent"
+ ],
+ [
+ "gr",
+ "am"
+ ],
+ [
+ "gra",
+ "m"
+ ],
+ [
+ "g",
+ "ram"
+ ],
+ [
+ "▁p",
+ "rivate"
+ ],
+ [
+ "▁priv",
+ "ate"
+ ],
+ [
+ "▁",
+ "private"
+ ],
+ [
+ "▁g",
+ "reat"
+ ],
+ [
+ "▁gre",
+ "at"
+ ],
+ [
+ "\"",
+ ";"
+ ],
+ [
+ "op",
+ "y"
+ ],
+ [
+ "o",
+ "py"
+ ],
+ [
+ "▁fe",
+ "el"
+ ],
+ [
+ "▁fee",
+ "l"
+ ],
+ [
+ "▁H",
+ "ow"
+ ],
+ [
+ "▁Ho",
+ "w"
+ ],
+ [
+ "▁",
+ "How"
+ ],
+ [
+ "//",
+ "//"
+ ],
+ [
+ "///",
+ "/"
+ ],
+ [
+ "/",
+ "///"
+ ],
+ [
+ "I",
+ "C"
+ ],
+ [
+ "▁d",
+ "r"
+ ],
+ [
+ "▁",
+ "dr"
+ ],
+ [
+ "ain",
+ "s"
+ ],
+ [
+ "ai",
+ "ns"
+ ],
+ [
+ "a",
+ "ins"
+ ],
+ [
+ "lo",
+ "ck"
+ ],
+ [
+ "loc",
+ "k"
+ ],
+ [
+ "l",
+ "ock"
+ ],
+ [
+ "E",
+ "n"
+ ],
+ [
+ "▁S",
+ "ch"
+ ],
+ [
+ "▁Sc",
+ "h"
+ ],
+ [
+ "▁",
+ "Sch"
+ ],
+ [
+ "▁m",
+ "at"
+ ],
+ [
+ "▁ma",
+ "t"
+ ],
+ [
+ "▁",
+ "mat"
+ ],
+ [
+ "▁h",
+ "ome"
+ ],
+ [
+ "▁hom",
+ "e"
+ ],
+ [
+ "▁ho",
+ "me"
+ ],
+ [
+ "▁",
+ "home"
+ ],
+ [
+ "per",
+ "ty"
+ ],
+ [
+ "pert",
+ "y"
+ ],
+ [
+ "te",
+ "st"
+ ],
+ [
+ "tes",
+ "t"
+ ],
+ [
+ "t",
+ "est"
+ ],
+ [
+ "lo",
+ "c"
+ ],
+ [
+ "l",
+ "oc"
+ ],
+ [
+ "▁w",
+ "om"
+ ],
+ [
+ "▁wo",
+ "m"
+ ],
+ [
+ "s",
+ "w"
+ ],
+ [
+ "ar",
+ "ly"
+ ],
+ [
+ "arl",
+ "y"
+ ],
+ [
+ "▁E",
+ "n"
+ ],
+ [
+ "▁",
+ "En"
+ ],
+ [
+ "▁к",
+ "о"
+ ],
+ [
+ "▁",
+ "ко"
+ ],
+ [
+ "de",
+ "n"
+ ],
+ [
+ "d",
+ "en"
+ ],
+ [
+ "ст",
+ "а"
+ ],
+ [
+ "с",
+ "та"
+ ],
+ [
+ "▁",
+ "а"
+ ],
+ [
+ "et",
+ "er"
+ ],
+ [
+ "ete",
+ "r"
+ ],
+ [
+ "e",
+ "ter"
+ ],
+ [
+ "▁incl",
+ "ud"
+ ],
+ [
+ "▁inclu",
+ "d"
+ ],
+ [
+ "UL",
+ "L"
+ ],
+ [
+ "U",
+ "LL"
+ ],
+ [
+ "▁m",
+ "em"
+ ],
+ [
+ "▁me",
+ "m"
+ ],
+ [
+ "▁",
+ "mem"
+ ],
+ [
+ "▁p",
+ "o"
+ ],
+ [
+ "▁",
+ "po"
+ ],
+ [
+ "▁l",
+ "ittle"
+ ],
+ [
+ "▁lit",
+ "tle"
+ ],
+ [
+ "▁litt",
+ "le"
+ ],
+ [
+ "▁a",
+ "rg"
+ ],
+ [
+ "▁ar",
+ "g"
+ ],
+ [
+ "▁",
+ "arg"
+ ],
+ [
+ "▁}",
+ ","
+ ],
+ [
+ "▁",
+ "},"
+ ],
+ [
+ "in",
+ "clude"
+ ],
+ [
+ "incl",
+ "ude"
+ ],
+ [
+ "et",
+ "a"
+ ],
+ [
+ "e",
+ "ta"
+ ],
+ [
+ "▁p",
+ "lace"
+ ],
+ [
+ "▁pl",
+ "ace"
+ ],
+ [
+ "▁plac",
+ "e"
+ ],
+ [
+ "▁",
+ "place"
+ ],
+ [
+ "id",
+ "th"
+ ],
+ [
+ "us",
+ "tom"
+ ],
+ [
+ "ust",
+ "om"
+ ],
+ [
+ "▁|",
+ "|"
+ ],
+ [
+ "▁",
+ "||"
+ ],
+ [
+ "▁t",
+ "em"
+ ],
+ [
+ "▁te",
+ "m"
+ ],
+ [
+ "▁",
+ "tem"
+ ],
+ [
+ "ri",
+ "ed"
+ ],
+ [
+ "rie",
+ "d"
+ ],
+ [
+ "r",
+ "ied"
+ ],
+ [
+ "▁f",
+ "act"
+ ],
+ [
+ "▁fac",
+ "t"
+ ],
+ [
+ "▁fa",
+ "ct"
+ ],
+ [
+ "▁",
+ "fact"
+ ],
+ [
+ "ien",
+ "ce"
+ ],
+ [
+ "i",
+ "ence"
+ ],
+ [
+ "▁P",
+ "l"
+ ],
+ [
+ "▁",
+ "Pl"
+ ],
+ [
+ "op",
+ "t"
+ ],
+ [
+ "o",
+ "pt"
+ ],
+ [
+ "el",
+ "e"
+ ],
+ [
+ "e",
+ "le"
+ ],
+ [
+ "g",
+ "o"
+ ],
+ [
+ "A",
+ "C"
+ ],
+ [
+ "in",
+ "ter"
+ ],
+ [
+ "int",
+ "er"
+ ],
+ [
+ "inte",
+ "r"
+ ],
+ [
+ "====",
+ "===="
+ ],
+ [
+ "()",
+ ","
+ ],
+ [
+ "(",
+ "),"
+ ],
+ [
+ "ot",
+ "s"
+ ],
+ [
+ "o",
+ "ts"
+ ],
+ [
+ "ra",
+ "l"
+ ],
+ [
+ "r",
+ "al"
+ ],
+ [
+ "iqu",
+ "e"
+ ],
+ [
+ "iq",
+ "ue"
+ ],
+ [
+ "i",
+ "que"
+ ],
+ [
+ "av",
+ "ing"
+ ],
+ [
+ "avi",
+ "ng"
+ ],
+ [
+ "a",
+ "ving"
+ ],
+ [
+ "m",
+ "l"
+ ],
+ [
+ "▁th",
+ "ought"
+ ],
+ [
+ "▁though",
+ "t"
+ ],
+ [
+ "▁thou",
+ "ght"
+ ],
+ [
+ "fr",
+ "ac"
+ ],
+ [
+ "f",
+ "rac"
+ ],
+ [
+ "▁c",
+ "are"
+ ],
+ [
+ "▁car",
+ "e"
+ ],
+ [
+ "▁ca",
+ "re"
+ ],
+ [
+ "▁",
+ "care"
+ ],
+ [
+ "()",
+ ");"
+ ],
+ [
+ "())",
+ ";"
+ ],
+ [
+ "(",
+ "));"
+ ],
+ [
+ "▁p",
+ "ut"
+ ],
+ [
+ "▁pu",
+ "t"
+ ],
+ [
+ "▁",
+ "put"
+ ],
+ [
+ "▁m",
+ "ight"
+ ],
+ [
+ "▁mi",
+ "ght"
+ ],
+ [
+ "▁mig",
+ "ht"
+ ],
+ [
+ "▁A",
+ "mer"
+ ],
+ [
+ "▁Am",
+ "er"
+ ],
+ [
+ "▁",
+ "Amer"
+ ],
+ [
+ "▁(",
+ "!"
+ ],
+ [
+ "▁",
+ "(!"
+ ],
+ [
+ "am",
+ "ple"
+ ],
+ [
+ "amp",
+ "le"
+ ],
+ [
+ "al",
+ "th"
+ ],
+ [
+ "alt",
+ "h"
+ ],
+ [
+ "▁f",
+ "ew"
+ ],
+ [
+ "▁fe",
+ "w"
+ ],
+ [
+ "▁st",
+ "ate"
+ ],
+ [
+ "▁stat",
+ "e"
+ ],
+ [
+ "▁sta",
+ "te"
+ ],
+ [
+ "▁",
+ "state"
+ ],
+ [
+ "su",
+ "b"
+ ],
+ [
+ "s",
+ "ub"
+ ],
+ [
+ "▁O",
+ "r"
+ ],
+ [
+ "▁",
+ "Or"
+ ],
+ [
+ "]",
+ ";"
+ ],
+ [
+ "▁s",
+ "ize"
+ ],
+ [
+ "▁si",
+ "ze"
+ ],
+ [
+ "▁",
+ "size"
+ ],
+ [
+ "▁S",
+ "p"
+ ],
+ [
+ "▁",
+ "Sp"
+ ],
+ [
+ "▁with",
+ "out"
+ ],
+ [
+ "▁",
+ "without"
+ ],
+ [
+ "▁p",
+ "oss"
+ ],
+ [
+ "▁pos",
+ "s"
+ ],
+ [
+ "▁po",
+ "ss"
+ ],
+ [
+ "▁",
+ "poss"
+ ],
+ [
+ "e",
+ "q"
+ ],
+ [
+ "pl",
+ "ay"
+ ],
+ [
+ "p",
+ "lay"
+ ],
+ [
+ "▁ex",
+ "pect"
+ ],
+ [
+ "▁exp",
+ "ect"
+ ],
+ [
+ "▁",
+ "expect"
+ ],
+ [
+ "▁se",
+ "cond"
+ ],
+ [
+ "▁sec",
+ "ond"
+ ],
+ [
+ "▁",
+ "second"
+ ],
+ [
+ "▁S",
+ "tring"
+ ],
+ [
+ "▁St",
+ "ring"
+ ],
+ [
+ "▁Str",
+ "ing"
+ ],
+ [
+ "▁",
+ "String"
+ ],
+ [
+ "ui",
+ "ld"
+ ],
+ [
+ "u",
+ "ild"
+ ],
+ [
+ "▁n",
+ "ext"
+ ],
+ [
+ "▁ne",
+ "xt"
+ ],
+ [
+ "▁",
+ "next"
+ ],
+ [
+ "+",
+ "+"
+ ],
+ [
+ "re",
+ "qu"
+ ],
+ [
+ "req",
+ "u"
+ ],
+ [
+ "r",
+ "equ"
+ ],
+ [
+ "▁A",
+ "ll"
+ ],
+ [
+ "▁Al",
+ "l"
+ ],
+ [
+ "▁",
+ "All"
+ ],
+ [
+ "▁m",
+ "en"
+ ],
+ [
+ "▁me",
+ "n"
+ ],
+ [
+ "▁",
+ "men"
+ ],
+ [
+ "▁W",
+ "hen"
+ ],
+ [
+ "▁Wh",
+ "en"
+ ],
+ [
+ "▁Whe",
+ "n"
+ ],
+ [
+ "▁",
+ "When"
+ ],
+ [
+ "it",
+ "er"
+ ],
+ [
+ "ite",
+ "r"
+ ],
+ [
+ "i",
+ "ter"
+ ],
+ [
+ "am",
+ "ent"
+ ],
+ [
+ "ame",
+ "nt"
+ ],
+ [
+ "amen",
+ "t"
+ ],
+ [
+ "a",
+ "ment"
+ ],
+ [
+ "ne",
+ "t"
+ ],
+ [
+ "n",
+ "et"
+ ],
+ [
+ "▁",
+ "К"
+ ],
+ [
+ "ro",
+ "n"
+ ],
+ [
+ "r",
+ "on"
+ ],
+ [
+ "ain",
+ "t"
+ ],
+ [
+ "ai",
+ "nt"
+ ],
+ [
+ "a",
+ "int"
+ ],
+ [
+ "▁I",
+ "s"
+ ],
+ [
+ "▁",
+ "Is"
+ ],
+ [
+ "в",
+ "е"
+ ],
+ [
+ "pe",
+ "nd"
+ ],
+ [
+ "pen",
+ "d"
+ ],
+ [
+ "p",
+ "end"
+ ],
+ [
+ "trans",
+ "lation"
+ ],
+ [
+ "transl",
+ "ation"
+ ],
+ [
+ "▁г",
+ "о"
+ ],
+ [
+ "▁",
+ "го"
+ ],
+ [
+ "ч",
+ "е"
+ ],
+ [
+ "▁v",
+ "an"
+ ],
+ [
+ "▁va",
+ "n"
+ ],
+ [
+ "▁",
+ "van"
+ ],
+ [
+ "▁an",
+ "other"
+ ],
+ [
+ "▁ano",
+ "ther"
+ ],
+ [
+ "▁re",
+ "t"
+ ],
+ [
+ "▁r",
+ "et"
+ ],
+ [
+ "▁",
+ "ret"
+ ],
+ [
+ "▁L",
+ "a"
+ ],
+ [
+ "▁",
+ "La"
+ ],
+ [
+ "Mo",
+ "d"
+ ],
+ [
+ "M",
+ "od"
+ ],
+ [
+ "IO",
+ "N"
+ ],
+ [
+ "I",
+ "ON"
+ ],
+ [
+ "li",
+ "st"
+ ],
+ [
+ "l",
+ "ist"
+ ],
+ [
+ "▁p",
+ "ost"
+ ],
+ [
+ "▁pos",
+ "t"
+ ],
+ [
+ "▁po",
+ "st"
+ ],
+ [
+ "▁",
+ "post"
+ ],
+ [
+ "d",
+ "a"
+ ],
+ [
+ "wa",
+ "re"
+ ],
+ [
+ "war",
+ "e"
+ ],
+ [
+ "w",
+ "are"
+ ],
+ [
+ "▁w",
+ "ord"
+ ],
+ [
+ "▁wor",
+ "d"
+ ],
+ [
+ "▁wo",
+ "rd"
+ ],
+ [
+ "▁",
+ "word"
+ ],
+ [
+ "Err",
+ "or"
+ ],
+ [
+ "Er",
+ "ror"
+ ],
+ [
+ "▁se",
+ "em"
+ ],
+ [
+ "▁see",
+ "m"
+ ],
+ [
+ "▁cont",
+ "in"
+ ],
+ [
+ "▁",
+ "contin"
+ ],
+ [
+ "at",
+ "ic"
+ ],
+ [
+ "ati",
+ "c"
+ ],
+ [
+ "▁th",
+ "ree"
+ ],
+ [
+ "▁thr",
+ "ee"
+ ],
+ [
+ "▁",
+ "three"
+ ],
+ [
+ "Ob",
+ "ject"
+ ],
+ [
+ "Obj",
+ "ect"
+ ],
+ [
+ "▁part",
+ "ic"
+ ],
+ [
+ "▁parti",
+ "c"
+ ],
+ [
+ "$",
+ "."
+ ],
+ [
+ "▁m",
+ "ark"
+ ],
+ [
+ "▁mar",
+ "k"
+ ],
+ [
+ "▁",
+ "mark"
+ ],
+ [
+ "▁v",
+ "is"
+ ],
+ [
+ "▁vi",
+ "s"
+ ],
+ [
+ "▁",
+ "vis"
+ ],
+ [
+ "r",
+ "c"
+ ],
+ [
+ "▁s",
+ "w"
+ ],
+ [
+ "▁",
+ "sw"
+ ],
+ [
+ "pt",
+ "ions"
+ ],
+ [
+ "ption",
+ "s"
+ ],
+ [
+ "▁b",
+ "reak"
+ ],
+ [
+ "▁bre",
+ "ak"
+ ],
+ [
+ "▁",
+ "break"
+ ],
+ [
+ "▁th",
+ "ings"
+ ],
+ [
+ "▁thing",
+ "s"
+ ],
+ [
+ "▁thin",
+ "gs"
+ ],
+ [
+ "ut",
+ "e"
+ ],
+ [
+ "u",
+ "te"
+ ],
+ [
+ "u",
+ "i"
+ ],
+ [
+ "▁T",
+ "hat"
+ ],
+ [
+ "▁Th",
+ "at"
+ ],
+ [
+ "▁",
+ "That"
+ ],
+ [
+ "ur",
+ "s"
+ ],
+ [
+ "u",
+ "rs"
+ ],
+ [
+ "g",
+ "l"
+ ],
+ [
+ "р",
+ "у"
+ ],
+ [
+ "▁f",
+ "ile"
+ ],
+ [
+ "▁fil",
+ "e"
+ ],
+ [
+ "▁fi",
+ "le"
+ ],
+ [
+ "▁",
+ "file"
+ ],
+ [
+ "us",
+ "e"
+ ],
+ [
+ "u",
+ "se"
+ ],
+ [
+ "ig",
+ "ned"
+ ],
+ [
+ "ign",
+ "ed"
+ ],
+ [
+ "igne",
+ "d"
+ ],
+ [
+ "par",
+ "t"
+ ],
+ [
+ "pa",
+ "rt"
+ ],
+ [
+ "p",
+ "art"
+ ],
+ [
+ "U",
+ "n"
+ ],
+ [
+ "▁e",
+ "qu"
+ ],
+ [
+ "▁eq",
+ "u"
+ ],
+ [
+ "▁",
+ "equ"
+ ],
+ [
+ "(",
+ "&"
+ ],
+ [
+ "▁l",
+ "ead"
+ ],
+ [
+ "▁le",
+ "ad"
+ ],
+ [
+ "r",
+ "m"
+ ],
+ [
+ "ain",
+ "ed"
+ ],
+ [
+ "ai",
+ "ned"
+ ],
+ [
+ "aine",
+ "d"
+ ],
+ [
+ "a",
+ "ined"
+ ],
+ [
+ "▁B",
+ "e"
+ ],
+ [
+ "▁",
+ "Be"
+ ],
+ [
+ "pat",
+ "h"
+ ],
+ [
+ "pa",
+ "th"
+ ],
+ [
+ "p",
+ "ath"
+ ],
+ [
+ "▁sm",
+ "all"
+ ],
+ [
+ "▁",
+ "small"
+ ],
+ [
+ "ag",
+ "er"
+ ],
+ [
+ "age",
+ "r"
+ ],
+ [
+ "a",
+ "ger"
+ ],
+ [
+ "▁al",
+ "ways"
+ ],
+ [
+ "▁",
+ "always"
+ ],
+ [
+ "▁E",
+ "l"
+ ],
+ [
+ "▁",
+ "El"
+ ],
+ [
+ "▁or",
+ "der"
+ ],
+ [
+ "▁ord",
+ "er"
+ ],
+ [
+ "▁",
+ "order"
+ ],
+ [
+ "▁e",
+ "y"
+ ],
+ [
+ "▁",
+ "ey"
+ ],
+ [
+ "▁w",
+ "on"
+ ],
+ [
+ "▁wo",
+ "n"
+ ],
+ [
+ "▁",
+ "won"
+ ],
+ [
+ "ap",
+ "e"
+ ],
+ [
+ "a",
+ "pe"
+ ],
+ [
+ "▁l",
+ "eft"
+ ],
+ [
+ "▁le",
+ "ft"
+ ],
+ [
+ "▁",
+ "left"
+ ],
+ [
+ "av",
+ "a"
+ ],
+ [
+ "a",
+ "va"
+ ],
+ [
+ "it",
+ "em"
+ ],
+ [
+ "ite",
+ "m"
+ ],
+ [
+ "i",
+ "tem"
+ ],
+ [
+ "ho",
+ "r"
+ ],
+ [
+ "h",
+ "or"
+ ],
+ [
+ "▁a",
+ "way"
+ ],
+ [
+ "▁aw",
+ "ay"
+ ],
+ [
+ "▁",
+ "away"
+ ],
+ [
+ "b",
+ "b"
+ ],
+ [
+ "fu",
+ "n"
+ ],
+ [
+ "f",
+ "un"
+ ],
+ [
+ "▁I",
+ "nd"
+ ],
+ [
+ "▁In",
+ "d"
+ ],
+ [
+ "▁",
+ "Ind"
+ ],
+ [
+ "m",
+ "b"
+ ],
+ [
+ "▁st",
+ "ruct"
+ ],
+ [
+ "▁str",
+ "uct"
+ ],
+ [
+ "▁stru",
+ "ct"
+ ],
+ [
+ "▁",
+ "struct"
+ ],
+ [
+ "▁pro",
+ "cess"
+ ],
+ [
+ "▁proc",
+ "ess"
+ ],
+ [
+ "▁proces",
+ "s"
+ ],
+ [
+ "▁",
+ "process"
+ ],
+ [
+ "▁s",
+ "upport"
+ ],
+ [
+ "▁sup",
+ "port"
+ ],
+ [
+ "▁supp",
+ "ort"
+ ],
+ [
+ "▁",
+ "support"
+ ],
+ [
+ ");",
+ "\r"
+ ],
+ [
+ ")",
+ ";\r"
+ ],
+ [
+ "ió",
+ "n"
+ ],
+ [
+ "i",
+ "ón"
+ ],
+ [
+ "L",
+ "O"
+ ],
+ [
+ "▁o",
+ "per"
+ ],
+ [
+ "▁op",
+ "er"
+ ],
+ [
+ "▁",
+ "oper"
+ ],
+ [
+ "U",
+ "T"
+ ],
+ [
+ "▁",
+ "·"
+ ],
+ [
+ "P",
+ "E"
+ ],
+ [
+ "lo",
+ "ad"
+ ],
+ [
+ "l",
+ "oad"
+ ],
+ [
+ "of",
+ "f"
+ ],
+ [
+ "o",
+ "ff"
+ ],
+ [
+ "▁N",
+ "o"
+ ],
+ [
+ "▁",
+ "No"
+ ],
+ [
+ "iv",
+ "es"
+ ],
+ [
+ "ive",
+ "s"
+ ],
+ [
+ "i",
+ "ves"
+ ],
+ [
+ "ic",
+ "an"
+ ],
+ [
+ "ica",
+ "n"
+ ],
+ [
+ "i",
+ "can"
+ ],
+ [
+ "▁v",
+ "e"
+ ],
+ [
+ "▁",
+ "ve"
+ ],
+ [
+ "act",
+ "ion"
+ ],
+ [
+ "a",
+ "ction"
+ ],
+ [
+ "'",
+ ";"
+ ],
+ [
+ "▁v",
+ "o"
+ ],
+ [
+ "▁",
+ "vo"
+ ],
+ [
+ "$",
+ ","
+ ],
+ [
+ "▁G",
+ "r"
+ ],
+ [
+ "▁",
+ "Gr"
+ ],
+ [
+ "pr",
+ "e"
+ ],
+ [
+ "p",
+ "re"
+ ],
+ [
+ "n",
+ "y"
+ ],
+ [
+ "ain",
+ "ing"
+ ],
+ [
+ "ai",
+ "ning"
+ ],
+ [
+ "a",
+ "ining"
+ ],
+ [
+ "io",
+ "r"
+ ],
+ [
+ "i",
+ "or"
+ ],
+ [
+ "in",
+ "it"
+ ],
+ [
+ "ini",
+ "t"
+ ],
+ [
+ "i",
+ "nit"
+ ],
+ [
+ "le",
+ "ction"
+ ],
+ [
+ "lect",
+ "ion"
+ ],
+ [
+ "l",
+ "ection"
+ ],
+ [
+ "ar",
+ "m"
+ ],
+ [
+ "a",
+ "rm"
+ ],
+ [
+ "um",
+ "n"
+ ],
+ [
+ "u",
+ "mn"
+ ],
+ [
+ "ag",
+ "s"
+ ],
+ [
+ "a",
+ "gs"
+ ],
+ [
+ "ц",
+ "и"
+ ],
+ [
+ "ск",
+ "о"
+ ],
+ [
+ "с",
+ "ко"
+ ],
+ [
+ "vers",
+ "ion"
+ ],
+ [
+ "v",
+ "ersion"
+ ],
+ [
+ "▁T",
+ "o"
+ ],
+ [
+ "▁",
+ "To"
+ ],
+ [
+ "▁re",
+ "f"
+ ],
+ [
+ "▁r",
+ "ef"
+ ],
+ [
+ "▁",
+ "ref"
+ ],
+ [
+ "st",
+ "and"
+ ],
+ [
+ "sta",
+ "nd"
+ ],
+ [
+ "stan",
+ "d"
+ ],
+ [
+ "▁A",
+ "t"
+ ],
+ [
+ "▁",
+ "At"
+ ],
+ [
+ "if",
+ "t"
+ ],
+ [
+ "i",
+ "ft"
+ ],
+ [
+ "▁e",
+ "in"
+ ],
+ [
+ "fa",
+ "ce"
+ ],
+ [
+ "fac",
+ "e"
+ ],
+ [
+ "f",
+ "ace"
+ ],
+ [
+ "b",
+ "o"
+ ],
+ [
+ "if",
+ "ied"
+ ],
+ [
+ "ifi",
+ "ed"
+ ],
+ [
+ "ve",
+ "d"
+ ],
+ [
+ "v",
+ "ed"
+ ],
+ [
+ "su",
+ "m"
+ ],
+ [
+ "s",
+ "um"
+ ],
+ [
+ "un",
+ "e"
+ ],
+ [
+ "u",
+ "ne"
+ ],
+ [
+ "it",
+ "al"
+ ],
+ [
+ "ita",
+ "l"
+ ],
+ [
+ "i",
+ "tal"
+ ],
+ [
+ "um",
+ "p"
+ ],
+ [
+ "u",
+ "mp"
+ ],
+ [
+ "com",
+ "m"
+ ],
+ [
+ "co",
+ "mm"
+ ],
+ [
+ "c",
+ "omm"
+ ],
+ [
+ "▁m",
+ "ov"
+ ],
+ [
+ "▁mo",
+ "v"
+ ],
+ [
+ "▁",
+ "mov"
+ ],
+ [
+ "el",
+ "t"
+ ],
+ [
+ "e",
+ "lt"
+ ],
+ [
+ "▁v",
+ "on"
+ ],
+ [
+ "▁vo",
+ "n"
+ ],
+ [
+ "vel",
+ "op"
+ ],
+ [
+ "ct",
+ "or"
+ ],
+ [
+ "c",
+ "tor"
+ ],
+ [
+ "he",
+ "ad"
+ ],
+ [
+ "h",
+ "ead"
+ ],
+ [
+ "cl",
+ "e"
+ ],
+ [
+ "c",
+ "le"
+ ],
+ [
+ "▁b",
+ "uild"
+ ],
+ [
+ "▁bu",
+ "ild"
+ ],
+ [
+ "▁",
+ "build"
+ ],
+ [
+ "in",
+ "c"
+ ],
+ [
+ "i",
+ "nc"
+ ],
+ [
+ ".",
+ "'"
+ ],
+ [
+ "b",
+ "s"
+ ],
+ [
+ "in",
+ "fo"
+ ],
+ [
+ "inf",
+ "o"
+ ],
+ [
+ "ch",
+ "n"
+ ],
+ [
+ "c",
+ "hn"
+ ],
+ [
+ "▁we",
+ "ek"
+ ],
+ [
+ "▁",
+ "week"
+ ],
+ [
+ "▁b",
+ "ook"
+ ],
+ [
+ "▁bo",
+ "ok"
+ ],
+ [
+ "▁",
+ "book"
+ ],
+ [
+ "H",
+ "E"
+ ],
+ [
+ "ba",
+ "r"
+ ],
+ [
+ "b",
+ "ar"
+ ],
+ [
+ "ic",
+ "ense"
+ ],
+ [
+ "▁W",
+ "hat"
+ ],
+ [
+ "▁Wh",
+ "at"
+ ],
+ [
+ "▁",
+ "What"
+ ],
+ [
+ "▁qu",
+ "est"
+ ],
+ [
+ "▁que",
+ "st"
+ ],
+ [
+ "▁q",
+ "uest"
+ ],
+ [
+ "▁",
+ "quest"
+ ],
+ [
+ "ur",
+ "ch"
+ ],
+ [
+ "at",
+ "o"
+ ],
+ [
+ "a",
+ "to"
+ ],
+ [
+ "le",
+ "ft"
+ ],
+ [
+ "l",
+ "eft"
+ ],
+ [
+ "▁m",
+ "ar"
+ ],
+ [
+ "▁ma",
+ "r"
+ ],
+ [
+ "▁",
+ "mar"
+ ],
+ [
+ "▁t",
+ "op"
+ ],
+ [
+ "▁to",
+ "p"
+ ],
+ [
+ "▁",
+ "top"
+ ],
+ [
+ "F",
+ "F"
+ ],
+ [
+ "▁f",
+ "riend"
+ ],
+ [
+ "▁",
+ "friend"
+ ],
+ [
+ "▁b",
+ "eh"
+ ],
+ [
+ "▁be",
+ "h"
+ ],
+ [
+ "▁f",
+ "ield"
+ ],
+ [
+ "▁fi",
+ "eld"
+ ],
+ [
+ "▁",
+ "field"
+ ],
+ [
+ "▁again",
+ "st"
+ ],
+ [
+ "ra",
+ "ct"
+ ],
+ [
+ "rac",
+ "t"
+ ],
+ [
+ "r",
+ "act"
+ ],
+ [
+ "iz",
+ "ation"
+ ],
+ [
+ "us",
+ "er"
+ ],
+ [
+ "use",
+ "r"
+ ],
+ [
+ "u",
+ "ser"
+ ],
+ [
+ "ch",
+ "en"
+ ],
+ [
+ "che",
+ "n"
+ ],
+ [
+ "c",
+ "hen"
+ ],
+ [
+ "▁ke",
+ "ep"
+ ],
+ [
+ "▁",
+ "keep"
+ ],
+ [
+ "A",
+ "D"
+ ],
+ [
+ "it",
+ "or"
+ ],
+ [
+ "ito",
+ "r"
+ ],
+ [
+ "i",
+ "tor"
+ ],
+ [
+ "▁n",
+ "on"
+ ],
+ [
+ "▁no",
+ "n"
+ ],
+ [
+ "▁",
+ "non"
+ ],
+ [
+ "ir",
+ "d"
+ ],
+ [
+ "i",
+ "rd"
+ ],
+ [
+ "op",
+ "e"
+ ],
+ [
+ "o",
+ "pe"
+ ],
+ [
+ "▁re",
+ "st"
+ ],
+ [
+ "▁r",
+ "est"
+ ],
+ [
+ "▁res",
+ "t"
+ ],
+ [
+ "▁",
+ "rest"
+ ],
+ [
+ "▁d",
+ "ev"
+ ],
+ [
+ "▁de",
+ "v"
+ ],
+ [
+ "▁",
+ "dev"
+ ],
+ [
+ "▁_",
+ "_"
+ ],
+ [
+ "▁",
+ "__"
+ ],
+ [
+ "▁u",
+ "na"
+ ],
+ [
+ "▁un",
+ "a"
+ ],
+ [
+ "▁",
+ "una"
+ ],
+ [
+ "▁t",
+ "erm"
+ ],
+ [
+ "▁te",
+ "rm"
+ ],
+ [
+ "▁ter",
+ "m"
+ ],
+ [
+ "▁",
+ "term"
+ ],
+ [
+ "I",
+ "S"
+ ],
+ [
+ "▁p",
+ "op"
+ ],
+ [
+ "▁po",
+ "p"
+ ],
+ [
+ "▁",
+ "pop"
+ ],
+ [
+ "ri",
+ "st"
+ ],
+ [
+ "ris",
+ "t"
+ ],
+ [
+ "r",
+ "ist"
+ ],
+ [
+ "▁s",
+ "ince"
+ ],
+ [
+ "▁sin",
+ "ce"
+ ],
+ [
+ "▁sinc",
+ "e"
+ ],
+ [
+ "▁",
+ "since"
+ ],
+ [
+ "ve",
+ "s"
+ ],
+ [
+ "v",
+ "es"
+ ],
+ [
+ "▁h",
+ "ard"
+ ],
+ [
+ "▁ha",
+ "rd"
+ ],
+ [
+ "▁har",
+ "d"
+ ],
+ [
+ "▁",
+ "hard"
+ ],
+ [
+ "p",
+ "i"
+ ],
+ [
+ "ut",
+ "il"
+ ],
+ [
+ "uti",
+ "l"
+ ],
+ [
+ "u",
+ "til"
+ ],
+ [
+ "▁s",
+ "oc"
+ ],
+ [
+ "▁so",
+ "c"
+ ],
+ [
+ "▁",
+ "soc"
+ ],
+ [
+ "en",
+ "e"
+ ],
+ [
+ "e",
+ "ne"
+ ],
+ [
+ "Ex",
+ "ception"
+ ],
+ [
+ "▁l",
+ "ocal"
+ ],
+ [
+ "▁loc",
+ "al"
+ ],
+ [
+ "▁lo",
+ "cal"
+ ],
+ [
+ "▁",
+ "local"
+ ],
+ [
+ "▁d",
+ "irect"
+ ],
+ [
+ "▁di",
+ "rect"
+ ],
+ [
+ "▁dire",
+ "ct"
+ ],
+ [
+ "▁dir",
+ "ect"
+ ],
+ [
+ "▁",
+ "direct"
+ ],
+ [
+ "▁s",
+ "ure"
+ ],
+ [
+ "▁su",
+ "re"
+ ],
+ [
+ "▁sur",
+ "e"
+ ],
+ [
+ "▁",
+ "sure"
+ ],
+ [
+ "▁b",
+ "ro"
+ ],
+ [
+ "▁br",
+ "o"
+ ],
+ [
+ "▁",
+ "bro"
+ ],
+ [
+ "▁d",
+ "a"
+ ],
+ [
+ "▁",
+ "da"
+ ],
+ [
+ "▁<",
+ "/"
+ ],
+ [
+ "▁",
+ ""
+ ],
+ [
+ "▁cur",
+ "rent"
+ ],
+ [
+ "▁curr",
+ "ent"
+ ],
+ [
+ "▁",
+ "current"
+ ],
+ [
+ "'",
+ ":"
+ ],
+ [
+ "W",
+ "h"
+ ],
+ [
+ "▁in",
+ "formation"
+ ],
+ [
+ "▁inform",
+ "ation"
+ ],
+ [
+ "▁",
+ "information"
+ ],
+ [
+ "▁i",
+ "de"
+ ],
+ [
+ "▁id",
+ "e"
+ ],
+ [
+ "▁",
+ "ide"
+ ],
+ [
+ "▁bet",
+ "ter"
+ ],
+ [
+ "Te",
+ "xt"
+ ],
+ [
+ "Tex",
+ "t"
+ ],
+ [
+ "T",
+ "ext"
+ ],
+ [
+ "ra",
+ "ph"
+ ],
+ [
+ "rap",
+ "h"
+ ],
+ [
+ "r",
+ "aph"
+ ],
+ [
+ "▁st",
+ "and"
+ ],
+ [
+ "▁stan",
+ "d"
+ ],
+ [
+ "▁sta",
+ "nd"
+ ],
+ [
+ "▁",
+ "stand"
+ ],
+ [
+ "▁c",
+ "heck"
+ ],
+ [
+ "▁che",
+ "ck"
+ ],
+ [
+ "▁",
+ "check"
+ ],
+ [
+ "▁",
+ "к"
+ ],
+ [
+ "▁n",
+ "a"
+ ],
+ [
+ "▁",
+ "na"
+ ],
+ [
+ "(",
+ "("
+ ],
+ [
+ "ou",
+ "th"
+ ],
+ [
+ "out",
+ "h"
+ ],
+ [
+ "o",
+ "uth"
+ ],
+ [
+ "ap",
+ "s"
+ ],
+ [
+ "a",
+ "ps"
+ ],
+ [
+ "▁u",
+ "nt"
+ ],
+ [
+ "▁un",
+ "t"
+ ],
+ [
+ "▁",
+ "unt"
+ ],
+ [
+ "b",
+ "f"
+ ],
+ [
+ "▁con",
+ "f"
+ ],
+ [
+ "▁co",
+ "nf"
+ ],
+ [
+ "▁",
+ "conf"
+ ],
+ [
+ "▁s",
+ "pe"
+ ],
+ [
+ "▁sp",
+ "e"
+ ],
+ [
+ "▁",
+ "spe"
+ ],
+ [
+ "it",
+ "le"
+ ],
+ [
+ "i",
+ "tle"
+ ],
+ [
+ "▁C",
+ "ol"
+ ],
+ [
+ "▁Co",
+ "l"
+ ],
+ [
+ "▁",
+ "Col"
+ ],
+ [
+ "cl",
+ "ass"
+ ],
+ [
+ "c",
+ "lass"
+ ],
+ [
+ "ur",
+ "al"
+ ],
+ [
+ "ura",
+ "l"
+ ],
+ [
+ "u",
+ "ral"
+ ],
+ [
+ "ber",
+ "s"
+ ],
+ [
+ "be",
+ "rs"
+ ],
+ [
+ "b",
+ "ers"
+ ],
+ [
+ "M",
+ "A"
+ ],
+ [
+ "ess",
+ "ion"
+ ],
+ [
+ "▁",
+ "М"
+ ],
+ [
+ "In",
+ "fo"
+ ],
+ [
+ "Inf",
+ "o"
+ ],
+ [
+ "▁B",
+ "r"
+ ],
+ [
+ "▁",
+ "Br"
+ ],
+ [
+ "▁e",
+ "as"
+ ],
+ [
+ "erv",
+ "ice"
+ ],
+ [
+ "au",
+ "s"
+ ],
+ [
+ "a",
+ "us"
+ ],
+ [
+ "ar",
+ "i"
+ ],
+ [
+ "a",
+ "ri"
+ ],
+ [
+ "п",
+ "о"
+ ],
+ [
+ "▁c",
+ "oun"
+ ],
+ [
+ "▁co",
+ "un"
+ ],
+ [
+ "▁cou",
+ "n"
+ ],
+ [
+ "д",
+ "е"
+ ],
+ [
+ "()",
+ ")"
+ ],
+ [
+ "(",
+ "))"
+ ],
+ [
+ "li",
+ "ng"
+ ],
+ [
+ "lin",
+ "g"
+ ],
+ [
+ "l",
+ "ing"
+ ],
+ [
+ "E",
+ "D"
+ ],
+ [
+ "ab",
+ "ly"
+ ],
+ [
+ "abl",
+ "y"
+ ],
+ [
+ "▁p",
+ "at"
+ ],
+ [
+ "▁pa",
+ "t"
+ ],
+ [
+ "▁",
+ "pat"
+ ],
+ [
+ "or",
+ "g"
+ ],
+ [
+ "o",
+ "rg"
+ ],
+ [
+ "▁i",
+ "d"
+ ],
+ [
+ "▁",
+ "id"
+ ],
+ [
+ "▁",
+ "г"
+ ],
+ [
+ "▁t",
+ "ell"
+ ],
+ [
+ "▁te",
+ "ll"
+ ],
+ [
+ "▁tel",
+ "l"
+ ],
+ [
+ "le",
+ "x"
+ ],
+ [
+ "l",
+ "ex"
+ ],
+ [
+ "▁al",
+ "low"
+ ],
+ [
+ "▁all",
+ "ow"
+ ],
+ [
+ "▁",
+ "allow"
+ ],
+ [
+ "re",
+ "en"
+ ],
+ [
+ "ree",
+ "n"
+ ],
+ [
+ "r",
+ "een"
+ ],
+ [
+ "m",
+ "y"
+ ],
+ [
+ "▁cons",
+ "ider"
+ ],
+ [
+ "▁consid",
+ "er"
+ ],
+ [
+ "▁te",
+ "am"
+ ],
+ [
+ "▁tea",
+ "m"
+ ],
+ [
+ "▁",
+ "team"
+ ],
+ [
+ "le",
+ "ase"
+ ],
+ [
+ "ht",
+ "t"
+ ],
+ [
+ "h",
+ "tt"
+ ],
+ [
+ "▁P",
+ "r"
+ ],
+ [
+ "▁",
+ "Pr"
+ ],
+ [
+ "/*",
+ "*"
+ ],
+ [
+ "/",
+ "**"
+ ],
+ [
+ "▁s",
+ "ing"
+ ],
+ [
+ "▁si",
+ "ng"
+ ],
+ [
+ "▁sin",
+ "g"
+ ],
+ [
+ "▁",
+ "sing"
+ ],
+ [
+ "Re",
+ "qu"
+ ],
+ [
+ "Req",
+ "u"
+ ],
+ [
+ "R",
+ "equ"
+ ],
+ [
+ "R",
+ "e"
+ ],
+ [
+ "id",
+ "es"
+ ],
+ [
+ "ide",
+ "s"
+ ],
+ [
+ "i",
+ "des"
+ ],
+ [
+ "ch",
+ "es"
+ ],
+ [
+ "che",
+ "s"
+ ],
+ [
+ "c",
+ "hes"
+ ],
+ [
+ "▁ob",
+ "ject"
+ ],
+ [
+ "▁obj",
+ "ect"
+ ],
+ [
+ "▁",
+ "object"
+ ],
+ [
+ "ial",
+ "ly"
+ ],
+ [
+ "i",
+ "ally"
+ ],
+ [
+ "B",
+ "y"
+ ],
+ [
+ "с",
+ "я"
+ ],
+ [
+ "id",
+ "ed"
+ ],
+ [
+ "ide",
+ "d"
+ ],
+ [
+ "i",
+ "ded"
+ ],
+ [
+ "▁f",
+ "ree"
+ ],
+ [
+ "▁fr",
+ "ee"
+ ],
+ [
+ "▁fre",
+ "e"
+ ],
+ [
+ "▁",
+ "free"
+ ],
+ [
+ "▁pro",
+ "ble"
+ ],
+ [
+ "▁prob",
+ "le"
+ ],
+ [
+ "ci",
+ "te"
+ ],
+ [
+ "cit",
+ "e"
+ ],
+ [
+ "c",
+ "ite"
+ ],
+ [
+ "▁)",
+ ";"
+ ],
+ [
+ "▁",
+ ");"
+ ],
+ [
+ "iss",
+ "ion"
+ ],
+ [
+ "▁d",
+ "uring"
+ ],
+ [
+ "▁du",
+ "ring"
+ ],
+ [
+ "▁dur",
+ "ing"
+ ],
+ [
+ "▁-",
+ "-"
+ ],
+ [
+ "▁",
+ "--"
+ ],
+ [
+ "it",
+ "her"
+ ],
+ [
+ "ith",
+ "er"
+ ],
+ [
+ "i",
+ "ther"
+ ],
+ [
+ "л",
+ "я"
+ ],
+ [
+ "▁l",
+ "eg"
+ ],
+ [
+ "▁le",
+ "g"
+ ],
+ [
+ "▁",
+ "leg"
+ ],
+ [
+ "▁s",
+ "it"
+ ],
+ [
+ "▁si",
+ "t"
+ ],
+ [
+ "ic",
+ "ally"
+ ],
+ [
+ "ical",
+ "ly"
+ ],
+ [
+ "▁k",
+ "ey"
+ ],
+ [
+ "▁ke",
+ "y"
+ ],
+ [
+ "▁",
+ "key"
+ ],
+ [
+ "le",
+ "g"
+ ],
+ [
+ "l",
+ "eg"
+ ],
+ [
+ "tr",
+ "a"
+ ],
+ [
+ "t",
+ "ra"
+ ],
+ [
+ "▁m",
+ "om"
+ ],
+ [
+ "▁mo",
+ "m"
+ ],
+ [
+ "▁ex",
+ "pl"
+ ],
+ [
+ "▁exp",
+ "l"
+ ],
+ [
+ "▁",
+ "expl"
+ ],
+ [
+ "▁de",
+ "velop"
+ ],
+ [
+ "▁",
+ "develop"
+ ],
+ [
+ "▁e",
+ "vent"
+ ],
+ [
+ "▁ev",
+ "ent"
+ ],
+ [
+ "▁even",
+ "t"
+ ],
+ [
+ "▁",
+ "event"
+ ],
+ [
+ "▁N",
+ "ULL"
+ ],
+ [
+ "▁",
+ "NULL"
+ ],
+ [
+ "oh",
+ "n"
+ ],
+ [
+ "o",
+ "hn"
+ ],
+ [
+ "▁//",
+ "/"
+ ],
+ [
+ "▁/",
+ "//"
+ ],
+ [
+ "▁",
+ "///"
+ ],
+ [
+ "▁bus",
+ "iness"
+ ],
+ [
+ "▁",
+ "business"
+ ],
+ [
+ "ч",
+ "а"
+ ],
+ [
+ "▁pro",
+ "f"
+ ],
+ [
+ "▁pr",
+ "of"
+ ],
+ [
+ "▁",
+ "prof"
+ ],
+ [
+ "er",
+ "ror"
+ ],
+ [
+ "err",
+ "or"
+ ],
+ [
+ "▁p",
+ "or"
+ ],
+ [
+ "▁po",
+ "r"
+ ],
+ [
+ "▁",
+ "por"
+ ],
+ [
+ "▁com",
+ "mun"
+ ],
+ [
+ "▁comm",
+ "un"
+ ],
+ [
+ "▁",
+ "commun"
+ ],
+ [
+ "In",
+ "d"
+ ],
+ [
+ "I",
+ "nd"
+ ],
+ [
+ "iu",
+ "m"
+ ],
+ [
+ "i",
+ "um"
+ ],
+ [
+ "Te",
+ "st"
+ ],
+ [
+ "T",
+ "est"
+ ],
+ [
+ "▁A",
+ "d"
+ ],
+ [
+ "▁",
+ "Ad"
+ ],
+ [
+ "ou",
+ "ble"
+ ],
+ [
+ "▁s",
+ "on"
+ ],
+ [
+ "▁so",
+ "n"
+ ],
+ [
+ "▁",
+ "son"
+ ],
+ [
+ "ri",
+ "te"
+ ],
+ [
+ "rit",
+ "e"
+ ],
+ [
+ "r",
+ "ite"
+ ],
+ [
+ "re",
+ "ady"
+ ],
+ [
+ "read",
+ "y"
+ ],
+ [
+ "rea",
+ "dy"
+ ],
+ [
+ "▁{",
+ "\r"
+ ],
+ [
+ "▁",
+ "{\r"
+ ],
+ [
+ "▁t",
+ "hing"
+ ],
+ [
+ "▁th",
+ "ing"
+ ],
+ [
+ "▁thin",
+ "g"
+ ],
+ [
+ "▁",
+ "thing"
+ ],
+ [
+ "н",
+ "я"
+ ],
+ [
+ "▁P",
+ "h"
+ ],
+ [
+ "▁",
+ "Ph"
+ ],
+ [
+ "pe",
+ "d"
+ ],
+ [
+ "p",
+ "ed"
+ ],
+ [
+ "с",
+ "ь"
+ ],
+ [
+ "iv",
+ "ed"
+ ],
+ [
+ "ive",
+ "d"
+ ],
+ [
+ "i",
+ "ved"
+ ],
+ [
+ "Y",
+ "ou"
+ ],
+ [
+ "ar",
+ "l"
+ ],
+ [
+ "a",
+ "rl"
+ ],
+ [
+ "con",
+ "st"
+ ],
+ [
+ "cons",
+ "t"
+ ],
+ [
+ "..",
+ "/"
+ ],
+ [
+ ".",
+ "./"
+ ],
+ [
+ "S",
+ "e"
+ ],
+ [
+ "S",
+ "h"
+ ],
+ [
+ "▁p",
+ "ower"
+ ],
+ [
+ "▁po",
+ "wer"
+ ],
+ [
+ "▁pow",
+ "er"
+ ],
+ [
+ "▁",
+ "power"
+ ],
+ [
+ "rib",
+ "ute"
+ ],
+ [
+ "ribut",
+ "e"
+ ],
+ [
+ "ribu",
+ "te"
+ ],
+ [
+ "▁M",
+ "y"
+ ],
+ [
+ "▁",
+ "My"
+ ],
+ [
+ "▁t",
+ "alk"
+ ],
+ [
+ "▁tal",
+ "k"
+ ],
+ [
+ "▁",
+ "talk"
+ ],
+ [
+ "it",
+ "ch"
+ ],
+ [
+ "▁c",
+ "alled"
+ ],
+ [
+ "▁call",
+ "ed"
+ ],
+ [
+ "▁cal",
+ "led"
+ ],
+ [
+ "▁",
+ "called"
+ ],
+ [
+ "▁c",
+ "ame"
+ ],
+ [
+ "▁cam",
+ "e"
+ ],
+ [
+ "▁ca",
+ "me"
+ ],
+ [
+ "▁be",
+ "lie"
+ ],
+ [
+ "▁bel",
+ "ie"
+ ],
+ [
+ "U",
+ "R"
+ ],
+ [
+ "Ad",
+ "d"
+ ],
+ [
+ "A",
+ "dd"
+ ],
+ [
+ "▁R",
+ "es"
+ ],
+ [
+ "▁Re",
+ "s"
+ ],
+ [
+ "▁",
+ "Res"
+ ],
+ [
+ "as",
+ "ter"
+ ],
+ [
+ "ast",
+ "er"
+ ],
+ [
+ "aste",
+ "r"
+ ],
+ [
+ "a",
+ "ster"
+ ],
+ [
+ "el",
+ "la"
+ ],
+ [
+ "ell",
+ "a"
+ ],
+ [
+ "e",
+ "lla"
+ ],
+ [
+ "ob",
+ "al"
+ ],
+ [
+ "oba",
+ "l"
+ ],
+ [
+ "o",
+ "bal"
+ ],
+ [
+ "▁u",
+ "ntil"
+ ],
+ [
+ "▁un",
+ "til"
+ ],
+ [
+ "▁unt",
+ "il"
+ ],
+ [
+ "▁",
+ "until"
+ ],
+ [
+ "▁h",
+ "um"
+ ],
+ [
+ "▁",
+ "hum"
+ ],
+ [
+ "C",
+ "O"
+ ],
+ [
+ "at",
+ "ely"
+ ],
+ [
+ "ate",
+ "ly"
+ ],
+ [
+ "atel",
+ "y"
+ ],
+ [
+ "##",
+ "##"
+ ],
+ [
+ "###",
+ "#"
+ ],
+ [
+ "#",
+ "###"
+ ],
+ [
+ "pu",
+ "blic"
+ ],
+ [
+ "pub",
+ "lic"
+ ],
+ [
+ "p",
+ "ublic"
+ ],
+ [
+ "[",
+ "]"
+ ],
+ [
+ "▁r",
+ "oom"
+ ],
+ [
+ "▁ro",
+ "om"
+ ],
+ [
+ "▁",
+ "room"
+ ],
+ [
+ "le",
+ "n"
+ ],
+ [
+ "l",
+ "en"
+ ],
+ [
+ "▁f",
+ "amily"
+ ],
+ [
+ "▁fam",
+ "ily"
+ ],
+ [
+ "▁famil",
+ "y"
+ ],
+ [
+ "▁",
+ "family"
+ ],
+ [
+ "po",
+ "r"
+ ],
+ [
+ "p",
+ "or"
+ ],
+ [
+ "▁pro",
+ "gram"
+ ],
+ [
+ "▁pr",
+ "ogram"
+ ],
+ [
+ "▁",
+ "program"
+ ],
+ [
+ "▁h",
+ "ist"
+ ],
+ [
+ "▁his",
+ "t"
+ ],
+ [
+ "▁hi",
+ "st"
+ ],
+ [
+ "▁",
+ "hist"
+ ],
+ [
+ "▁m",
+ "us"
+ ],
+ [
+ "▁mu",
+ "s"
+ ],
+ [
+ "▁",
+ "mus"
+ ],
+ [
+ "ar",
+ "ge"
+ ],
+ [
+ "arg",
+ "e"
+ ],
+ [
+ "on",
+ "ey"
+ ],
+ [
+ "one",
+ "y"
+ ],
+ [
+ "o",
+ "ney"
+ ],
+ [
+ "I",
+ "m"
+ ],
+ [
+ "el",
+ "se"
+ ],
+ [
+ "els",
+ "e"
+ ],
+ [
+ "ail",
+ "s"
+ ],
+ [
+ "ai",
+ "ls"
+ ],
+ [
+ "a",
+ "ils"
+ ],
+ [
+ "a",
+ "f"
+ ],
+ [
+ "▁l",
+ "ove"
+ ],
+ [
+ "▁lo",
+ "ve"
+ ],
+ [
+ "▁lov",
+ "e"
+ ],
+ [
+ "▁",
+ "love"
+ ],
+ [
+ "ä",
+ "r"
+ ],
+ [
+ "as",
+ "es"
+ ],
+ [
+ "ase",
+ "s"
+ ],
+ [
+ "a",
+ "ses"
+ ],
+ [
+ "ph",
+ "a"
+ ],
+ [
+ "p",
+ "ha"
+ ],
+ [
+ "ou",
+ "rs"
+ ],
+ [
+ "our",
+ "s"
+ ],
+ [
+ "o",
+ "urs"
+ ],
+ [
+ "di",
+ "s"
+ ],
+ [
+ "d",
+ "is"
+ ],
+ [
+ "ma",
+ "p"
+ ],
+ [
+ "m",
+ "ap"
+ ],
+ [
+ "iv",
+ "er"
+ ],
+ [
+ "ive",
+ "r"
+ ],
+ [
+ "i",
+ "ver"
+ ],
+ [
+ "ö",
+ "r"
+ ],
+ [
+ "▁B",
+ "l"
+ ],
+ [
+ "▁",
+ "Bl"
+ ],
+ [
+ "at",
+ "eg"
+ ],
+ [
+ "ate",
+ "g"
+ ],
+ [
+ "st",
+ "ate"
+ ],
+ [
+ "stat",
+ "e"
+ ],
+ [
+ "sta",
+ "te"
+ ],
+ [
+ "St",
+ "ate"
+ ],
+ [
+ "Stat",
+ "e"
+ ],
+ [
+ "er",
+ "tain"
+ ],
+ [
+ "ert",
+ "ain"
+ ],
+ [
+ "erta",
+ "in"
+ ],
+ [
+ "▁e",
+ "ffect"
+ ],
+ [
+ "▁eff",
+ "ect"
+ ],
+ [
+ "▁",
+ "effect"
+ ],
+ [
+ "pr",
+ "int"
+ ],
+ [
+ "▁b",
+ "ig"
+ ],
+ [
+ "▁bi",
+ "g"
+ ],
+ [
+ "▁",
+ "big"
+ ],
+ [
+ "in",
+ "dex"
+ ],
+ [
+ "ind",
+ "ex"
+ ],
+ [
+ "inde",
+ "x"
+ ],
+ [
+ "▁p",
+ "ub"
+ ],
+ [
+ "▁pu",
+ "b"
+ ],
+ [
+ "▁",
+ "pub"
+ ],
+ [
+ "ve",
+ "rt"
+ ],
+ [
+ "ver",
+ "t"
+ ],
+ [
+ "v",
+ "ert"
+ ],
+ [
+ "er",
+ "o"
+ ],
+ [
+ "e",
+ "ro"
+ ],
+ [
+ "m",
+ "d"
+ ],
+ [
+ "▁m",
+ "ethod"
+ ],
+ [
+ "▁meth",
+ "od"
+ ],
+ [
+ "▁",
+ "method"
+ ],
+ [
+ "▁g",
+ "ame"
+ ],
+ [
+ "▁gam",
+ "e"
+ ],
+ [
+ "▁ga",
+ "me"
+ ],
+ [
+ "▁",
+ "game"
+ ],
+ [
+ "ri",
+ "es"
+ ],
+ [
+ "rie",
+ "s"
+ ],
+ [
+ "r",
+ "ies"
+ ],
+ [
+ "le",
+ "te"
+ ],
+ [
+ "let",
+ "e"
+ ],
+ [
+ "l",
+ "ete"
+ ],
+ [
+ "It",
+ "em"
+ ],
+ [
+ "I",
+ "tem"
+ ],
+ [
+ "IN",
+ "G"
+ ],
+ [
+ "I",
+ "NG"
+ ],
+ [
+ "re",
+ "sent"
+ ],
+ [
+ "res",
+ "ent"
+ ],
+ [
+ "al",
+ "ity"
+ ],
+ [
+ "ali",
+ "ty"
+ ],
+ [
+ "pt",
+ "y"
+ ],
+ [
+ "p",
+ "ty"
+ ],
+ [
+ "le",
+ "y"
+ ],
+ [
+ "l",
+ "ey"
+ ],
+ [
+ "oc",
+ "ument"
+ ],
+ [
+ "▁b",
+ "eg"
+ ],
+ [
+ "▁be",
+ "g"
+ ],
+ [
+ "T",
+ "R"
+ ],
+ [
+ "}",
+ "."
+ ],
+ [
+ "▁sch",
+ "ool"
+ ],
+ [
+ "▁",
+ "school"
+ ],
+ [
+ "he",
+ "s"
+ ],
+ [
+ "h",
+ "es"
+ ],
+ [
+ "д",
+ "о"
+ ],
+ [
+ "▁l",
+ "ot"
+ ],
+ [
+ "▁lo",
+ "t"
+ ],
+ [
+ "▁",
+ "lot"
+ ],
+ [
+ "▁t",
+ "ook"
+ ],
+ [
+ "▁to",
+ "ok"
+ ],
+ [
+ "▁too",
+ "k"
+ ],
+ [
+ "▁a",
+ "dv"
+ ],
+ [
+ "▁ad",
+ "v"
+ ],
+ [
+ "▁",
+ "adv"
+ ],
+ [
+ "▁c",
+ "ap"
+ ],
+ [
+ "▁ca",
+ "p"
+ ],
+ [
+ "▁",
+ "cap"
+ ],
+ [
+ "M",
+ "P"
+ ],
+ [
+ "un",
+ "k"
+ ],
+ [
+ "▁l",
+ "ight"
+ ],
+ [
+ "▁li",
+ "ght"
+ ],
+ [
+ "▁lig",
+ "ht"
+ ],
+ [
+ "▁",
+ "light"
+ ],
+ [
+ "▁l",
+ "ater"
+ ],
+ [
+ "▁la",
+ "ter"
+ ],
+ [
+ "▁late",
+ "r"
+ ],
+ [
+ "▁lat",
+ "er"
+ ],
+ [
+ ".",
+ ","
+ ],
+ [
+ "Ke",
+ "y"
+ ],
+ [
+ "K",
+ "ey"
+ ],
+ [
+ "it",
+ "ions"
+ ],
+ [
+ "ition",
+ "s"
+ ],
+ [
+ "iti",
+ "ons"
+ ],
+ [
+ "▁en",
+ "ough"
+ ],
+ [
+ "▁/",
+ "**"
+ ],
+ [
+ "▁/*",
+ "*"
+ ],
+ [
+ "▁",
+ "/**"
+ ],
+ [
+ "▁w",
+ "ent"
+ ],
+ [
+ "▁we",
+ "nt"
+ ],
+ [
+ "▁wen",
+ "t"
+ ],
+ [
+ "ã",
+ "o"
+ ],
+ [
+ "▁th",
+ "ough"
+ ],
+ [
+ "▁thou",
+ "gh"
+ ],
+ [
+ "▁",
+ "though"
+ ],
+ [
+ "▁g",
+ "roup"
+ ],
+ [
+ "▁gr",
+ "oup"
+ ],
+ [
+ "▁gro",
+ "up"
+ ],
+ [
+ "▁",
+ "group"
+ ],
+ [
+ "▁me",
+ "an"
+ ],
+ [
+ "▁",
+ "mean"
+ ],
+ [
+ "ск",
+ "и"
+ ],
+ [
+ "с",
+ "ки"
+ ],
+ [
+ "A",
+ "P"
+ ],
+ [
+ "▁n",
+ "um"
+ ],
+ [
+ "▁nu",
+ "m"
+ ],
+ [
+ "▁",
+ "num"
+ ],
+ [
+ "▁c",
+ "ond"
+ ],
+ [
+ "▁con",
+ "d"
+ ],
+ [
+ "▁co",
+ "nd"
+ ],
+ [
+ "▁",
+ "cond"
+ ],
+ [
+ "н",
+ "і"
+ ],
+ [
+ "▁g",
+ "iven"
+ ],
+ [
+ "▁giv",
+ "en"
+ ],
+ [
+ "▁give",
+ "n"
+ ],
+ [
+ "▁gi",
+ "ven"
+ ],
+ [
+ "▁w",
+ "hy"
+ ],
+ [
+ "▁wh",
+ "y"
+ ],
+ [
+ "▁",
+ "why"
+ ],
+ [
+ "▁re",
+ "ce"
+ ],
+ [
+ "▁rec",
+ "e"
+ ],
+ [
+ "▁s",
+ "ide"
+ ],
+ [
+ "▁si",
+ "de"
+ ],
+ [
+ "▁sid",
+ "e"
+ ],
+ [
+ "▁",
+ "side"
+ ],
+ [
+ "▁f",
+ "ar"
+ ],
+ [
+ "▁fa",
+ "r"
+ ],
+ [
+ "▁",
+ "far"
+ ],
+ [
+ "Con",
+ "text"
+ ],
+ [
+ "Cont",
+ "ext"
+ ],
+ [
+ "м",
+ "е"
+ ],
+ [
+ "▁l",
+ "og"
+ ],
+ [
+ "▁lo",
+ "g"
+ ],
+ [
+ "▁",
+ "log"
+ ],
+ [
+ "Vi",
+ "ew"
+ ],
+ [
+ "V",
+ "iew"
+ ],
+ [
+ "▁<",
+ "<"
+ ],
+ [
+ "▁",
+ "<<"
+ ],
+ [
+ "fi",
+ "l"
+ ],
+ [
+ "f",
+ "il"
+ ],
+ [
+ "ac",
+ "es"
+ ],
+ [
+ "ace",
+ "s"
+ ],
+ [
+ "a",
+ "ces"
+ ],
+ [
+ "en",
+ "cy"
+ ],
+ [
+ "enc",
+ "y"
+ ],
+ [
+ "oa",
+ "d"
+ ],
+ [
+ "o",
+ "ad"
+ ],
+ [
+ "er",
+ "ed"
+ ],
+ [
+ "ere",
+ "d"
+ ],
+ [
+ "e",
+ "red"
+ ],
+ [
+ "▁pro",
+ "duct"
+ ],
+ [
+ "▁produ",
+ "ct"
+ ],
+ [
+ "▁prod",
+ "uct"
+ ],
+ [
+ "▁",
+ "product"
+ ],
+ [
+ "E",
+ "T"
+ ],
+ [
+ "▁p",
+ "aram"
+ ],
+ [
+ "▁par",
+ "am"
+ ],
+ [
+ "▁para",
+ "m"
+ ],
+ [
+ "▁pa",
+ "ram"
+ ],
+ [
+ "▁",
+ "param"
+ ],
+ [
+ "▁p",
+ "rote"
+ ],
+ [
+ "▁pro",
+ "te"
+ ],
+ [
+ "▁pr",
+ "ote"
+ ],
+ [
+ "▁prot",
+ "e"
+ ],
+ [
+ "▁",
+ "prote"
+ ],
+ [
+ "te",
+ "s"
+ ],
+ [
+ "t",
+ "es"
+ ],
+ [
+ "Tim",
+ "e"
+ ],
+ [
+ "T",
+ "ime"
+ ],
+ [
+ "j",
+ "e"
+ ],
+ [
+ "ol",
+ "ution"
+ ],
+ [
+ "olut",
+ "ion"
+ ],
+ [
+ "▁р",
+ "а"
+ ],
+ [
+ "▁",
+ "ра"
+ ],
+ [
+ "▁mon",
+ "th"
+ ],
+ [
+ "▁mont",
+ "h"
+ ],
+ [
+ "▁",
+ "month"
+ ],
+ [
+ "fer",
+ "ence"
+ ],
+ [
+ "fe",
+ "rence"
+ ],
+ [
+ "▁a",
+ "ppe"
+ ],
+ [
+ "▁app",
+ "e"
+ ],
+ [
+ "▁ap",
+ "pe"
+ ],
+ [
+ "▁",
+ "appe"
+ ],
+ [
+ "▁f",
+ "ace"
+ ],
+ [
+ "▁fac",
+ "e"
+ ],
+ [
+ "▁fa",
+ "ce"
+ ],
+ [
+ "▁",
+ "face"
+ ],
+ [
+ "en",
+ "ed"
+ ],
+ [
+ "ene",
+ "d"
+ ],
+ [
+ "e",
+ "ned"
+ ],
+ [
+ "tr",
+ "act"
+ ],
+ [
+ "tra",
+ "ct"
+ ],
+ [
+ "t",
+ "ract"
+ ],
+ [
+ "▁l",
+ "ess"
+ ],
+ [
+ "▁le",
+ "ss"
+ ],
+ [
+ "▁les",
+ "s"
+ ],
+ [
+ "▁",
+ "less"
+ ],
+ [
+ "A",
+ "S"
+ ],
+ [
+ "é",
+ "e"
+ ],
+ [
+ "▁g",
+ "ive"
+ ],
+ [
+ "▁giv",
+ "e"
+ ],
+ [
+ "▁gi",
+ "ve"
+ ],
+ [
+ "▁k",
+ "ind"
+ ],
+ [
+ "▁ki",
+ "nd"
+ ],
+ [
+ "▁kin",
+ "d"
+ ],
+ [
+ "▁",
+ "kind"
+ ],
+ [
+ "▁c",
+ "ount"
+ ],
+ [
+ "▁co",
+ "unt"
+ ],
+ [
+ "▁coun",
+ "t"
+ ],
+ [
+ "▁cou",
+ "nt"
+ ],
+ [
+ "▁",
+ "count"
+ ],
+ [
+ "co",
+ "unt"
+ ],
+ [
+ "cou",
+ "nt"
+ ],
+ [
+ "c",
+ "ount"
+ ],
+ [
+ "▁s",
+ "top"
+ ],
+ [
+ "▁st",
+ "op"
+ ],
+ [
+ "▁sto",
+ "p"
+ ],
+ [
+ "▁",
+ "stop"
+ ],
+ [
+ "▁g",
+ "over"
+ ],
+ [
+ "▁go",
+ "ver"
+ ],
+ [
+ "k",
+ "a"
+ ],
+ [
+ "▁err",
+ "or"
+ ],
+ [
+ "▁er",
+ "ror"
+ ],
+ [
+ "▁",
+ "error"
+ ],
+ [
+ "en",
+ "ces"
+ ],
+ [
+ "ence",
+ "s"
+ ],
+ [
+ "enc",
+ "es"
+ ],
+ [
+ "▁m",
+ "il"
+ ],
+ [
+ "▁mi",
+ "l"
+ ],
+ [
+ "▁",
+ "mil"
+ ],
+ [
+ "al",
+ "f"
+ ],
+ [
+ "yn",
+ "c"
+ ],
+ [
+ "y",
+ "nc"
+ ],
+ [
+ "vi",
+ "ous"
+ ],
+ [
+ "v",
+ "ious"
+ ],
+ [
+ "h",
+ "o"
+ ],
+ [
+ "▁n",
+ "ight"
+ ],
+ [
+ "▁ni",
+ "ght"
+ ],
+ [
+ "▁",
+ "night"
+ ],
+ [
+ "er",
+ "a"
+ ],
+ [
+ "e",
+ "ra"
+ ],
+ [
+ "▁п",
+ "ро"
+ ],
+ [
+ "▁пр",
+ "о"
+ ],
+ [
+ "▁",
+ "про"
+ ],
+ [
+ "▁s",
+ "ol"
+ ],
+ [
+ "▁so",
+ "l"
+ ],
+ [
+ "▁",
+ "sol"
+ ],
+ [
+ "me",
+ "n"
+ ],
+ [
+ "m",
+ "en"
+ ],
+ [
+ "▁w",
+ "ater"
+ ],
+ [
+ "▁wat",
+ "er"
+ ],
+ [
+ "▁wa",
+ "ter"
+ ],
+ [
+ "▁",
+ "water"
+ ],
+ [
+ "er",
+ "ing"
+ ],
+ [
+ "eri",
+ "ng"
+ ],
+ [
+ "e",
+ "ring"
+ ],
+ [
+ "▁l",
+ "im"
+ ],
+ [
+ "▁li",
+ "m"
+ ],
+ [
+ "▁",
+ "lim"
+ ],
+ [
+ "Par",
+ "am"
+ ],
+ [
+ "P",
+ "aram"
+ ],
+ [
+ "▁h",
+ "ouse"
+ ],
+ [
+ "▁hous",
+ "e"
+ ],
+ [
+ "▁ho",
+ "use"
+ ],
+ [
+ "▁",
+ "house"
+ ],
+ [
+ "▁S",
+ "ystem"
+ ],
+ [
+ "▁",
+ "System"
+ ],
+ [
+ "▁p",
+ "ay"
+ ],
+ [
+ "▁pa",
+ "y"
+ ],
+ [
+ "▁",
+ "pay"
+ ],
+ [
+ "▁:",
+ "="
+ ],
+ [
+ "ur",
+ "o"
+ ],
+ [
+ "u",
+ "ro"
+ ],
+ [
+ "oc",
+ "i"
+ ],
+ [
+ "o",
+ "ci"
+ ],
+ [
+ "z",
+ "y"
+ ],
+ [
+ "▁al",
+ "ready"
+ ],
+ [
+ ",",
+ "\\"
+ ],
+ [
+ "le",
+ "ngth"
+ ],
+ [
+ "l",
+ "ength"
+ ],
+ [
+ "▁s",
+ "i"
+ ],
+ [
+ "▁",
+ "si"
+ ],
+ [
+ "▁inter",
+ "est"
+ ],
+ [
+ "▁inte",
+ "rest"
+ ],
+ [
+ "▁",
+ "interest"
+ ],
+ [
+ "af",
+ "f"
+ ],
+ [
+ "a",
+ "ff"
+ ],
+ [
+ "ct",
+ "ed"
+ ],
+ [
+ "c",
+ "ted"
+ ],
+ [
+ "ent",
+ "ion"
+ ],
+ [
+ "enti",
+ "on"
+ ],
+ [
+ "▁д",
+ "о"
+ ],
+ [
+ "▁",
+ "до"
+ ],
+ [
+ "um",
+ "e"
+ ],
+ [
+ "u",
+ "me"
+ ],
+ [
+ "▁app",
+ "ro"
+ ],
+ [
+ "▁ap",
+ "pro"
+ ],
+ [
+ "▁",
+ "appro"
+ ],
+ [
+ "br",
+ "e"
+ ],
+ [
+ "b",
+ "re"
+ ],
+ [
+ "I",
+ "G"
+ ],
+ [
+ "▁th",
+ "row"
+ ],
+ [
+ "▁thr",
+ "ow"
+ ],
+ [
+ "▁thro",
+ "w"
+ ],
+ [
+ "▁",
+ "throw"
+ ],
+ [
+ "math",
+ "cal"
+ ],
+ [
+ "ir",
+ "l"
+ ],
+ [
+ "i",
+ "rl"
+ ],
+ [
+ "▁p",
+ "rom"
+ ],
+ [
+ "▁pro",
+ "m"
+ ],
+ [
+ "▁pr",
+ "om"
+ ],
+ [
+ "▁",
+ "prom"
+ ],
+ [
+ "os",
+ "s"
+ ],
+ [
+ "o",
+ "ss"
+ ],
+ [
+ "▁re",
+ "quest"
+ ],
+ [
+ "▁requ",
+ "est"
+ ],
+ [
+ "▁req",
+ "uest"
+ ],
+ [
+ "▁",
+ "request"
+ ],
+ [
+ "equ",
+ "ation"
+ ],
+ [
+ "eq",
+ "uation"
+ ],
+ [
+ "ol",
+ "ogy"
+ ],
+ [
+ "olog",
+ "y"
+ ],
+ [
+ "olo",
+ "gy"
+ ],
+ [
+ "mi",
+ "t"
+ ],
+ [
+ "m",
+ "it"
+ ],
+ [
+ "▁p",
+ "ack"
+ ],
+ [
+ "▁pa",
+ "ck"
+ ],
+ [
+ "▁pac",
+ "k"
+ ],
+ [
+ "▁",
+ "pack"
+ ],
+ [
+ "in",
+ "o"
+ ],
+ [
+ "i",
+ "no"
+ ],
+ [
+ "ar",
+ "ray"
+ ],
+ [
+ "arr",
+ "ay"
+ ],
+ [
+ "z",
+ "a"
+ ],
+ [
+ "ti",
+ "l"
+ ],
+ [
+ "t",
+ "il"
+ ],
+ [
+ "U",
+ "N"
+ ],
+ [
+ "▁p",
+ "resent"
+ ],
+ [
+ "▁pre",
+ "sent"
+ ],
+ [
+ "▁pres",
+ "ent"
+ ],
+ [
+ "▁",
+ "present"
+ ],
+ [
+ "▁or",
+ "gan"
+ ],
+ [
+ "▁org",
+ "an"
+ ],
+ [
+ "▁",
+ "organ"
+ ],
+ [
+ "Fil",
+ "e"
+ ],
+ [
+ "Fi",
+ "le"
+ ],
+ [
+ "F",
+ "ile"
+ ],
+ [
+ "▁o",
+ "rig"
+ ],
+ [
+ "▁or",
+ "ig"
+ ],
+ [
+ "▁",
+ "orig"
+ ],
+ [
+ "▁f",
+ "ull"
+ ],
+ [
+ "▁ful",
+ "l"
+ ],
+ [
+ "▁fu",
+ "ll"
+ ],
+ [
+ "▁",
+ "full"
+ ],
+ [
+ "is",
+ "tr"
+ ],
+ [
+ "ist",
+ "r"
+ ],
+ [
+ "i",
+ "str"
+ ],
+ [
+ "▁f",
+ "lo"
+ ],
+ [
+ "▁fl",
+ "o"
+ ],
+ [
+ "h",
+ "r"
+ ],
+ [
+ "▁as",
+ "sert"
+ ],
+ [
+ "▁ass",
+ "ert"
+ ],
+ [
+ "▁",
+ "assert"
+ ],
+ [
+ "ar",
+ "ds"
+ ],
+ [
+ "ard",
+ "s"
+ ],
+ [
+ "ur",
+ "l"
+ ],
+ [
+ "u",
+ "rl"
+ ],
+ [
+ "en",
+ "n"
+ ],
+ [
+ "e",
+ "nn"
+ ],
+ [
+ "s",
+ "l"
+ ],
+ [
+ "▁",
+ "А"
+ ],
+ [
+ "▁c",
+ "ho"
+ ],
+ [
+ "▁ch",
+ "o"
+ ],
+ [
+ "▁",
+ "cho"
+ ],
+ [
+ "▁l",
+ "evel"
+ ],
+ [
+ "▁le",
+ "vel"
+ ],
+ [
+ "▁lev",
+ "el"
+ ],
+ [
+ "▁",
+ "level"
+ ],
+ [
+ "O",
+ "T"
+ ],
+ [
+ "wo",
+ "rd"
+ ],
+ [
+ "wor",
+ "d"
+ ],
+ [
+ "w",
+ "ord"
+ ],
+ [
+ "▁b",
+ "ody"
+ ],
+ [
+ "▁bo",
+ "dy"
+ ],
+ [
+ "▁bod",
+ "y"
+ ],
+ [
+ "▁",
+ "body"
+ ],
+ [
+ "▁u",
+ "ser"
+ ],
+ [
+ "▁us",
+ "er"
+ ],
+ [
+ "▁use",
+ "r"
+ ],
+ [
+ "▁",
+ "user"
+ ],
+ [
+ "í",
+ "a"
+ ],
+ [
+ "Q",
+ "u"
+ ],
+ [
+ "▁m",
+ "ain"
+ ],
+ [
+ "▁ma",
+ "in"
+ ],
+ [
+ "▁mai",
+ "n"
+ ],
+ [
+ "▁",
+ "main"
+ ],
+ [
+ "A",
+ "B"
+ ],
+ [
+ "pl",
+ "oy"
+ ],
+ [
+ "plo",
+ "y"
+ ],
+ [
+ "Ev",
+ "ent"
+ ],
+ [
+ "Even",
+ "t"
+ ],
+ [
+ "E",
+ "vent"
+ ],
+ [
+ "▁s",
+ "uper"
+ ],
+ [
+ "▁su",
+ "per"
+ ],
+ [
+ "▁sup",
+ "er"
+ ],
+ [
+ "▁",
+ "super"
+ ],
+ [
+ "ok",
+ "en"
+ ],
+ [
+ "oke",
+ "n"
+ ],
+ [
+ "o",
+ "ken"
+ ],
+ [
+ "▁",
+ "Н"
+ ],
+ [
+ "A",
+ "s"
+ ],
+ [
+ "th",
+ "ers"
+ ],
+ [
+ "ther",
+ "s"
+ ],
+ [
+ "the",
+ "rs"
+ ],
+ [
+ "м",
+ "о"
+ ],
+ [
+ "к",
+ "у"
+ ],
+ [
+ "▁d",
+ "ays"
+ ],
+ [
+ "▁day",
+ "s"
+ ],
+ [
+ "▁da",
+ "ys"
+ ],
+ [
+ "▁",
+ "days"
+ ],
+ [
+ "▁d",
+ "one"
+ ],
+ [
+ "▁do",
+ "ne"
+ ],
+ [
+ "▁don",
+ "e"
+ ],
+ [
+ "▁",
+ "done"
+ ],
+ [
+ "▁v",
+ "iew"
+ ],
+ [
+ "▁vi",
+ "ew"
+ ],
+ [
+ "▁vie",
+ "w"
+ ],
+ [
+ "▁",
+ "view"
+ ],
+ [
+ "si",
+ "de"
+ ],
+ [
+ "sid",
+ "e"
+ ],
+ [
+ "s",
+ "ide"
+ ],
+ [
+ "с",
+ "и"
+ ],
+ [
+ "')",
+ ";"
+ ],
+ [
+ "'",
+ ");"
+ ],
+ [
+ "▁v",
+ "ol"
+ ],
+ [
+ "▁vo",
+ "l"
+ ],
+ [
+ "▁",
+ "vol"
+ ],
+ [
+ "▁t",
+ "ot"
+ ],
+ [
+ "▁to",
+ "t"
+ ],
+ [
+ "▁",
+ "tot"
+ ],
+ [
+ "ca",
+ "se"
+ ],
+ [
+ "cas",
+ "e"
+ ],
+ [
+ "c",
+ "ase"
+ ],
+ [
+ "▁a",
+ "ff"
+ ],
+ [
+ "▁af",
+ "f"
+ ],
+ [
+ "▁",
+ "aff"
+ ],
+ [
+ "Requ",
+ "est"
+ ],
+ [
+ "Re",
+ "quest"
+ ],
+ [
+ "Req",
+ "uest"
+ ],
+ [
+ "▁M",
+ "an"
+ ],
+ [
+ "▁Ma",
+ "n"
+ ],
+ [
+ "▁",
+ "Man"
+ ],
+ [
+ "\\",
+ "\\"
+ ],
+ [
+ "▁J",
+ "ohn"
+ ],
+ [
+ "▁Jo",
+ "hn"
+ ],
+ [
+ "▁Joh",
+ "n"
+ ],
+ [
+ "▁",
+ "John"
+ ],
+ [
+ "▁",
+ "Б"
+ ],
+ [
+ "or",
+ "th"
+ ],
+ [
+ "ort",
+ "h"
+ ],
+ [
+ "▁j",
+ "e"
+ ],
+ [
+ "▁",
+ "je"
+ ],
+ [
+ "▁u",
+ "ne"
+ ],
+ [
+ "▁un",
+ "e"
+ ],
+ [
+ "▁",
+ "une"
+ ],
+ [
+ "l",
+ "a"
+ ],
+ [
+ "[",
+ "\""
+ ],
+ [
+ "fi",
+ "eld"
+ ],
+ [
+ "f",
+ "ield"
+ ],
+ [
+ "▁U",
+ "S"
+ ],
+ [
+ "▁",
+ "US"
+ ],
+ [
+ "ic",
+ "o"
+ ],
+ [
+ "i",
+ "co"
+ ],
+ [
+ "▁per",
+ "form"
+ ],
+ [
+ "▁perf",
+ "orm"
+ ],
+ [
+ "▁",
+ "perform"
+ ],
+ [
+ "ail",
+ "able"
+ ],
+ [
+ "Con",
+ "fig"
+ ],
+ [
+ "Conf",
+ "ig"
+ ],
+ [
+ "O",
+ "r"
+ ],
+ [
+ "▁mod",
+ "el"
+ ],
+ [
+ "▁mo",
+ "del"
+ ],
+ [
+ "▁mode",
+ "l"
+ ],
+ [
+ "▁",
+ "model"
+ ],
+ [
+ "al",
+ "es"
+ ],
+ [
+ "ale",
+ "s"
+ ],
+ [
+ "a",
+ "les"
+ ],
+ [
+ "▁c",
+ "reate"
+ ],
+ [
+ "▁cre",
+ "ate"
+ ],
+ [
+ "▁creat",
+ "e"
+ ],
+ [
+ "▁",
+ "create"
+ ],
+ [
+ "▁a",
+ "nn"
+ ],
+ [
+ "▁an",
+ "n"
+ ],
+ [
+ "▁",
+ "ann"
+ ],
+ [
+ "an",
+ "ces"
+ ],
+ [
+ "ance",
+ "s"
+ ],
+ [
+ "anc",
+ "es"
+ ],
+ [
+ "I",
+ "L"
+ ],
+ [
+ "in",
+ "ation"
+ ],
+ [
+ "▁I",
+ "m"
+ ],
+ [
+ "▁",
+ "Im"
+ ],
+ [
+ "an",
+ "te"
+ ],
+ [
+ "ant",
+ "e"
+ ],
+ [
+ "a",
+ "nte"
+ ],
+ [
+ "an",
+ "a"
+ ],
+ [
+ "a",
+ "na"
+ ],
+ [
+ "а",
+ "н"
+ ],
+ [
+ "▁t",
+ "old"
+ ],
+ [
+ "▁to",
+ "ld"
+ ],
+ [
+ "con",
+ "fig"
+ ],
+ [
+ "conf",
+ "ig"
+ ],
+ [
+ "\"",
+ "]"
+ ],
+ [
+ "me",
+ "t"
+ ],
+ [
+ "m",
+ "et"
+ ],
+ [
+ "l",
+ "t"
+ ],
+ [
+ "▁t",
+ "ext"
+ ],
+ [
+ "▁te",
+ "xt"
+ ],
+ [
+ "▁tex",
+ "t"
+ ],
+ [
+ "▁",
+ "text"
+ ],
+ [
+ "▁M",
+ "ay"
+ ],
+ [
+ "▁Ma",
+ "y"
+ ],
+ [
+ "▁",
+ "May"
+ ],
+ [
+ "▁o",
+ "rg"
+ ],
+ [
+ "▁or",
+ "g"
+ ],
+ [
+ "▁",
+ "org"
+ ],
+ [
+ "▁p",
+ "ort"
+ ],
+ [
+ "▁po",
+ "rt"
+ ],
+ [
+ "▁por",
+ "t"
+ ],
+ [
+ "▁",
+ "port"
+ ],
+ [
+ "P",
+ "l"
+ ],
+ [
+ "ent",
+ "ly"
+ ],
+ [
+ "▁d",
+ "oor"
+ ],
+ [
+ "▁do",
+ "or"
+ ],
+ [
+ "▁",
+ "door"
+ ],
+ [
+ "U",
+ "S"
+ ],
+ [
+ "▁(",
+ "*"
+ ],
+ [
+ "▁",
+ "(*"
+ ],
+ [
+ "k",
+ "t"
+ ],
+ [
+ "E",
+ "S"
+ ],
+ [
+ "ent",
+ "ial"
+ ],
+ [
+ "enti",
+ "al"
+ ],
+ [
+ "▁is",
+ "s"
+ ],
+ [
+ "▁i",
+ "ss"
+ ],
+ [
+ "▁",
+ "iss"
+ ],
+ [
+ "▁in",
+ "c"
+ ],
+ [
+ "▁i",
+ "nc"
+ ],
+ [
+ "▁",
+ "inc"
+ ],
+ [
+ "No",
+ "de"
+ ],
+ [
+ "N",
+ "ode"
+ ],
+ [
+ "iv",
+ "ely"
+ ],
+ [
+ "ive",
+ "ly"
+ ],
+ [
+ "ivel",
+ "y"
+ ],
+ [
+ "▁as",
+ "ked"
+ ],
+ [
+ "▁ask",
+ "ed"
+ ],
+ [
+ "ir",
+ "t"
+ ],
+ [
+ "i",
+ "rt"
+ ],
+ [
+ "▁T",
+ "e"
+ ],
+ [
+ "▁",
+ "Te"
+ ],
+ [
+ "▁re",
+ "port"
+ ],
+ [
+ "▁rep",
+ "ort"
+ ],
+ [
+ "▁repo",
+ "rt"
+ ],
+ [
+ "▁",
+ "report"
+ ],
+ [
+ "▁c",
+ "hang"
+ ],
+ [
+ "▁ch",
+ "ang"
+ ],
+ [
+ "▁cha",
+ "ng"
+ ],
+ [
+ "ст",
+ "и"
+ ],
+ [
+ "с",
+ "ти"
+ ],
+ [
+ "▁a",
+ "long"
+ ],
+ [
+ "▁al",
+ "ong"
+ ],
+ [
+ "▁ch",
+ "ange"
+ ],
+ [
+ "▁chang",
+ "e"
+ ],
+ [
+ "▁",
+ "change"
+ ],
+ [
+ "Si",
+ "ze"
+ ],
+ [
+ "S",
+ "ize"
+ ],
+ [
+ "▁e",
+ "ver"
+ ],
+ [
+ "▁ev",
+ "er"
+ ],
+ [
+ "▁",
+ "ever"
+ ],
+ [
+ "▁o",
+ "cc"
+ ],
+ [
+ "▁oc",
+ "c"
+ ],
+ [
+ "▁",
+ "occ"
+ ],
+ [
+ "ur",
+ "y"
+ ],
+ [
+ "u",
+ "ry"
+ ],
+ [
+ "▁m",
+ "ind"
+ ],
+ [
+ "▁min",
+ "d"
+ ],
+ [
+ "▁mi",
+ "nd"
+ ],
+ [
+ "▁",
+ "mind"
+ ],
+ [
+ "or",
+ "der"
+ ],
+ [
+ "ord",
+ "er"
+ ],
+ [
+ "po",
+ "int"
+ ],
+ [
+ "p",
+ "oint"
+ ],
+ [
+ "ст",
+ "о"
+ ],
+ [
+ "с",
+ "то"
+ ],
+ [
+ "▁w",
+ "he"
+ ],
+ [
+ "▁wh",
+ "e"
+ ],
+ [
+ "▁",
+ "whe"
+ ],
+ [
+ "▁import",
+ "ant"
+ ],
+ [
+ "▁",
+ "important"
+ ],
+ [
+ "de",
+ "s"
+ ],
+ [
+ "d",
+ "es"
+ ],
+ [
+ "▁N",
+ "ot"
+ ],
+ [
+ "▁No",
+ "t"
+ ],
+ [
+ "▁",
+ "Not"
+ ],
+ [
+ "▁w",
+ "rit"
+ ],
+ [
+ "▁wr",
+ "it"
+ ],
+ [
+ "▁",
+ "writ"
+ ],
+ [
+ "▁e",
+ "yes"
+ ],
+ [
+ "▁ey",
+ "es"
+ ],
+ [
+ "▁eye",
+ "s"
+ ],
+ [
+ "▁d",
+ "esc"
+ ],
+ [
+ "▁de",
+ "sc"
+ ],
+ [
+ "▁des",
+ "c"
+ ],
+ [
+ "▁",
+ "desc"
+ ],
+ [
+ "mo",
+ "st"
+ ],
+ [
+ "mos",
+ "t"
+ ],
+ [
+ "m",
+ "ost"
+ ],
+ [
+ "k",
+ "s"
+ ],
+ [
+ "▁b",
+ "it"
+ ],
+ [
+ "▁bi",
+ "t"
+ ],
+ [
+ "▁",
+ "bit"
+ ],
+ [
+ "▁su",
+ "ccess"
+ ],
+ [
+ "▁suc",
+ "cess"
+ ],
+ [
+ "▁succ",
+ "ess"
+ ],
+ [
+ "▁",
+ "success"
+ ],
+ [
+ "т",
+ "ь"
+ ],
+ [
+ "б",
+ "о"
+ ],
+ [
+ "co",
+ "re"
+ ],
+ [
+ "cor",
+ "e"
+ ],
+ [
+ "c",
+ "ore"
+ ],
+ [
+ "}",
+ "("
+ ],
+ [
+ "▁ar",
+ "ray"
+ ],
+ [
+ "▁arr",
+ "ay"
+ ],
+ [
+ "▁",
+ "array"
+ ],
+ [
+ "li",
+ "n"
+ ],
+ [
+ "l",
+ "in"
+ ],
+ [
+ "li",
+ "sh"
+ ],
+ [
+ "l",
+ "ish"
+ ],
+ [
+ "▁follow",
+ "ing"
+ ],
+ [
+ "Fi",
+ "eld"
+ ],
+ [
+ "F",
+ "ield"
+ ],
+ [
+ "id",
+ "s"
+ ],
+ [
+ "i",
+ "ds"
+ ],
+ [
+ "hi",
+ "ng"
+ ],
+ [
+ "hin",
+ "g"
+ ],
+ [
+ "h",
+ "ing"
+ ],
+ [
+ "▁c",
+ "al"
+ ],
+ [
+ "▁ca",
+ "l"
+ ],
+ [
+ "▁",
+ "cal"
+ ],
+ [
+ "I",
+ "s"
+ ],
+ [
+ "ar",
+ "ing"
+ ],
+ [
+ "ari",
+ "ng"
+ ],
+ [
+ "arin",
+ "g"
+ ],
+ [
+ "a",
+ "ring"
+ ],
+ [
+ "le",
+ "v"
+ ],
+ [
+ "l",
+ "ev"
+ ],
+ [
+ "al",
+ "t"
+ ],
+ [
+ "a",
+ "lt"
+ ],
+ [
+ "C",
+ "H"
+ ],
+ [
+ "▁d",
+ "é"
+ ],
+ [
+ "al",
+ "pha"
+ ],
+ [
+ "alph",
+ "a"
+ ],
+ [
+ "▁f",
+ "our"
+ ],
+ [
+ "▁fo",
+ "ur"
+ ],
+ [
+ "▁fou",
+ "r"
+ ],
+ [
+ "▁",
+ "four"
+ ],
+ [
+ "▁l",
+ "aw"
+ ],
+ [
+ "▁la",
+ "w"
+ ],
+ [
+ "▁",
+ "law"
+ ],
+ [
+ "▁с",
+ "е"
+ ],
+ [
+ "▁",
+ "се"
+ ],
+ [
+ "ir",
+ "on"
+ ],
+ [
+ "iro",
+ "n"
+ ],
+ [
+ "i",
+ "ron"
+ ],
+ [
+ "▁d",
+ "isc"
+ ],
+ [
+ "▁dis",
+ "c"
+ ],
+ [
+ "▁di",
+ "sc"
+ ],
+ [
+ "с",
+ "е"
+ ],
+ [
+ "ke",
+ "n"
+ ],
+ [
+ "k",
+ "en"
+ ],
+ [
+ "no",
+ "de"
+ ],
+ [
+ "nod",
+ "e"
+ ],
+ [
+ "n",
+ "ode"
+ ],
+ [
+ "▁P",
+ "ar"
+ ],
+ [
+ "▁Pa",
+ "r"
+ ],
+ [
+ "▁",
+ "Par"
+ ],
+ [
+ "▁E",
+ "ng"
+ ],
+ [
+ "▁En",
+ "g"
+ ],
+ [
+ "▁",
+ "Eng"
+ ],
+ [
+ "▁m",
+ "ove"
+ ],
+ [
+ "▁mov",
+ "e"
+ ],
+ [
+ "▁mo",
+ "ve"
+ ],
+ [
+ "▁",
+ "move"
+ ],
+ [
+ "▁L",
+ "icense"
+ ],
+ [
+ "▁Lic",
+ "ense"
+ ],
+ [
+ "▁",
+ "License"
+ ],
+ [
+ "cu",
+ "l"
+ ],
+ [
+ "c",
+ "ul"
+ ],
+ [
+ "ion",
+ "e"
+ ],
+ [
+ "io",
+ "ne"
+ ],
+ [
+ "i",
+ "one"
+ ],
+ [
+ ")",
+ "$"
+ ],
+ [
+ "▁t",
+ "w"
+ ],
+ [
+ "▁",
+ "tw"
+ ],
+ [
+ "W",
+ "e"
+ ],
+ [
+ "se",
+ "l"
+ ],
+ [
+ "s",
+ "el"
+ ],
+ [
+ "▁W",
+ "ith"
+ ],
+ [
+ "▁Wi",
+ "th"
+ ],
+ [
+ "▁",
+ "With"
+ ],
+ [
+ "▁on",
+ "ce"
+ ],
+ [
+ "▁",
+ "once"
+ ],
+ [
+ "Serv",
+ "ice"
+ ],
+ [
+ "S",
+ "ervice"
+ ],
+ [
+ "bo",
+ "l"
+ ],
+ [
+ "b",
+ "ol"
+ ],
+ [
+ "ur",
+ "ed"
+ ],
+ [
+ "ure",
+ "d"
+ ],
+ [
+ "u",
+ "red"
+ ],
+ [
+ "id",
+ "a"
+ ],
+ [
+ "i",
+ "da"
+ ],
+ [
+ "▁Q",
+ "u"
+ ],
+ [
+ "▁",
+ "Qu"
+ ],
+ [
+ "▁g",
+ "row"
+ ],
+ [
+ "▁gr",
+ "ow"
+ ],
+ [
+ "▁gro",
+ "w"
+ ],
+ [
+ "▁",
+ "grow"
+ ],
+ [
+ "▁c",
+ "onne"
+ ],
+ [
+ "▁con",
+ "ne"
+ ],
+ [
+ "▁conn",
+ "e"
+ ],
+ [
+ "▁",
+ "conne"
+ ],
+ [
+ "E",
+ "X"
+ ],
+ [
+ "▁h",
+ "tt"
+ ],
+ [
+ "▁",
+ "htt"
+ ],
+ [
+ "▁}",
+ ";"
+ ],
+ [
+ "▁",
+ "};"
+ ],
+ [
+ "▁w",
+ "alk"
+ ],
+ [
+ "▁wal",
+ "k"
+ ],
+ [
+ "▁",
+ "walk"
+ ],
+ [
+ "▁in",
+ "it"
+ ],
+ [
+ "▁i",
+ "nit"
+ ],
+ [
+ "▁",
+ "init"
+ ],
+ [
+ "na",
+ "l"
+ ],
+ [
+ "n",
+ "al"
+ ],
+ [
+ "en",
+ "der"
+ ],
+ [
+ "end",
+ "er"
+ ],
+ [
+ "ende",
+ "r"
+ ],
+ [
+ "e",
+ "nder"
+ ],
+ [
+ "cri",
+ "ption"
+ ],
+ [
+ "cript",
+ "ion"
+ ],
+ [
+ "mb",
+ "er"
+ ],
+ [
+ "m",
+ "ber"
+ ],
+ [
+ "le",
+ "cted"
+ ],
+ [
+ "lect",
+ "ed"
+ ],
+ [
+ "p",
+ "o"
+ ],
+ [
+ "▁n",
+ "il"
+ ],
+ [
+ "▁ni",
+ "l"
+ ],
+ [
+ "▁",
+ "nil"
+ ],
+ [
+ "▁p",
+ "rob"
+ ],
+ [
+ "▁pro",
+ "b"
+ ],
+ [
+ "▁pr",
+ "ob"
+ ],
+ [
+ "▁",
+ "prob"
+ ],
+ [
+ "ч",
+ "и"
+ ],
+ [
+ "▁S",
+ "te"
+ ],
+ [
+ "▁St",
+ "e"
+ ],
+ [
+ "▁",
+ "Ste"
+ ],
+ [
+ "is",
+ "on"
+ ],
+ [
+ "iso",
+ "n"
+ ],
+ [
+ "i",
+ "son"
+ ],
+ [
+ "an",
+ "ds"
+ ],
+ [
+ "and",
+ "s"
+ ],
+ [
+ "os",
+ "ed"
+ ],
+ [
+ "ose",
+ "d"
+ ],
+ [
+ "o",
+ "sed"
+ ],
+ [
+ "ж",
+ "е"
+ ],
+ [
+ "▁H",
+ "is"
+ ],
+ [
+ "▁Hi",
+ "s"
+ ],
+ [
+ "▁",
+ "His"
+ ],
+ [
+ "ü",
+ "r"
+ ],
+ [
+ "Ma",
+ "n"
+ ],
+ [
+ "M",
+ "an"
+ ],
+ [
+ "El",
+ "ement"
+ ],
+ [
+ "Elem",
+ "ent"
+ ],
+ [
+ "E",
+ "lement"
+ ],
+ [
+ "▁a",
+ "ble"
+ ],
+ [
+ "▁ab",
+ "le"
+ ],
+ [
+ "▁",
+ "able"
+ ],
+ [
+ "In",
+ "dex"
+ ],
+ [
+ "Ind",
+ "ex"
+ ],
+ [
+ "se",
+ "arch"
+ ],
+ [
+ "s",
+ "earch"
+ ],
+ [
+ "▁m",
+ "ag"
+ ],
+ [
+ "▁ma",
+ "g"
+ ],
+ [
+ "▁",
+ "mag"
+ ],
+ [
+ "а",
+ "р"
+ ],
+ [
+ "▁c",
+ "ourse"
+ ],
+ [
+ "▁cour",
+ "se"
+ ],
+ [
+ "▁cours",
+ "e"
+ ],
+ [
+ "▁",
+ "course"
+ ],
+ [
+ "▁C",
+ "ar"
+ ],
+ [
+ "▁Ca",
+ "r"
+ ],
+ [
+ "▁",
+ "Car"
+ ],
+ [
+ "▁e",
+ "xp"
+ ],
+ [
+ "▁ex",
+ "p"
+ ],
+ [
+ "▁",
+ "exp"
+ ],
+ [
+ "ap",
+ "h"
+ ],
+ [
+ "a",
+ "ph"
+ ],
+ [
+ "▁m",
+ "it"
+ ],
+ [
+ "▁mi",
+ "t"
+ ],
+ [
+ "▁",
+ "mit"
+ ],
+ [
+ "▁does",
+ "n"
+ ],
+ [
+ "▁def",
+ "ault"
+ ],
+ [
+ "▁",
+ "default"
+ ],
+ [
+ "/",
+ ">"
+ ],
+ [
+ "ai",
+ "m"
+ ],
+ [
+ "a",
+ "im"
+ ],
+ [
+ "▁s",
+ "ervice"
+ ],
+ [
+ "▁serv",
+ "ice"
+ ],
+ [
+ "▁",
+ "service"
+ ],
+ [
+ "▁with",
+ "in"
+ ],
+ [
+ "an",
+ "gu"
+ ],
+ [
+ "ang",
+ "u"
+ ],
+ [
+ "▁",
+ "Д"
+ ],
+ [
+ "uf",
+ "fer"
+ ],
+ [
+ "uff",
+ "er"
+ ],
+ [
+ "A",
+ "G"
+ ],
+ [
+ "▁D",
+ "o"
+ ],
+ [
+ "▁",
+ "Do"
+ ],
+ [
+ "▁in",
+ "cre"
+ ],
+ [
+ "▁inc",
+ "re"
+ ],
+ [
+ "▁under",
+ "stand"
+ ],
+ [
+ "}",
+ "^"
+ ],
+ [
+ "▁look",
+ "ed"
+ ],
+ [
+ "▁lo",
+ "oked"
+ ],
+ [
+ "ge",
+ "n"
+ ],
+ [
+ "g",
+ "en"
+ ],
+ [
+ "ail",
+ "ed"
+ ],
+ [
+ "ai",
+ "led"
+ ],
+ [
+ "a",
+ "iled"
+ ],
+ [
+ "▁",
+ "е"
+ ],
+ [
+ "ay",
+ "er"
+ ],
+ [
+ "aye",
+ "r"
+ ],
+ [
+ "a",
+ "yer"
+ ],
+ [
+ "▁O",
+ "ne"
+ ],
+ [
+ "▁On",
+ "e"
+ ],
+ [
+ "▁",
+ "One"
+ ],
+ [
+ "▁b",
+ "as"
+ ],
+ [
+ "▁ba",
+ "s"
+ ],
+ [
+ "▁",
+ "bas"
+ ],
+ [
+ "▁j",
+ "ob"
+ ],
+ [
+ "▁jo",
+ "b"
+ ],
+ [
+ "▁",
+ "job"
+ ],
+ [
+ "m",
+ "u"
+ ],
+ [
+ "bu",
+ "t"
+ ],
+ [
+ "b",
+ "ut"
+ ],
+ [
+ "el",
+ "ta"
+ ],
+ [
+ "elt",
+ "a"
+ ],
+ [
+ "▁Ch",
+ "rist"
+ ],
+ [
+ "▁Chris",
+ "t"
+ ],
+ [
+ "▁",
+ "Christ"
+ ],
+ [
+ "ur",
+ "ation"
+ ],
+ [
+ "▁re",
+ "cord"
+ ],
+ [
+ "▁rec",
+ "ord"
+ ],
+ [
+ "▁",
+ "record"
+ ],
+ [
+ "▁Un",
+ "ivers"
+ ],
+ [
+ "▁",
+ "Univers"
+ ],
+ [
+ "iv",
+ "id"
+ ],
+ [
+ "ivi",
+ "d"
+ ],
+ [
+ "i",
+ "vid"
+ ],
+ [
+ "val",
+ "id"
+ ],
+ [
+ "▁",
+ "Р"
+ ],
+ [
+ "▁h",
+ "old"
+ ],
+ [
+ "▁hol",
+ "d"
+ ],
+ [
+ "▁ho",
+ "ld"
+ ],
+ [
+ "▁",
+ "hold"
+ ],
+ [
+ "▁t",
+ "able"
+ ],
+ [
+ "▁tab",
+ "le"
+ ],
+ [
+ "▁ta",
+ "ble"
+ ],
+ [
+ "▁",
+ "table"
+ ],
+ [
+ "on",
+ "es"
+ ],
+ [
+ "one",
+ "s"
+ ],
+ [
+ "o",
+ "nes"
+ ],
+ [
+ "lin",
+ "k"
+ ],
+ [
+ "l",
+ "ink"
+ ],
+ [
+ "▁G",
+ "e"
+ ],
+ [
+ "▁",
+ "Ge"
+ ],
+ [
+ "▁of",
+ "fer"
+ ],
+ [
+ "▁off",
+ "er"
+ ],
+ [
+ "st",
+ "er"
+ ],
+ [
+ "ste",
+ "r"
+ ],
+ [
+ "s",
+ "ter"
+ ],
+ [
+ "For",
+ "m"
+ ],
+ [
+ "F",
+ "orm"
+ ],
+ [
+ "=",
+ "{"
+ ],
+ [
+ "▁н",
+ "е"
+ ],
+ [
+ "▁",
+ "не"
+ ],
+ [
+ "st",
+ "ance"
+ ],
+ [
+ "stan",
+ "ce"
+ ],
+ [
+ "▁g",
+ "overn"
+ ],
+ [
+ "▁go",
+ "vern"
+ ],
+ [
+ "▁gover",
+ "n"
+ ],
+ [
+ "▁",
+ "govern"
+ ],
+ [
+ "▁te",
+ "chn"
+ ],
+ [
+ "▁tech",
+ "n"
+ ],
+ [
+ "▁",
+ "techn"
+ ],
+ [
+ "▁p",
+ "rim"
+ ],
+ [
+ "▁pr",
+ "im"
+ ],
+ [
+ "▁pri",
+ "m"
+ ],
+ [
+ "▁",
+ "prim"
+ ],
+ [
+ "*",
+ "."
+ ],
+ [
+ "ch",
+ "o"
+ ],
+ [
+ "c",
+ "ho"
+ ],
+ [
+ "ma",
+ "x"
+ ],
+ [
+ "m",
+ "ax"
+ ],
+ [
+ "▁f",
+ "ore"
+ ],
+ [
+ "▁for",
+ "e"
+ ],
+ [
+ "▁fo",
+ "re"
+ ],
+ [
+ "▁",
+ "fore"
+ ],
+ [
+ "▁C",
+ "an"
+ ],
+ [
+ "▁Ca",
+ "n"
+ ],
+ [
+ "▁",
+ "Can"
+ ],
+ [
+ "▁pol",
+ "it"
+ ],
+ [
+ "▁po",
+ "lit"
+ ],
+ [
+ "▁",
+ "polit"
+ ],
+ [
+ "or",
+ "ies"
+ ],
+ [
+ "ori",
+ "es"
+ ],
+ [
+ "orie",
+ "s"
+ ],
+ [
+ "o",
+ "ries"
+ ],
+ [
+ "▁t",
+ "imes"
+ ],
+ [
+ "▁time",
+ "s"
+ ],
+ [
+ "▁tim",
+ "es"
+ ],
+ [
+ "▁ti",
+ "mes"
+ ],
+ [
+ "▁",
+ "times"
+ ],
+ [
+ "▁d",
+ "ans"
+ ],
+ [
+ "▁da",
+ "ns"
+ ],
+ [
+ "▁dan",
+ "s"
+ ],
+ [
+ "▁a",
+ "ir"
+ ],
+ [
+ "▁ai",
+ "r"
+ ],
+ [
+ "▁",
+ "air"
+ ],
+ [
+ "▁any",
+ "thing"
+ ],
+ [
+ "▁s",
+ "ever"
+ ],
+ [
+ "▁se",
+ "ver"
+ ],
+ [
+ "ac",
+ "y"
+ ],
+ [
+ "a",
+ "cy"
+ ],
+ [
+ "}",
+ "_"
+ ],
+ [
+ "H",
+ "e"
+ ],
+ [
+ "▁l",
+ "east"
+ ],
+ [
+ "▁le",
+ "ast"
+ ],
+ [
+ "ip",
+ "s"
+ ],
+ [
+ "i",
+ "ps"
+ ],
+ [
+ "EN",
+ "T"
+ ],
+ [
+ "E",
+ "NT"
+ ],
+ [
+ "d",
+ "o"
+ ],
+ [
+ "▁о",
+ "т"
+ ],
+ [
+ "▁",
+ "от"
+ ],
+ [
+ "▁c",
+ "ost"
+ ],
+ [
+ "▁co",
+ "st"
+ ],
+ [
+ "▁cos",
+ "t"
+ ],
+ [
+ "▁",
+ "cost"
+ ],
+ [
+ ".",
+ "”"
+ ],
+ [
+ "▁child",
+ "ren"
+ ],
+ [
+ "▁",
+ "children"
+ ],
+ [
+ "ab",
+ "ility"
+ ],
+ [
+ "abil",
+ "ity"
+ ],
+ [
+ "Bu",
+ "t"
+ ],
+ [
+ "B",
+ "ut"
+ ],
+ [
+ "▁p",
+ "ath"
+ ],
+ [
+ "▁pat",
+ "h"
+ ],
+ [
+ "▁pa",
+ "th"
+ ],
+ [
+ "▁",
+ "path"
+ ],
+ [
+ "res",
+ "ult"
+ ],
+ [
+ "ac",
+ "ter"
+ ],
+ [
+ "act",
+ "er"
+ ],
+ [
+ "▁e",
+ "lement"
+ ],
+ [
+ "▁el",
+ "ement"
+ ],
+ [
+ "▁ele",
+ "ment"
+ ],
+ [
+ "▁elem",
+ "ent"
+ ],
+ [
+ "▁",
+ "element"
+ ],
+ [
+ "e",
+ "e"
+ ],
+ [
+ "▁w",
+ "ait"
+ ],
+ [
+ "▁wa",
+ "it"
+ ],
+ [
+ "▁",
+ "wait"
+ ],
+ [
+ "▁m",
+ "oney"
+ ],
+ [
+ "▁mon",
+ "ey"
+ ],
+ [
+ "▁mo",
+ "ney"
+ ],
+ [
+ "Ma",
+ "p"
+ ],
+ [
+ "M",
+ "ap"
+ ],
+ [
+ "t",
+ "d"
+ ],
+ [
+ "oi",
+ "n"
+ ],
+ [
+ "o",
+ "in"
+ ],
+ [
+ "iv",
+ "ing"
+ ],
+ [
+ "ivi",
+ "ng"
+ ],
+ [
+ "i",
+ "ving"
+ ],
+ [
+ "ic",
+ "ht"
+ ],
+ [
+ "ich",
+ "t"
+ ],
+ [
+ "i",
+ "cht"
+ ],
+ [
+ "ic",
+ "y"
+ ],
+ [
+ "i",
+ "cy"
+ ],
+ [
+ "sc",
+ "h"
+ ],
+ [
+ "s",
+ "ch"
+ ],
+ [
+ "st",
+ "e"
+ ],
+ [
+ "s",
+ "te"
+ ],
+ [
+ "д",
+ "у"
+ ],
+ [
+ "or",
+ "ed"
+ ],
+ [
+ "ore",
+ "d"
+ ],
+ [
+ "o",
+ "red"
+ ],
+ [
+ "ou",
+ "d"
+ ],
+ [
+ "o",
+ "ud"
+ ],
+ [
+ "il",
+ "le"
+ ],
+ [
+ "ill",
+ "e"
+ ],
+ [
+ "i",
+ "lle"
+ ],
+ [
+ "is",
+ "ed"
+ ],
+ [
+ "ise",
+ "d"
+ ],
+ [
+ "i",
+ "sed"
+ ],
+ [
+ "pl",
+ "ication"
+ ],
+ [
+ "plic",
+ "ation"
+ ],
+ [
+ "▁c",
+ "ustom"
+ ],
+ [
+ "▁cust",
+ "om"
+ ],
+ [
+ "▁",
+ "custom"
+ ],
+ [
+ "▁h",
+ "aving"
+ ],
+ [
+ "▁ha",
+ "ving"
+ ],
+ [
+ "▁hav",
+ "ing"
+ ],
+ [
+ "pon",
+ "ent"
+ ],
+ [
+ "po",
+ "nent"
+ ],
+ [
+ "▁B",
+ "y"
+ ],
+ [
+ "▁",
+ "By"
+ ],
+ [
+ "ul",
+ "es"
+ ],
+ [
+ "ule",
+ "s"
+ ],
+ [
+ "u",
+ "les"
+ ],
+ [
+ "ue",
+ "d"
+ ],
+ [
+ "u",
+ "ed"
+ ],
+ [
+ "at",
+ "ter"
+ ],
+ [
+ "att",
+ "er"
+ ],
+ [
+ "atte",
+ "r"
+ ],
+ [
+ "An",
+ "d"
+ ],
+ [
+ "A",
+ "nd"
+ ],
+ [
+ "it",
+ "ive"
+ ],
+ [
+ "iti",
+ "ve"
+ ],
+ [
+ "De",
+ "f"
+ ],
+ [
+ "D",
+ "ef"
+ ],
+ [
+ "▁m",
+ "oment"
+ ],
+ [
+ "▁mom",
+ "ent"
+ ],
+ [
+ "▁mo",
+ "ment"
+ ],
+ [
+ "▁",
+ "moment"
+ ],
+ [
+ "at",
+ "erial"
+ ],
+ [
+ "ate",
+ "rial"
+ ],
+ [
+ "ater",
+ "ial"
+ ],
+ [
+ "Cl",
+ "ass"
+ ],
+ [
+ "C",
+ "lass"
+ ],
+ [
+ "og",
+ "raph"
+ ],
+ [
+ "ograp",
+ "h"
+ ],
+ [
+ "o",
+ "graph"
+ ],
+ [
+ "ik",
+ "e"
+ ],
+ [
+ "i",
+ "ke"
+ ],
+ [
+ "▁l",
+ "arge"
+ ],
+ [
+ "▁larg",
+ "e"
+ ],
+ [
+ "▁",
+ "large"
+ ],
+ [
+ "▁#",
+ "###"
+ ],
+ [
+ "▁##",
+ "##"
+ ],
+ [
+ "▁###",
+ "#"
+ ],
+ [
+ "▁",
+ "####"
+ ],
+ [
+ "▁e",
+ "ither"
+ ],
+ [
+ "du",
+ "ct"
+ ],
+ [
+ "duc",
+ "t"
+ ],
+ [
+ "d",
+ "uct"
+ ],
+ [
+ "▁T",
+ "hen"
+ ],
+ [
+ "▁The",
+ "n"
+ ],
+ [
+ "▁Th",
+ "en"
+ ],
+ [
+ "▁",
+ "Then"
+ ],
+ [
+ "▁G",
+ "u"
+ ],
+ [
+ "▁",
+ "Gu"
+ ],
+ [
+ "ole",
+ "an"
+ ],
+ [
+ "o",
+ "lean"
+ ],
+ [
+ "pe",
+ "rt"
+ ],
+ [
+ "per",
+ "t"
+ ],
+ [
+ "p",
+ "ert"
+ ],
+ [
+ "▁G",
+ "et"
+ ],
+ [
+ "▁Ge",
+ "t"
+ ],
+ [
+ "▁",
+ "Get"
+ ],
+ [
+ "▁A",
+ "b"
+ ],
+ [
+ "▁",
+ "Ab"
+ ],
+ [
+ "▁sh",
+ "ort"
+ ],
+ [
+ "▁",
+ "short"
+ ],
+ [
+ "O",
+ "n"
+ ],
+ [
+ "im",
+ "ent"
+ ],
+ [
+ "ime",
+ "nt"
+ ],
+ [
+ "imen",
+ "t"
+ ],
+ [
+ "i",
+ "ment"
+ ],
+ [
+ "▁pro",
+ "ject"
+ ],
+ [
+ "▁",
+ "project"
+ ],
+ [
+ "cri",
+ "pt"
+ ],
+ [
+ "cr",
+ "ipt"
+ ],
+ [
+ "c",
+ "ript"
+ ],
+ [
+ "▁incl",
+ "uding"
+ ],
+ [
+ "▁includ",
+ "ing"
+ ],
+ [
+ "▁inclu",
+ "ding"
+ ],
+ [
+ "▁",
+ "including"
+ ],
+ [
+ "ни",
+ "я"
+ ],
+ [
+ "▁m",
+ "aking"
+ ],
+ [
+ "▁ma",
+ "king"
+ ],
+ [
+ "▁",
+ "making"
+ ],
+ [
+ "▁some",
+ "one"
+ ],
+ [
+ "▁F",
+ "l"
+ ],
+ [
+ "▁",
+ "Fl"
+ ],
+ [
+ "▁s",
+ "at"
+ ],
+ [
+ "▁sa",
+ "t"
+ ],
+ [
+ "▁",
+ "sat"
+ ],
+ [
+ "▁comp",
+ "any"
+ ],
+ [
+ "▁compan",
+ "y"
+ ],
+ [
+ "▁",
+ "company"
+ ],
+ [
+ "oc",
+ "us"
+ ],
+ [
+ "p",
+ "u"
+ ],
+ [
+ "▁G",
+ "od"
+ ],
+ [
+ "▁Go",
+ "d"
+ ],
+ [
+ "▁",
+ "God"
+ ],
+ [
+ "if",
+ "ication"
+ ],
+ [
+ "ific",
+ "ation"
+ ],
+ [
+ "N",
+ "o"
+ ],
+ [
+ "▁s",
+ "n"
+ ],
+ [
+ "▁",
+ "sn"
+ ],
+ [
+ "an",
+ "o"
+ ],
+ [
+ "a",
+ "no"
+ ],
+ [
+ "g",
+ "a"
+ ],
+ [
+ "▁a",
+ "u"
+ ],
+ [
+ "▁",
+ "au"
+ ],
+ [
+ "▁c",
+ "ou"
+ ],
+ [
+ "▁co",
+ "u"
+ ],
+ [
+ "▁",
+ "cou"
+ ],
+ [
+ "á",
+ "s"
+ ],
+ [
+ "en",
+ "ded"
+ ],
+ [
+ "end",
+ "ed"
+ ],
+ [
+ "ende",
+ "d"
+ ],
+ [
+ "т",
+ "у"
+ ],
+ [
+ "ob",
+ "er"
+ ],
+ [
+ "obe",
+ "r"
+ ],
+ [
+ "o",
+ "ber"
+ ],
+ [
+ "▁n",
+ "othing"
+ ],
+ [
+ "▁not",
+ "hing"
+ ],
+ [
+ "▁no",
+ "thing"
+ ],
+ [
+ "▁n",
+ "et"
+ ],
+ [
+ "▁ne",
+ "t"
+ ],
+ [
+ "▁",
+ "net"
+ ],
+ [
+ "▁p",
+ "ot"
+ ],
+ [
+ "▁po",
+ "t"
+ ],
+ [
+ "▁",
+ "pot"
+ ],
+ [
+ "▁t",
+ "yp"
+ ],
+ [
+ "▁ty",
+ "p"
+ ],
+ [
+ "▁",
+ "typ"
+ ],
+ [
+ "▁it",
+ "em"
+ ],
+ [
+ "▁i",
+ "tem"
+ ],
+ [
+ "▁",
+ "item"
+ ],
+ [
+ "re",
+ "w"
+ ],
+ [
+ "r",
+ "ew"
+ ],
+ [
+ "At",
+ "t"
+ ],
+ [
+ "A",
+ "tt"
+ ],
+ [
+ "▁you",
+ "ng"
+ ],
+ [
+ "▁yo",
+ "ung"
+ ],
+ [
+ "}",
+ "\r"
+ ],
+ [
+ "nd",
+ "er"
+ ],
+ [
+ "nde",
+ "r"
+ ],
+ [
+ "n",
+ "der"
+ ],
+ [
+ "st",
+ "art"
+ ],
+ [
+ "sta",
+ "rt"
+ ],
+ [
+ "star",
+ "t"
+ ],
+ [
+ "▁S",
+ "c"
+ ],
+ [
+ "▁",
+ "Sc"
+ ],
+ [
+ "*",
+ ")"
+ ],
+ [
+ "▁e",
+ "nc"
+ ],
+ [
+ "▁en",
+ "c"
+ ],
+ [
+ "▁",
+ "enc"
+ ],
+ [
+ "▁w",
+ "omen"
+ ],
+ [
+ "▁wom",
+ "en"
+ ],
+ [
+ "▁wo",
+ "men"
+ ],
+ [
+ "▁look",
+ "ing"
+ ],
+ [
+ "▁lo",
+ "oking"
+ ],
+ [
+ "▁",
+ "looking"
+ ],
+ [
+ "▁р",
+ "о"
+ ],
+ [
+ "▁",
+ "ро"
+ ],
+ [
+ "▁he",
+ "alth"
+ ],
+ [
+ "▁heal",
+ "th"
+ ],
+ [
+ "▁",
+ "health"
+ ],
+ [
+ "Pat",
+ "h"
+ ],
+ [
+ "P",
+ "ath"
+ ],
+ [
+ "▁A",
+ "fter"
+ ],
+ [
+ "▁Af",
+ "ter"
+ ],
+ [
+ "▁",
+ "After"
+ ],
+ [
+ "▁m",
+ "ult"
+ ],
+ [
+ "▁mu",
+ "lt"
+ ],
+ [
+ "▁mul",
+ "t"
+ ],
+ [
+ "▁",
+ "mult"
+ ],
+ [
+ "▁{",
+ "\\"
+ ],
+ [
+ "▁",
+ "{\\"
+ ],
+ [
+ "▁l",
+ "and"
+ ],
+ [
+ "▁la",
+ "nd"
+ ],
+ [
+ "▁lan",
+ "d"
+ ],
+ [
+ "▁",
+ "land"
+ ],
+ [
+ "or",
+ "ld"
+ ],
+ [
+ "▁D",
+ "es"
+ ],
+ [
+ "▁De",
+ "s"
+ ],
+ [
+ "▁",
+ "Des"
+ ],
+ [
+ "▁e",
+ "ng"
+ ],
+ [
+ "▁en",
+ "g"
+ ],
+ [
+ "▁",
+ "eng"
+ ],
+ [
+ "in",
+ "put"
+ ],
+ [
+ "▁P",
+ "ol"
+ ],
+ [
+ "▁Po",
+ "l"
+ ],
+ [
+ "▁",
+ "Pol"
+ ],
+ [
+ "\"",
+ "\""
+ ],
+ [
+ "Co",
+ "de"
+ ],
+ [
+ "C",
+ "ode"
+ ],
+ [
+ "▁s",
+ "upp"
+ ],
+ [
+ "▁su",
+ "pp"
+ ],
+ [
+ "▁sup",
+ "p"
+ ],
+ [
+ "▁",
+ "supp"
+ ],
+ [
+ "ain",
+ "er"
+ ],
+ [
+ "ai",
+ "ner"
+ ],
+ [
+ "aine",
+ "r"
+ ],
+ [
+ "a",
+ "iner"
+ ],
+ [
+ "he",
+ "ck"
+ ],
+ [
+ "▁m",
+ "or"
+ ],
+ [
+ "▁mo",
+ "r"
+ ],
+ [
+ "▁",
+ "mor"
+ ],
+ [
+ "▁m",
+ "ill"
+ ],
+ [
+ "▁mil",
+ "l"
+ ],
+ [
+ "▁mi",
+ "ll"
+ ],
+ [
+ "▁",
+ "mill"
+ ],
+ [
+ "▁a",
+ "w"
+ ],
+ [
+ "▁",
+ "aw"
+ ],
+ [
+ "f",
+ "s"
+ ],
+ [
+ "▁do",
+ "ing"
+ ],
+ [
+ "ting",
+ "s"
+ ],
+ [
+ "t",
+ "ings"
+ ],
+ [
+ "ad",
+ "es"
+ ],
+ [
+ "ade",
+ "s"
+ ],
+ [
+ "a",
+ "des"
+ ],
+ [
+ "▁to",
+ "get"
+ ],
+ [
+ "▁c",
+ "ertain"
+ ],
+ [
+ "▁cert",
+ "ain"
+ ],
+ [
+ "▁cer",
+ "tain"
+ ],
+ [
+ "▁t",
+ "ogether"
+ ],
+ [
+ "▁toget",
+ "her"
+ ],
+ [
+ "C",
+ "E"
+ ],
+ [
+ "ide",
+ "o"
+ ],
+ [
+ "▁Amer",
+ "ican"
+ ],
+ [
+ "▁America",
+ "n"
+ ],
+ [
+ "▁",
+ "American"
+ ],
+ [
+ "on",
+ "y"
+ ],
+ [
+ "o",
+ "ny"
+ ],
+ [
+ "id",
+ "d"
+ ],
+ [
+ "i",
+ "dd"
+ ],
+ [
+ "I",
+ "I"
+ ],
+ [
+ "ge",
+ "d"
+ ],
+ [
+ "g",
+ "ed"
+ ],
+ [
+ "ab",
+ "les"
+ ],
+ [
+ "able",
+ "s"
+ ],
+ [
+ "abl",
+ "es"
+ ],
+ [
+ "a",
+ "bles"
+ ],
+ [
+ "▁ide",
+ "nt"
+ ],
+ [
+ "▁id",
+ "ent"
+ ],
+ [
+ "▁",
+ "ident"
+ ],
+ [
+ "io",
+ "d"
+ ],
+ [
+ "i",
+ "od"
+ ],
+ [
+ "▁p",
+ "arent"
+ ],
+ [
+ "▁par",
+ "ent"
+ ],
+ [
+ "▁pa",
+ "rent"
+ ],
+ [
+ "▁pare",
+ "nt"
+ ],
+ [
+ "▁",
+ "parent"
+ ],
+ [
+ "F",
+ "or"
+ ],
+ [
+ "amb",
+ "da"
+ ],
+ [
+ "an",
+ "do"
+ ],
+ [
+ "and",
+ "o"
+ ],
+ [
+ "=",
+ "\\"
+ ],
+ [
+ "ag",
+ "ed"
+ ],
+ [
+ "age",
+ "d"
+ ],
+ [
+ "a",
+ "ged"
+ ],
+ [
+ "en",
+ "ding"
+ ],
+ [
+ "end",
+ "ing"
+ ],
+ [
+ "In",
+ "t"
+ ],
+ [
+ "I",
+ "nt"
+ ],
+ [
+ "▁poss",
+ "ible"
+ ],
+ [
+ "▁",
+ "possible"
+ ],
+ [
+ "▁с",
+ "о"
+ ],
+ [
+ "▁",
+ "со"
+ ],
+ [
+ "iv",
+ "ity"
+ ],
+ [
+ "ivi",
+ "ty"
+ ],
+ [
+ "nu",
+ "m"
+ ],
+ [
+ "n",
+ "um"
+ ],
+ [
+ "r",
+ "t"
+ ],
+ [
+ "aj",
+ "or"
+ ],
+ [
+ "ajo",
+ "r"
+ ],
+ [
+ "a",
+ "jor"
+ ],
+ [
+ "cre",
+ "ate"
+ ],
+ [
+ "creat",
+ "e"
+ ],
+ [
+ "c",
+ "reate"
+ ],
+ [
+ "ri",
+ "de"
+ ],
+ [
+ "rid",
+ "e"
+ ],
+ [
+ "r",
+ "ide"
+ ],
+ [
+ "▁k",
+ "new"
+ ],
+ [
+ "▁kn",
+ "ew"
+ ],
+ [
+ "▁kne",
+ "w"
+ ],
+ [
+ "bi",
+ "t"
+ ],
+ [
+ "b",
+ "it"
+ ],
+ [
+ "it",
+ "ional"
+ ],
+ [
+ "ition",
+ "al"
+ ],
+ [
+ "iti",
+ "onal"
+ ],
+ [
+ "▁l",
+ "ik"
+ ],
+ [
+ "▁li",
+ "k"
+ ],
+ [
+ "▁",
+ "lik"
+ ],
+ [
+ "▁H",
+ "er"
+ ],
+ [
+ "▁He",
+ "r"
+ ],
+ [
+ "▁",
+ "Her"
+ ],
+ [
+ "ens",
+ "ion"
+ ],
+ [
+ "\"",
+ "."
+ ],
+ [
+ "ot",
+ "o"
+ ],
+ [
+ "o",
+ "to"
+ ],
+ [
+ "▁ex",
+ "ist"
+ ],
+ [
+ "▁",
+ "exist"
+ ],
+ [
+ "ak",
+ "en"
+ ],
+ [
+ "ake",
+ "n"
+ ],
+ [
+ "a",
+ "ken"
+ ],
+ [
+ "▁act",
+ "ually"
+ ],
+ [
+ "▁actual",
+ "ly"
+ ],
+ [
+ "c",
+ "a"
+ ],
+ [
+ "▁",
+ "Г"
+ ],
+ [
+ "х",
+ "о"
+ ],
+ [
+ "in",
+ "n"
+ ],
+ [
+ "i",
+ "nn"
+ ],
+ [
+ "Al",
+ "l"
+ ],
+ [
+ "A",
+ "ll"
+ ],
+ [
+ "bu",
+ "f"
+ ],
+ [
+ "b",
+ "uf"
+ ],
+ [
+ "▁M",
+ "e"
+ ],
+ [
+ "▁",
+ "Me"
+ ],
+ [
+ "▁s",
+ "een"
+ ],
+ [
+ "▁se",
+ "en"
+ ],
+ [
+ "▁see",
+ "n"
+ ],
+ [
+ "▁",
+ "seen"
+ ],
+ [
+ "op",
+ "s"
+ ],
+ [
+ "o",
+ "ps"
+ ],
+ [
+ "No",
+ "t"
+ ],
+ [
+ "N",
+ "ot"
+ ],
+ [
+ "▁cont",
+ "rol"
+ ],
+ [
+ "▁contr",
+ "ol"
+ ],
+ [
+ "▁contro",
+ "l"
+ ],
+ [
+ "▁",
+ "control"
+ ],
+ [
+ "▁res",
+ "pon"
+ ],
+ [
+ "▁resp",
+ "on"
+ ],
+ [
+ "▁",
+ "respon"
+ ],
+ [
+ "}",
+ ";"
+ ],
+ [
+ "il",
+ "t"
+ ],
+ [
+ "i",
+ "lt"
+ ],
+ [
+ "is",
+ "k"
+ ],
+ [
+ "i",
+ "sk"
+ ],
+ [
+ "▁b",
+ "ad"
+ ],
+ [
+ "▁ba",
+ "d"
+ ],
+ [
+ "▁",
+ "bad"
+ ],
+ [
+ "▁o",
+ "ften"
+ ],
+ [
+ "▁of",
+ "ten"
+ ],
+ [
+ "▁p",
+ "ast"
+ ],
+ [
+ "▁pas",
+ "t"
+ ],
+ [
+ "▁pa",
+ "st"
+ ],
+ [
+ "ap",
+ "er"
+ ],
+ [
+ "ape",
+ "r"
+ ],
+ [
+ "a",
+ "per"
+ ],
+ [
+ "▁re",
+ "ason"
+ ],
+ [
+ "▁",
+ "reason"
+ ],
+ [
+ "et",
+ "ers"
+ ],
+ [
+ "eter",
+ "s"
+ ],
+ [
+ "ete",
+ "rs"
+ ],
+ [
+ "e",
+ "ters"
+ ],
+ [
+ "▁w",
+ "anted"
+ ],
+ [
+ "▁want",
+ "ed"
+ ],
+ [
+ "ur",
+ "a"
+ ],
+ [
+ "u",
+ "ra"
+ ],
+ [
+ "ta",
+ "ble"
+ ],
+ [
+ "tab",
+ "le"
+ ],
+ [
+ "t",
+ "able"
+ ],
+ [
+ "or",
+ "mal"
+ ],
+ [
+ "orm",
+ "al"
+ ],
+ [
+ "wid",
+ "th"
+ ],
+ [
+ "w",
+ "idth"
+ ],
+ [
+ "г",
+ "а"
+ ],
+ [
+ "pt",
+ "r"
+ ],
+ [
+ "p",
+ "tr"
+ ],
+ [
+ "▁d",
+ "est"
+ ],
+ [
+ "▁de",
+ "st"
+ ],
+ [
+ "▁des",
+ "t"
+ ],
+ [
+ "▁",
+ "dest"
+ ],
+ [
+ "▁de",
+ "sign"
+ ],
+ [
+ "▁des",
+ "ign"
+ ],
+ [
+ "▁",
+ "design"
+ ],
+ [
+ "▁s",
+ "ound"
+ ],
+ [
+ "▁so",
+ "und"
+ ],
+ [
+ "▁sou",
+ "nd"
+ ],
+ [
+ "▁",
+ "sound"
+ ],
+ [
+ "▁p",
+ "lan"
+ ],
+ [
+ "▁pl",
+ "an"
+ ],
+ [
+ "▁",
+ "plan"
+ ],
+ [
+ "▁b",
+ "ase"
+ ],
+ [
+ "▁bas",
+ "e"
+ ],
+ [
+ "▁ba",
+ "se"
+ ],
+ [
+ "▁",
+ "base"
+ ],
+ [
+ "ha",
+ "nd"
+ ],
+ [
+ "han",
+ "d"
+ ],
+ [
+ "h",
+ "and"
+ ],
+ [
+ "g",
+ "s"
+ ],
+ [
+ "▁s",
+ "ays"
+ ],
+ [
+ "▁sa",
+ "ys"
+ ],
+ [
+ "▁say",
+ "s"
+ ],
+ [
+ "fun",
+ "ction"
+ ],
+ [
+ "f",
+ "unction"
+ ],
+ [
+ "▁t",
+ "ri"
+ ],
+ [
+ "▁tr",
+ "i"
+ ],
+ [
+ "▁",
+ "tri"
+ ],
+ [
+ "m",
+ "t"
+ ],
+ [
+ "▁in",
+ "vest"
+ ],
+ [
+ "▁inv",
+ "est"
+ ],
+ [
+ "▁av",
+ "ailable"
+ ],
+ [
+ "▁",
+ "available"
+ ],
+ [
+ "ay",
+ "out"
+ ],
+ [
+ "a",
+ "yout"
+ ],
+ [
+ "▁o",
+ "ch"
+ ],
+ [
+ "▁oc",
+ "h"
+ ],
+ [
+ "▁",
+ "och"
+ ],
+ [
+ "▁l",
+ "as"
+ ],
+ [
+ "▁la",
+ "s"
+ ],
+ [
+ "▁",
+ "las"
+ ],
+ [
+ "il",
+ "led"
+ ],
+ [
+ "ill",
+ "ed"
+ ],
+ [
+ "ille",
+ "d"
+ ],
+ [
+ "V",
+ "al"
+ ],
+ [
+ "▁",
+ "ф"
+ ],
+ [
+ "ie",
+ "ty"
+ ],
+ [
+ "iet",
+ "y"
+ ],
+ [
+ "i",
+ "ety"
+ ],
+ [
+ "mo",
+ "n"
+ ],
+ [
+ "m",
+ "on"
+ ],
+ [
+ "Ha",
+ "nd"
+ ],
+ [
+ "H",
+ "and"
+ ],
+ [
+ "F",
+ "r"
+ ],
+ [
+ "ia",
+ "m"
+ ],
+ [
+ "i",
+ "am"
+ ],
+ [
+ "pa",
+ "ce"
+ ],
+ [
+ "p",
+ "ace"
+ ],
+ [
+ "▁O",
+ "b"
+ ],
+ [
+ "▁",
+ "Ob"
+ ],
+ [
+ "▁p",
+ "ara"
+ ],
+ [
+ "▁par",
+ "a"
+ ],
+ [
+ "▁pa",
+ "ra"
+ ],
+ [
+ "▁",
+ "para"
+ ],
+ [
+ "▁me",
+ "et"
+ ],
+ [
+ "▁s",
+ "um"
+ ],
+ [
+ "▁su",
+ "m"
+ ],
+ [
+ "▁",
+ "sum"
+ ],
+ [
+ "M",
+ "essage"
+ ],
+ [
+ "ic",
+ "i"
+ ],
+ [
+ "i",
+ "ci"
+ ],
+ [
+ "▁k",
+ "nown"
+ ],
+ [
+ "▁kn",
+ "own"
+ ],
+ [
+ "▁know",
+ "n"
+ ],
+ [
+ "▁",
+ "known"
+ ],
+ [
+ "▁g",
+ "en"
+ ],
+ [
+ "▁ge",
+ "n"
+ ],
+ [
+ "▁",
+ "gen"
+ ],
+ [
+ "am",
+ "ma"
+ ],
+ [
+ "amm",
+ "a"
+ ],
+ [
+ "a",
+ "mma"
+ ],
+ [
+ "ar",
+ "r"
+ ],
+ [
+ "a",
+ "rr"
+ ],
+ [
+ "▁t",
+ "re"
+ ],
+ [
+ "▁tr",
+ "e"
+ ],
+ [
+ "▁",
+ "tre"
+ ],
+ [
+ "ok",
+ "e"
+ ],
+ [
+ "o",
+ "ke"
+ ],
+ [
+ "ut",
+ "h"
+ ],
+ [
+ "u",
+ "th"
+ ],
+ [
+ "~",
+ "\\"
+ ],
+ [
+ "▁exper",
+ "ience"
+ ],
+ [
+ "▁experi",
+ "ence"
+ ],
+ [
+ "ic",
+ "le"
+ ],
+ [
+ "icl",
+ "e"
+ ],
+ [
+ "i",
+ "cle"
+ ],
+ [
+ "▁I",
+ "l"
+ ],
+ [
+ "▁",
+ "Il"
+ ],
+ [
+ "▁s",
+ "ent"
+ ],
+ [
+ "▁se",
+ "nt"
+ ],
+ [
+ "▁sen",
+ "t"
+ ],
+ [
+ "▁",
+ "sent"
+ ],
+ [
+ "▁o",
+ "thers"
+ ],
+ [
+ "▁other",
+ "s"
+ ],
+ [
+ "▁",
+ "others"
+ ],
+ [
+ "▁s",
+ "oft"
+ ],
+ [
+ "▁so",
+ "ft"
+ ],
+ [
+ "▁",
+ "soft"
+ ],
+ [
+ "I",
+ "P"
+ ],
+ [
+ "▁m",
+ "ax"
+ ],
+ [
+ "▁ma",
+ "x"
+ ],
+ [
+ "▁",
+ "max"
+ ],
+ [
+ "ba",
+ "ll"
+ ],
+ [
+ "bal",
+ "l"
+ ],
+ [
+ "b",
+ "all"
+ ],
+ [
+ "▁mark",
+ "et"
+ ],
+ [
+ "▁mar",
+ "ket"
+ ],
+ [
+ "▁",
+ "market"
+ ],
+ [
+ "▁p",
+ "our"
+ ],
+ [
+ "▁po",
+ "ur"
+ ],
+ [
+ "▁pou",
+ "r"
+ ],
+ [
+ "pr",
+ "ession"
+ ],
+ [
+ "press",
+ "ion"
+ ],
+ [
+ "p",
+ "ression"
+ ],
+ [
+ "ep",
+ "s"
+ ],
+ [
+ "e",
+ "ps"
+ ],
+ [
+ "▁s",
+ "aw"
+ ],
+ [
+ "▁sa",
+ "w"
+ ],
+ [
+ "▁a",
+ "cross"
+ ],
+ [
+ "▁ac",
+ "ross"
+ ],
+ [
+ "▁S",
+ "u"
+ ],
+ [
+ "▁",
+ "Su"
+ ],
+ [
+ "O",
+ "ver"
+ ],
+ [
+ "ни",
+ "е"
+ ],
+ [
+ "ul",
+ "ation"
+ ],
+ [
+ "u",
+ "lation"
+ ],
+ [
+ "▁R",
+ "eg"
+ ],
+ [
+ "▁Re",
+ "g"
+ ],
+ [
+ "▁",
+ "Reg"
+ ],
+ [
+ "▁+",
+ "="
+ ],
+ [
+ "▁",
+ "+="
+ ],
+ [
+ "bo",
+ "dy"
+ ],
+ [
+ "b",
+ "ody"
+ ],
+ [
+ ")",
+ "\\"
+ ],
+ [
+ "▁pr",
+ "int"
+ ],
+ [
+ "▁pri",
+ "nt"
+ ],
+ [
+ "▁prin",
+ "t"
+ ],
+ [
+ "▁",
+ "print"
+ ],
+ [
+ "▁п",
+ "ри"
+ ],
+ [
+ "▁пр",
+ "и"
+ ],
+ [
+ "▁",
+ "при"
+ ],
+ [
+ "d",
+ "b"
+ ],
+ [
+ "our",
+ "ces"
+ ],
+ [
+ "ource",
+ "s"
+ ],
+ [
+ "ward",
+ "s"
+ ],
+ [
+ "war",
+ "ds"
+ ],
+ [
+ "w",
+ "ards"
+ ],
+ [
+ "▁bl",
+ "ack"
+ ],
+ [
+ "▁",
+ "black"
+ ],
+ [
+ "с",
+ "о"
+ ],
+ [
+ "il",
+ "i"
+ ],
+ [
+ "i",
+ "li"
+ ],
+ [
+ "▁E",
+ "d"
+ ],
+ [
+ "▁",
+ "Ed"
+ ],
+ [
+ "▁com",
+ "plet"
+ ],
+ [
+ "▁comp",
+ "let"
+ ],
+ [
+ "▁compl",
+ "et"
+ ],
+ [
+ "▁s",
+ "ingle"
+ ],
+ [
+ "▁sing",
+ "le"
+ ],
+ [
+ "▁sin",
+ "gle"
+ ],
+ [
+ "▁",
+ "single"
+ ],
+ [
+ "▁I",
+ "N"
+ ],
+ [
+ "▁",
+ "IN"
+ ],
+ [
+ "ac",
+ "hed"
+ ],
+ [
+ "ach",
+ "ed"
+ ],
+ [
+ "ache",
+ "d"
+ ],
+ [
+ "a",
+ "ched"
+ ],
+ [
+ "b",
+ "t"
+ ],
+ [
+ "▁c",
+ "ode"
+ ],
+ [
+ "▁co",
+ "de"
+ ],
+ [
+ "▁cod",
+ "e"
+ ],
+ [
+ "▁",
+ "code"
+ ],
+ [
+ "▁b",
+ "ool"
+ ],
+ [
+ "▁bo",
+ "ol"
+ ],
+ [
+ "▁",
+ "bool"
+ ],
+ [
+ "▁a",
+ "rea"
+ ],
+ [
+ "▁are",
+ "a"
+ ],
+ [
+ "▁ar",
+ "ea"
+ ],
+ [
+ "▁",
+ "area"
+ ],
+ [
+ "▁re",
+ "quire"
+ ],
+ [
+ "▁requ",
+ "ire"
+ ],
+ [
+ "▁",
+ "require"
+ ],
+ [
+ "▁pro",
+ "blem"
+ ],
+ [
+ "▁proble",
+ "m"
+ ],
+ [
+ "▁prob",
+ "lem"
+ ],
+ [
+ "ac",
+ "ed"
+ ],
+ [
+ "ace",
+ "d"
+ ],
+ [
+ "a",
+ "ced"
+ ],
+ [
+ "Eq",
+ "u"
+ ],
+ [
+ "E",
+ "qu"
+ ],
+ [
+ "▁con",
+ "fig"
+ ],
+ [
+ "▁conf",
+ "ig"
+ ],
+ [
+ "▁",
+ "config"
+ ],
+ [
+ "ve",
+ "c"
+ ],
+ [
+ "v",
+ "ec"
+ ],
+ [
+ "ne",
+ "y"
+ ],
+ [
+ "n",
+ "ey"
+ ],
+ [
+ "c",
+ "y"
+ ],
+ [
+ "A",
+ "l"
+ ],
+ [
+ "▁acc",
+ "ount"
+ ],
+ [
+ "▁ac",
+ "count"
+ ],
+ [
+ "▁",
+ "account"
+ ],
+ [
+ "ym",
+ "bol"
+ ],
+ [
+ "▁s",
+ "te"
+ ],
+ [
+ "▁st",
+ "e"
+ ],
+ [
+ "▁",
+ "ste"
+ ],
+ [
+ "ge",
+ "s"
+ ],
+ [
+ "g",
+ "es"
+ ],
+ [
+ "Ar",
+ "ray"
+ ],
+ [
+ "Arr",
+ "ay"
+ ],
+ [
+ "em",
+ "pl"
+ ],
+ [
+ "emp",
+ "l"
+ ],
+ [
+ "con",
+ "text"
+ ],
+ [
+ "cont",
+ "ext"
+ ],
+ [
+ "De",
+ "s"
+ ],
+ [
+ "D",
+ "es"
+ ],
+ [
+ "Res",
+ "ult"
+ ],
+ [
+ "ec",
+ "ut"
+ ],
+ [
+ "e",
+ "cut"
+ ],
+ [
+ "▁t",
+ "arget"
+ ],
+ [
+ "▁tar",
+ "get"
+ ],
+ [
+ "▁",
+ "target"
+ ],
+ [
+ "▁get",
+ "ting"
+ ],
+ [
+ "\"",
+ "/>"
+ ],
+ [
+ "og",
+ "le"
+ ],
+ [
+ "o",
+ "gle"
+ ],
+ [
+ "▁him",
+ "self"
+ ],
+ [
+ "▁was",
+ "n"
+ ],
+ [
+ "▁wa",
+ "sn"
+ ],
+ [
+ "▁b",
+ "lock"
+ ],
+ [
+ "▁bl",
+ "ock"
+ ],
+ [
+ "▁blo",
+ "ck"
+ ],
+ [
+ "▁",
+ "block"
+ ],
+ [
+ "▁a",
+ "nt"
+ ],
+ [
+ "▁an",
+ "t"
+ ],
+ [
+ "▁",
+ "ant"
+ ],
+ [
+ "▁Y",
+ "ork"
+ ],
+ [
+ "▁be",
+ "come"
+ ],
+ [
+ "▁bec",
+ "ome"
+ ],
+ [
+ "if",
+ "f"
+ ],
+ [
+ "i",
+ "ff"
+ ],
+ [
+ "port",
+ "s"
+ ],
+ [
+ "por",
+ "ts"
+ ],
+ [
+ "p",
+ "orts"
+ ],
+ [
+ "re",
+ "ate"
+ ],
+ [
+ "reat",
+ "e"
+ ],
+ [
+ "rea",
+ "te"
+ ],
+ [
+ "=",
+ "'"
+ ],
+ [
+ "c",
+ "d"
+ ],
+ [
+ "loc",
+ "ation"
+ ],
+ [
+ "l",
+ "ocation"
+ ],
+ [
+ "е",
+ "т"
+ ],
+ [
+ "▁a",
+ "ccess"
+ ],
+ [
+ "▁acc",
+ "ess"
+ ],
+ [
+ "▁ac",
+ "cess"
+ ],
+ [
+ "▁",
+ "access"
+ ],
+ [
+ "gr",
+ "ess"
+ ],
+ [
+ "gre",
+ "ss"
+ ],
+ [
+ "gres",
+ "s"
+ ],
+ [
+ "g",
+ "ress"
+ ],
+ [
+ "ro",
+ "s"
+ ],
+ [
+ "r",
+ "os"
+ ],
+ [
+ "U",
+ "p"
+ ],
+ [
+ "▁work",
+ "ing"
+ ],
+ [
+ "▁wor",
+ "king"
+ ],
+ [
+ "▁",
+ "working"
+ ],
+ [
+ "▁A",
+ "m"
+ ],
+ [
+ "▁",
+ "Am"
+ ],
+ [
+ "iq",
+ "u"
+ ],
+ [
+ "i",
+ "qu"
+ ],
+ [
+ "ce",
+ "r"
+ ],
+ [
+ "c",
+ "er"
+ ],
+ [
+ "▁(",
+ "("
+ ],
+ [
+ "▁",
+ "(("
+ ],
+ [
+ "▁P",
+ "er"
+ ],
+ [
+ "▁Pe",
+ "r"
+ ],
+ [
+ "▁",
+ "Per"
+ ],
+ [
+ "▁f",
+ "unc"
+ ],
+ [
+ "▁fun",
+ "c"
+ ],
+ [
+ "▁fu",
+ "nc"
+ ],
+ [
+ "▁",
+ "func"
+ ],
+ [
+ "▁g",
+ "irl"
+ ],
+ [
+ "▁gi",
+ "rl"
+ ],
+ [
+ "▁gir",
+ "l"
+ ],
+ [
+ "▁",
+ "girl"
+ ],
+ [
+ "▁ab",
+ "ove"
+ ],
+ [
+ "pe",
+ "n"
+ ],
+ [
+ "p",
+ "en"
+ ],
+ [
+ "п",
+ "и"
+ ],
+ [
+ "id",
+ "o"
+ ],
+ [
+ "i",
+ "do"
+ ],
+ [
+ "▁v",
+ "ersion"
+ ],
+ [
+ "▁vers",
+ "ion"
+ ],
+ [
+ "▁",
+ "version"
+ ],
+ [
+ "T",
+ "Y"
+ ],
+ [
+ "▁",
+ ";"
+ ],
+ [
+ "ma",
+ "ry"
+ ],
+ [
+ "mar",
+ "y"
+ ],
+ [
+ "m",
+ "ary"
+ ],
+ [
+ "ab",
+ "led"
+ ],
+ [
+ "able",
+ "d"
+ ],
+ [
+ "abl",
+ "ed"
+ ],
+ [
+ "a",
+ "bled"
+ ],
+ [
+ "an",
+ "nel"
+ ],
+ [
+ "ann",
+ "el"
+ ],
+ [
+ "anne",
+ "l"
+ ],
+ [
+ "▁ex",
+ "ample"
+ ],
+ [
+ "▁exam",
+ "ple"
+ ],
+ [
+ "▁",
+ "example"
+ ],
+ [
+ "▁con",
+ "text"
+ ],
+ [
+ "▁cont",
+ "ext"
+ ],
+ [
+ "▁",
+ "context"
+ ],
+ [
+ "O",
+ "P"
+ ],
+ [
+ "▁re",
+ "d"
+ ],
+ [
+ "▁r",
+ "ed"
+ ],
+ [
+ "▁",
+ "red"
+ ],
+ [
+ "▁c",
+ "ir"
+ ],
+ [
+ "▁ci",
+ "r"
+ ],
+ [
+ "▁",
+ "cir"
+ ],
+ [
+ "s",
+ "m"
+ ],
+ [
+ "Lo",
+ "g"
+ ],
+ [
+ "L",
+ "og"
+ ],
+ [
+ "▁s",
+ "pace"
+ ],
+ [
+ "▁sp",
+ "ace"
+ ],
+ [
+ "▁",
+ "space"
+ ],
+ [
+ "▁f",
+ "ut"
+ ],
+ [
+ "▁fu",
+ "t"
+ ],
+ [
+ "▁G",
+ "ener"
+ ],
+ [
+ "▁Ge",
+ "ner"
+ ],
+ [
+ "▁Gen",
+ "er"
+ ],
+ [
+ "▁Gene",
+ "r"
+ ],
+ [
+ "▁",
+ "Gener"
+ ],
+ [
+ "il",
+ "ls"
+ ],
+ [
+ "ill",
+ "s"
+ ],
+ [
+ "▁d",
+ "ri"
+ ],
+ [
+ "▁dr",
+ "i"
+ ],
+ [
+ "_",
+ "."
+ ],
+ [
+ "▁f",
+ "elt"
+ ],
+ [
+ "▁fe",
+ "lt"
+ ],
+ [
+ "▁fel",
+ "t"
+ ],
+ [
+ "▁o",
+ "ffic"
+ ],
+ [
+ "▁of",
+ "fic"
+ ],
+ [
+ "▁off",
+ "ic"
+ ],
+ [
+ "▁=",
+ "=="
+ ],
+ [
+ "▁==",
+ "="
+ ],
+ [
+ "▁",
+ "==="
+ ],
+ [
+ "i",
+ "i"
+ ],
+ [
+ "▁start",
+ "ed"
+ ],
+ [
+ "▁star",
+ "ted"
+ ],
+ [
+ "▁",
+ "Т"
+ ],
+ [
+ "▁}",
+ ");"
+ ],
+ [
+ "▁})",
+ ";"
+ ],
+ [
+ "▁",
+ "});"
+ ],
+ [
+ "j",
+ "s"
+ ],
+ [
+ "▁fr",
+ "ont"
+ ],
+ [
+ "▁fro",
+ "nt"
+ ],
+ [
+ "▁",
+ "front"
+ ],
+ [
+ "▁al",
+ "most"
+ ],
+ [
+ "ir",
+ "m"
+ ],
+ [
+ "i",
+ "rm"
+ ],
+ [
+ "!",
+ "\""
+ ],
+ [
+ "sign",
+ "ed"
+ ],
+ [
+ "sig",
+ "ned"
+ ],
+ [
+ "s",
+ "igned"
+ ],
+ [
+ "▁y",
+ "et"
+ ],
+ [
+ "▁ye",
+ "t"
+ ],
+ [
+ "▁t",
+ "rad"
+ ],
+ [
+ "▁tr",
+ "ad"
+ ],
+ [
+ "▁tra",
+ "d"
+ ],
+ [
+ "ient",
+ "s"
+ ],
+ [
+ "ien",
+ "ts"
+ ],
+ [
+ "i",
+ "ents"
+ ],
+ [
+ "am",
+ "a"
+ ],
+ [
+ "a",
+ "ma"
+ ],
+ [
+ "▁in",
+ "put"
+ ],
+ [
+ "▁",
+ "input"
+ ],
+ [
+ "li",
+ "m"
+ ],
+ [
+ "l",
+ "im"
+ ],
+ [
+ "п",
+ "а"
+ ],
+ [
+ "▁к",
+ "а"
+ ],
+ [
+ "▁",
+ "ка"
+ ],
+ [
+ "▁c",
+ "amp"
+ ],
+ [
+ "▁cam",
+ "p"
+ ],
+ [
+ "▁ca",
+ "mp"
+ ],
+ [
+ "▁",
+ "camp"
+ ],
+ [
+ "ib",
+ "r"
+ ],
+ [
+ "i",
+ "br"
+ ],
+ [
+ "fe",
+ "ct"
+ ],
+ [
+ "f",
+ "ect"
+ ],
+ [
+ "un",
+ "t"
+ ],
+ [
+ "u",
+ "nt"
+ ],
+ [
+ "▁h",
+ "alf"
+ ],
+ [
+ "▁hal",
+ "f"
+ ],
+ [
+ "▁",
+ "half"
+ ],
+ [
+ "▁c",
+ "over"
+ ],
+ [
+ "▁co",
+ "ver"
+ ],
+ [
+ "▁cov",
+ "er"
+ ],
+ [
+ "▁",
+ "cover"
+ ],
+ [
+ "angu",
+ "age"
+ ],
+ [
+ "▁b",
+ "en"
+ ],
+ [
+ "▁be",
+ "n"
+ ],
+ [
+ "▁",
+ "ben"
+ ],
+ [
+ "h",
+ "a"
+ ],
+ [
+ "▁d",
+ "iff"
+ ],
+ [
+ "▁di",
+ "ff"
+ ],
+ [
+ "▁dif",
+ "f"
+ ],
+ [
+ "▁",
+ "diff"
+ ],
+ [
+ "_",
+ "\\"
+ ],
+ [
+ "▁о",
+ "б"
+ ],
+ [
+ "▁",
+ "об"
+ ],
+ [
+ "]",
+ ")"
+ ],
+ [
+ "od",
+ "es"
+ ],
+ [
+ "ode",
+ "s"
+ ],
+ [
+ "o",
+ "des"
+ ],
+ [
+ "he",
+ "l"
+ ],
+ [
+ "h",
+ "el"
+ ],
+ [
+ "io",
+ "s"
+ ],
+ [
+ "i",
+ "os"
+ ],
+ [
+ "▁",
+ "О"
+ ],
+ [
+ "▁m",
+ "ot"
+ ],
+ [
+ "▁mo",
+ "t"
+ ],
+ [
+ "▁",
+ "mot"
+ ],
+ [
+ "▁s",
+ "ocial"
+ ],
+ [
+ "▁so",
+ "cial"
+ ],
+ [
+ "▁soc",
+ "ial"
+ ],
+ [
+ "▁soci",
+ "al"
+ ],
+ [
+ "▁",
+ "social"
+ ],
+ [
+ "////",
+ "////"
+ ],
+ [
+ "▁s",
+ "tre"
+ ],
+ [
+ "▁st",
+ "re"
+ ],
+ [
+ "▁str",
+ "e"
+ ],
+ [
+ "▁",
+ "stre"
+ ],
+ [
+ "gr",
+ "ound"
+ ],
+ [
+ "gro",
+ "und"
+ ],
+ [
+ "g",
+ "round"
+ ],
+ [
+ "і",
+ "в"
+ ],
+ [
+ "ob",
+ "ject"
+ ],
+ [
+ "obj",
+ "ect"
+ ],
+ [
+ "pl",
+ "es"
+ ],
+ [
+ "ple",
+ "s"
+ ],
+ [
+ "p",
+ "les"
+ ],
+ [
+ "re",
+ "ed"
+ ],
+ [
+ "ree",
+ "d"
+ ],
+ [
+ "r",
+ "eed"
+ ],
+ [
+ "▁e",
+ "en"
+ ],
+ [
+ "▁",
+ "een"
+ ],
+ [
+ "▁b",
+ "ased"
+ ],
+ [
+ "▁bas",
+ "ed"
+ ],
+ [
+ "▁base",
+ "d"
+ ],
+ [
+ "▁ba",
+ "sed"
+ ],
+ [
+ "▁",
+ "based"
+ ],
+ [
+ "▁r",
+ "ange"
+ ],
+ [
+ "▁ran",
+ "ge"
+ ],
+ [
+ "▁rang",
+ "e"
+ ],
+ [
+ "▁",
+ "range"
+ ],
+ [
+ "A",
+ "n"
+ ],
+ [
+ "ur",
+ "g"
+ ],
+ [
+ "u",
+ "rg"
+ ],
+ [
+ "▁le",
+ "arn"
+ ],
+ [
+ "▁lear",
+ "n"
+ ],
+ [
+ "▁",
+ "learn"
+ ],
+ [
+ "▁e",
+ "xc"
+ ],
+ [
+ "▁ex",
+ "c"
+ ],
+ [
+ "▁",
+ "exc"
+ ],
+ [
+ "▁im",
+ "p"
+ ],
+ [
+ "▁i",
+ "mp"
+ ],
+ [
+ "▁",
+ "imp"
+ ],
+ [
+ "▁me",
+ "ans"
+ ],
+ [
+ "▁mean",
+ "s"
+ ],
+ [
+ "▁w",
+ "ur"
+ ],
+ [
+ "en",
+ "ds"
+ ],
+ [
+ "end",
+ "s"
+ ],
+ [
+ "vo",
+ "id"
+ ],
+ [
+ "v",
+ "oid"
+ ],
+ [
+ "▁s",
+ "td"
+ ],
+ [
+ "▁st",
+ "d"
+ ],
+ [
+ "▁",
+ "std"
+ ],
+ [
+ "▁part",
+ "icular"
+ ],
+ [
+ "▁partic",
+ "ular"
+ ],
+ [
+ "▁particul",
+ "ar"
+ ],
+ [
+ "▁parti",
+ "cular"
+ ],
+ [
+ "j",
+ "a"
+ ],
+ [
+ "▁s",
+ "ource"
+ ],
+ [
+ "▁sour",
+ "ce"
+ ],
+ [
+ "▁",
+ "source"
+ ],
+ [
+ "def",
+ "ault"
+ ],
+ [
+ "p",
+ "y"
+ ],
+ [
+ "▁a",
+ "ls"
+ ],
+ [
+ "▁al",
+ "s"
+ ],
+ [
+ "▁",
+ "als"
+ ],
+ [
+ "sc",
+ "ri"
+ ],
+ [
+ "scr",
+ "i"
+ ],
+ [
+ "s",
+ "cri"
+ ],
+ [
+ "st",
+ "atus"
+ ],
+ [
+ "stat",
+ "us"
+ ],
+ [
+ "▁st",
+ "ory"
+ ],
+ [
+ "▁stor",
+ "y"
+ ],
+ [
+ "▁sto",
+ "ry"
+ ],
+ [
+ "▁",
+ "story"
+ ],
+ [
+ "▁b",
+ "egin"
+ ],
+ [
+ "▁be",
+ "gin"
+ ],
+ [
+ "▁beg",
+ "in"
+ ],
+ [
+ "▁",
+ "begin"
+ ],
+ [
+ "▁pos",
+ "ition"
+ ],
+ [
+ "▁posit",
+ "ion"
+ ],
+ [
+ "▁",
+ "position"
+ ],
+ [
+ "▁spec",
+ "ial"
+ ],
+ [
+ "▁spe",
+ "cial"
+ ],
+ [
+ "▁",
+ "special"
+ ],
+ [
+ "ph",
+ "p"
+ ],
+ [
+ "p",
+ "hp"
+ ],
+ [
+ "▁b",
+ "ar"
+ ],
+ [
+ "▁ba",
+ "r"
+ ],
+ [
+ "▁",
+ "bar"
+ ],
+ [
+ "▁p",
+ "ract"
+ ],
+ [
+ "▁pr",
+ "act"
+ ],
+ [
+ "▁pra",
+ "ct"
+ ],
+ [
+ "▁prac",
+ "t"
+ ],
+ [
+ "cal",
+ "l"
+ ],
+ [
+ "ca",
+ "ll"
+ ],
+ [
+ "c",
+ "all"
+ ],
+ [
+ "▁d",
+ "as"
+ ],
+ [
+ "▁da",
+ "s"
+ ],
+ [
+ "▁",
+ "das"
+ ],
+ [
+ "▁r",
+ "ad"
+ ],
+ [
+ "▁ra",
+ "d"
+ ],
+ [
+ "▁",
+ "rad"
+ ],
+ [
+ "▁cl",
+ "ose"
+ ],
+ [
+ "▁clos",
+ "e"
+ ],
+ [
+ "▁clo",
+ "se"
+ ],
+ [
+ "▁",
+ "close"
+ ],
+ [
+ "ww",
+ "w"
+ ],
+ [
+ "w",
+ "ww"
+ ],
+ [
+ "ер",
+ "е"
+ ],
+ [
+ "е",
+ "ре"
+ ],
+ [
+ "g",
+ "u"
+ ],
+ [
+ "▁E",
+ "r"
+ ],
+ [
+ "▁",
+ "Er"
+ ],
+ [
+ "▁d",
+ "om"
+ ],
+ [
+ "▁do",
+ "m"
+ ],
+ [
+ "▁",
+ "dom"
+ ],
+ [
+ "A",
+ "M"
+ ],
+ [
+ "▁b",
+ "ed"
+ ],
+ [
+ "▁be",
+ "d"
+ ],
+ [
+ "▁",
+ "bed"
+ ],
+ [
+ "▁sever",
+ "al"
+ ],
+ [
+ "au",
+ "l"
+ ],
+ [
+ "a",
+ "ul"
+ ],
+ [
+ "bo",
+ "x"
+ ],
+ [
+ "b",
+ "ox"
+ ],
+ [
+ "▁l",
+ "ow"
+ ],
+ [
+ "▁lo",
+ "w"
+ ],
+ [
+ "▁",
+ "low"
+ ],
+ [
+ "pa",
+ "ck"
+ ],
+ [
+ "p",
+ "ack"
+ ],
+ [
+ "Re",
+ "g"
+ ],
+ [
+ "R",
+ "eg"
+ ],
+ [
+ "O",
+ "f"
+ ],
+ [
+ "at",
+ "ures"
+ ],
+ [
+ "ature",
+ "s"
+ ],
+ [
+ "atur",
+ "es"
+ ],
+ [
+ "atu",
+ "res"
+ ],
+ [
+ "é",
+ "n"
+ ],
+ [
+ "ed",
+ "er"
+ ],
+ [
+ "ede",
+ "r"
+ ],
+ [
+ "e",
+ "der"
+ ],
+ [
+ "uild",
+ "er"
+ ],
+ [
+ "ca",
+ "st"
+ ],
+ [
+ "cas",
+ "t"
+ ],
+ [
+ "c",
+ "ast"
+ ],
+ [
+ "con",
+ "om"
+ ],
+ [
+ "co",
+ "nom"
+ ],
+ [
+ "c",
+ "onom"
+ ],
+ [
+ "ra",
+ "ft"
+ ],
+ [
+ "raf",
+ "t"
+ ],
+ [
+ "r",
+ "aft"
+ ],
+ [
+ "▁m",
+ "akes"
+ ],
+ [
+ "▁make",
+ "s"
+ ],
+ [
+ "▁ma",
+ "kes"
+ ],
+ [
+ "Lo",
+ "c"
+ ],
+ [
+ "L",
+ "oc"
+ ],
+ [
+ "ht",
+ "tp"
+ ],
+ [
+ "htt",
+ "p"
+ ],
+ [
+ "h",
+ "ttp"
+ ],
+ [
+ "▁a",
+ "bs"
+ ],
+ [
+ "▁ab",
+ "s"
+ ],
+ [
+ "▁",
+ "abs"
+ ],
+ [
+ "re",
+ "sh"
+ ],
+ [
+ "res",
+ "h"
+ ],
+ [
+ "r",
+ "esh"
+ ],
+ [
+ "▁W",
+ "ill"
+ ],
+ [
+ "▁Wil",
+ "l"
+ ],
+ [
+ "▁Wi",
+ "ll"
+ ],
+ [
+ "▁",
+ "Will"
+ ],
+ [
+ "bre",
+ "ak"
+ ],
+ [
+ "b",
+ "reak"
+ ],
+ [
+ "▁o",
+ "ptions"
+ ],
+ [
+ "▁opt",
+ "ions"
+ ],
+ [
+ "▁option",
+ "s"
+ ],
+ [
+ "▁",
+ "options"
+ ],
+ [
+ "fo",
+ "rt"
+ ],
+ [
+ "for",
+ "t"
+ ],
+ [
+ "f",
+ "ort"
+ ],
+ [
+ "▁и",
+ "з"
+ ],
+ [
+ "▁",
+ "из"
+ ],
+ [
+ "▁a",
+ "nal"
+ ],
+ [
+ "▁an",
+ "al"
+ ],
+ [
+ "▁",
+ "anal"
+ ],
+ [
+ "▁e",
+ "nv"
+ ],
+ [
+ "▁en",
+ "v"
+ ],
+ [
+ "▁",
+ "env"
+ ],
+ [
+ "(",
+ "{"
+ ],
+ [
+ "ev",
+ "ent"
+ ],
+ [
+ "even",
+ "t"
+ ],
+ [
+ "eve",
+ "nt"
+ ],
+ [
+ "e",
+ "vent"
+ ],
+ [
+ "▁p",
+ "age"
+ ],
+ [
+ "▁pa",
+ "ge"
+ ],
+ [
+ "▁pag",
+ "e"
+ ],
+ [
+ "▁",
+ "page"
+ ],
+ [
+ "ter",
+ "nal"
+ ],
+ [
+ "tern",
+ "al"
+ ],
+ [
+ "▁d",
+ "istribut"
+ ],
+ [
+ "▁dist",
+ "ribut"
+ ],
+ [
+ "▁f",
+ "ood"
+ ],
+ [
+ "▁fo",
+ "od"
+ ],
+ [
+ "▁foo",
+ "d"
+ ],
+ [
+ "▁",
+ "food"
+ ],
+ [
+ "che",
+ "ck"
+ ],
+ [
+ "c",
+ "heck"
+ ],
+ [
+ "C",
+ "K"
+ ],
+ [
+ "▁в",
+ "о"
+ ],
+ [
+ "▁",
+ "во"
+ ],
+ [
+ "as",
+ "sert"
+ ],
+ [
+ "ass",
+ "ert"
+ ],
+ [
+ "asse",
+ "rt"
+ ],
+ [
+ "á",
+ "n"
+ ],
+ [
+ "ba",
+ "se"
+ ],
+ [
+ "bas",
+ "e"
+ ],
+ [
+ "b",
+ "ase"
+ ],
+ [
+ "▁w",
+ "hole"
+ ],
+ [
+ "▁wh",
+ "ole"
+ ],
+ [
+ "▁who",
+ "le"
+ ],
+ [
+ "ac",
+ "ión"
+ ],
+ [
+ "ació",
+ "n"
+ ],
+ [
+ "aci",
+ "ón"
+ ],
+ [
+ "a",
+ "ción"
+ ],
+ [
+ "O",
+ "D"
+ ],
+ [
+ "▁turn",
+ "ed"
+ ],
+ [
+ "▁tur",
+ "ned"
+ ],
+ [
+ "ig",
+ "ma"
+ ],
+ [
+ "▁res",
+ "ponse"
+ ],
+ [
+ "▁respon",
+ "se"
+ ],
+ [
+ "▁respons",
+ "e"
+ ],
+ [
+ "▁",
+ "response"
+ ],
+ [
+ "▁Univers",
+ "ity"
+ ],
+ [
+ "▁d",
+ "iv"
+ ],
+ [
+ "▁di",
+ "v"
+ ],
+ [
+ "▁",
+ "div"
+ ],
+ [
+ "ap",
+ "ter"
+ ],
+ [
+ "apt",
+ "er"
+ ],
+ [
+ "▁result",
+ "s"
+ ],
+ [
+ "▁",
+ "results"
+ ],
+ [
+ "▁re",
+ "present"
+ ],
+ [
+ "▁rep",
+ "resent"
+ ],
+ [
+ "▁every",
+ "thing"
+ ],
+ [
+ "▁C",
+ "ent"
+ ],
+ [
+ "▁Ce",
+ "nt"
+ ],
+ [
+ "▁",
+ "Cent"
+ ],
+ [
+ "ut",
+ "es"
+ ],
+ [
+ "ute",
+ "s"
+ ],
+ [
+ "u",
+ "tes"
+ ],
+ [
+ "ri",
+ "x"
+ ],
+ [
+ "r",
+ "ix"
+ ],
+ [
+ "▁S",
+ "ome"
+ ],
+ [
+ "▁So",
+ "me"
+ ],
+ [
+ "▁Som",
+ "e"
+ ],
+ [
+ "▁",
+ "Some"
+ ],
+ [
+ "▁be",
+ "hind"
+ ],
+ [
+ "▁beh",
+ "ind"
+ ],
+ [
+ "▁c",
+ "reat"
+ ],
+ [
+ "▁cre",
+ "at"
+ ],
+ [
+ "▁",
+ "creat"
+ ],
+ [
+ "pl",
+ "ace"
+ ],
+ [
+ "plac",
+ "e"
+ ],
+ [
+ "p",
+ "lace"
+ ],
+ [
+ "s",
+ "u"
+ ],
+ [
+ "▁P",
+ "art"
+ ],
+ [
+ "▁Par",
+ "t"
+ ],
+ [
+ "▁Pa",
+ "rt"
+ ],
+ [
+ "▁",
+ "Part"
+ ],
+ [
+ "um",
+ "b"
+ ],
+ [
+ "u",
+ "mb"
+ ],
+ [
+ "math",
+ "bb"
+ ],
+ [
+ "pi",
+ "ng"
+ ],
+ [
+ "pin",
+ "g"
+ ],
+ [
+ "p",
+ "ing"
+ ],
+ [
+ "▁m",
+ "atch"
+ ],
+ [
+ "▁mat",
+ "ch"
+ ],
+ [
+ "▁",
+ "match"
+ ],
+ [
+ "O",
+ "ut"
+ ],
+ [
+ "do",
+ "m"
+ ],
+ [
+ "d",
+ "om"
+ ],
+ [
+ "▁s",
+ "itu"
+ ],
+ [
+ "▁sit",
+ "u"
+ ],
+ [
+ "▁si",
+ "tu"
+ ],
+ [
+ "d",
+ "r"
+ ],
+ [
+ "ar",
+ "a"
+ ],
+ [
+ "a",
+ "ra"
+ ],
+ [
+ "▁w",
+ "indow"
+ ],
+ [
+ "▁wind",
+ "ow"
+ ],
+ [
+ "▁",
+ "window"
+ ],
+ [
+ "n",
+ "s"
+ ],
+ [
+ "lish",
+ "ed"
+ ],
+ [
+ "l",
+ "ished"
+ ],
+ [
+ "▁V",
+ "er"
+ ],
+ [
+ "▁Ve",
+ "r"
+ ],
+ [
+ "▁",
+ "Ver"
+ ],
+ [
+ "▁m",
+ "essage"
+ ],
+ [
+ "▁mess",
+ "age"
+ ],
+ [
+ "▁",
+ "message"
+ ],
+ [
+ "▁E",
+ "m"
+ ],
+ [
+ "▁",
+ "Em"
+ ],
+ [
+ "▁h",
+ "uman"
+ ],
+ [
+ "▁hum",
+ "an"
+ ],
+ [
+ "▁",
+ "human"
+ ],
+ [
+ "per",
+ "ties"
+ ],
+ [
+ "pert",
+ "ies"
+ ],
+ [
+ "л",
+ "у"
+ ],
+ [
+ "le",
+ "m"
+ ],
+ [
+ "l",
+ "em"
+ ],
+ [
+ "OR",
+ "T"
+ ],
+ [
+ "O",
+ "RT"
+ ],
+ [
+ "▁e",
+ "arly"
+ ],
+ [
+ "▁ear",
+ "ly"
+ ],
+ [
+ "▁qu",
+ "ick"
+ ],
+ [
+ "▁qui",
+ "ck"
+ ],
+ [
+ "▁",
+ "quick"
+ ],
+ [
+ "▁т",
+ "а"
+ ],
+ [
+ "▁",
+ "та"
+ ],
+ [
+ "ro",
+ "id"
+ ],
+ [
+ "r",
+ "oid"
+ ],
+ [
+ "▁c",
+ "ountry"
+ ],
+ [
+ "▁coun",
+ "try"
+ ],
+ [
+ "▁count",
+ "ry"
+ ],
+ [
+ "▁countr",
+ "y"
+ ],
+ [
+ "▁",
+ "country"
+ ],
+ [
+ "▁d",
+ "ue"
+ ],
+ [
+ "▁du",
+ "e"
+ ],
+ [
+ "▁",
+ "due"
+ ],
+ [
+ "▁D",
+ "ie"
+ ],
+ [
+ "▁Di",
+ "e"
+ ],
+ [
+ "▁",
+ "Die"
+ ],
+ [
+ "▁t",
+ "rying"
+ ],
+ [
+ "▁tr",
+ "ying"
+ ],
+ [
+ "▁try",
+ "ing"
+ ],
+ [
+ "▁l",
+ "ive"
+ ],
+ [
+ "▁li",
+ "ve"
+ ],
+ [
+ "▁liv",
+ "e"
+ ],
+ [
+ "▁",
+ "live"
+ ],
+ [
+ "▁p",
+ "ress"
+ ],
+ [
+ "▁pre",
+ "ss"
+ ],
+ [
+ "▁pr",
+ "ess"
+ ],
+ [
+ "▁pres",
+ "s"
+ ],
+ [
+ "▁",
+ "press"
+ ],
+ [
+ "IN",
+ "T"
+ ],
+ [
+ "I",
+ "NT"
+ ],
+ [
+ "W",
+ "ith"
+ ],
+ [
+ "ov",
+ "ed"
+ ],
+ [
+ "ove",
+ "d"
+ ],
+ [
+ "o",
+ "ved"
+ ],
+ [
+ "▁spec",
+ "ific"
+ ],
+ [
+ "▁",
+ "specific"
+ ],
+ [
+ "▁f",
+ "all"
+ ],
+ [
+ "▁fa",
+ "ll"
+ ],
+ [
+ "▁fal",
+ "l"
+ ],
+ [
+ "▁",
+ "fall"
+ ],
+ [
+ "u",
+ "k"
+ ],
+ [
+ "y",
+ "l"
+ ],
+ [
+ "▁gener",
+ "al"
+ ],
+ [
+ "▁gen",
+ "eral"
+ ],
+ [
+ "▁gene",
+ "ral"
+ ],
+ [
+ "▁",
+ "general"
+ ],
+ [
+ "м",
+ "у"
+ ],
+ [
+ "н",
+ "у"
+ ],
+ [
+ "▁n",
+ "ames"
+ ],
+ [
+ "▁name",
+ "s"
+ ],
+ [
+ "▁na",
+ "mes"
+ ],
+ [
+ "▁nam",
+ "es"
+ ],
+ [
+ "▁",
+ "names"
+ ],
+ [
+ "wh",
+ "ere"
+ ],
+ [
+ "whe",
+ "re"
+ ],
+ [
+ "w",
+ "here"
+ ],
+ [
+ "▁The",
+ "se"
+ ],
+ [
+ "▁Th",
+ "ese"
+ ],
+ [
+ "▁",
+ "These"
+ ],
+ [
+ "▁s",
+ "il"
+ ],
+ [
+ "▁si",
+ "l"
+ ],
+ [
+ "▁",
+ "sil"
+ ],
+ [
+ "é",
+ "t"
+ ],
+ [
+ "▁e",
+ "ner"
+ ],
+ [
+ "▁en",
+ "er"
+ ],
+ [
+ "▁",
+ "ener"
+ ],
+ [
+ "▁N",
+ "ow"
+ ],
+ [
+ "▁No",
+ "w"
+ ],
+ [
+ "▁",
+ "Now"
+ ],
+ [
+ "▁add",
+ "ress"
+ ],
+ [
+ "▁addr",
+ "ess"
+ ],
+ [
+ "▁",
+ "address"
+ ],
+ [
+ "Res",
+ "ponse"
+ ],
+ [
+ "▁M",
+ "r"
+ ],
+ [
+ "▁",
+ "Mr"
+ ],
+ [
+ "▁an",
+ "sw"
+ ],
+ [
+ "▁ans",
+ "w"
+ ],
+ [
+ "▁fil",
+ "m"
+ ],
+ [
+ "▁fi",
+ "lm"
+ ],
+ [
+ "▁",
+ "film"
+ ],
+ [
+ "▁str",
+ "ong"
+ ],
+ [
+ "▁stro",
+ "ng"
+ ],
+ [
+ "▁",
+ "strong"
+ ],
+ [
+ "▁b",
+ "ring"
+ ],
+ [
+ "▁br",
+ "ing"
+ ],
+ [
+ "▁Un",
+ "ited"
+ ],
+ [
+ "▁Unit",
+ "ed"
+ ],
+ [
+ "▁g",
+ "e"
+ ],
+ [
+ "▁",
+ "ge"
+ ],
+ [
+ "▁w",
+ "oman"
+ ],
+ [
+ "▁wom",
+ "an"
+ ],
+ [
+ "▁wo",
+ "man"
+ ],
+ [
+ "▁",
+ "woman"
+ ],
+ [
+ "Ne",
+ "w"
+ ],
+ [
+ "N",
+ "ew"
+ ],
+ [
+ "et",
+ "t"
+ ],
+ [
+ "e",
+ "tt"
+ ],
+ [
+ ".",
+ ")"
+ ],
+ [
+ "en",
+ "ame"
+ ],
+ [
+ "ena",
+ "me"
+ ],
+ [
+ "e",
+ "name"
+ ],
+ [
+ "▁A",
+ "N"
+ ],
+ [
+ "▁",
+ "AN"
+ ],
+ [
+ "▁de",
+ "scrib"
+ ],
+ [
+ "▁desc",
+ "rib"
+ ],
+ [
+ "з",
+ "а"
+ ],
+ [
+ "is",
+ "ing"
+ ],
+ [
+ "isi",
+ "ng"
+ ],
+ [
+ "i",
+ "sing"
+ ],
+ [
+ "E",
+ "L"
+ ],
+ [
+ "q",
+ "l"
+ ],
+ [
+ "▁f",
+ "ur"
+ ],
+ [
+ "▁fu",
+ "r"
+ ],
+ [
+ "▁",
+ "fur"
+ ],
+ [
+ "y",
+ "ing"
+ ],
+ [
+ "▁C",
+ "al"
+ ],
+ [
+ "▁Ca",
+ "l"
+ ],
+ [
+ "▁",
+ "Cal"
+ ],
+ [
+ "▁D",
+ "r"
+ ],
+ [
+ "▁",
+ "Dr"
+ ],
+ [
+ "ER",
+ "R"
+ ],
+ [
+ "E",
+ "RR"
+ ],
+ [
+ "▁\\",
+ "\\"
+ ],
+ [
+ "▁",
+ "\\\\"
+ ],
+ [
+ "an",
+ "gle"
+ ],
+ [
+ "ang",
+ "le"
+ ],
+ [
+ "ur",
+ "ope"
+ ],
+ [
+ "uro",
+ "pe"
+ ],
+ [
+ "urop",
+ "e"
+ ],
+ [
+ "▁c",
+ "ity"
+ ],
+ [
+ "▁cit",
+ "y"
+ ],
+ [
+ "▁ci",
+ "ty"
+ ],
+ [
+ "▁",
+ "city"
+ ],
+ [
+ "▁in",
+ "dex"
+ ],
+ [
+ "▁ind",
+ "ex"
+ ],
+ [
+ "▁inde",
+ "x"
+ ],
+ [
+ "▁",
+ "index"
+ ],
+ [
+ "▁a",
+ "ction"
+ ],
+ [
+ "▁act",
+ "ion"
+ ],
+ [
+ "▁",
+ "action"
+ ],
+ [
+ "▁How",
+ "ever"
+ ],
+ [
+ "▁",
+ "However"
+ ],
+ [
+ "▁f",
+ "ig"
+ ],
+ [
+ "▁fi",
+ "g"
+ ],
+ [
+ "▁",
+ "fig"
+ ],
+ [
+ "ia",
+ "s"
+ ],
+ [
+ "i",
+ "as"
+ ],
+ [
+ "▁quest",
+ "ion"
+ ],
+ [
+ "▁",
+ "question"
+ ],
+ [
+ "▁J",
+ "an"
+ ],
+ [
+ "▁Ja",
+ "n"
+ ],
+ [
+ "▁",
+ "Jan"
+ ],
+ [
+ "▁M",
+ "ed"
+ ],
+ [
+ "▁Me",
+ "d"
+ ],
+ [
+ "▁",
+ "Med"
+ ],
+ [
+ "▁C",
+ "ont"
+ ],
+ [
+ "▁Con",
+ "t"
+ ],
+ [
+ "▁Co",
+ "nt"
+ ],
+ [
+ "▁",
+ "Cont"
+ ],
+ [
+ "am",
+ "ed"
+ ],
+ [
+ "ame",
+ "d"
+ ],
+ [
+ "a",
+ "med"
+ ],
+ [
+ "Cal",
+ "l"
+ ],
+ [
+ "C",
+ "all"
+ ],
+ [
+ "pl",
+ "ied"
+ ],
+ [
+ "tt",
+ "y"
+ ],
+ [
+ "t",
+ "ty"
+ ],
+ [
+ "▁ind",
+ "ivid"
+ ],
+ [
+ "pa",
+ "ge"
+ ],
+ [
+ "pag",
+ "e"
+ ],
+ [
+ "p",
+ "age"
+ ],
+ [
+ "▁c",
+ "omb"
+ ],
+ [
+ "▁com",
+ "b"
+ ],
+ [
+ "▁co",
+ "mb"
+ ],
+ [
+ "▁",
+ "comb"
+ ],
+ [
+ "se",
+ "ction"
+ ],
+ [
+ "sect",
+ "ion"
+ ],
+ [
+ "s",
+ "ection"
+ ],
+ [
+ "▁C",
+ "omm"
+ ],
+ [
+ "▁Com",
+ "m"
+ ],
+ [
+ "▁Co",
+ "mm"
+ ],
+ [
+ "▁",
+ "Comm"
+ ],
+ [
+ "ue",
+ "l"
+ ],
+ [
+ "u",
+ "el"
+ ],
+ [
+ "▁h",
+ "et"
+ ],
+ [
+ "▁he",
+ "t"
+ ],
+ [
+ "▁",
+ "het"
+ ],
+ [
+ "▁B",
+ "ar"
+ ],
+ [
+ "▁Ba",
+ "r"
+ ],
+ [
+ "▁",
+ "Bar"
+ ],
+ [
+ "ag",
+ "ement"
+ ],
+ [
+ "age",
+ "ment"
+ ],
+ [
+ "agem",
+ "ent"
+ ],
+ [
+ "fi",
+ "n"
+ ],
+ [
+ "f",
+ "in"
+ ],
+ [
+ "▁m",
+ "ajor"
+ ],
+ [
+ "▁ma",
+ "jor"
+ ],
+ [
+ "▁maj",
+ "or"
+ ],
+ [
+ "▁",
+ "major"
+ ],
+ [
+ "op",
+ "er"
+ ],
+ [
+ "ope",
+ "r"
+ ],
+ [
+ "o",
+ "per"
+ ],
+ [
+ "ap",
+ "i"
+ ],
+ [
+ "a",
+ "pi"
+ ],
+ [
+ "ro",
+ "om"
+ ],
+ [
+ "r",
+ "oom"
+ ],
+ [
+ "▁",
+ "„"
+ ],
+ [
+ "▁h",
+ "ab"
+ ],
+ [
+ "▁ha",
+ "b"
+ ],
+ [
+ "▁",
+ "hab"
+ ],
+ [
+ "з",
+ "и"
+ ],
+ [
+ "▁a",
+ "uf"
+ ],
+ [
+ "▁au",
+ "f"
+ ],
+ [
+ "▁",
+ "auf"
+ ],
+ [
+ "cur",
+ "rent"
+ ],
+ [
+ "curr",
+ "ent"
+ ],
+ [
+ "n",
+ "i"
+ ],
+ [
+ "▁in",
+ "clude"
+ ],
+ [
+ "▁incl",
+ "ude"
+ ],
+ [
+ "▁includ",
+ "e"
+ ],
+ [
+ "▁inclu",
+ "de"
+ ],
+ [
+ "▁",
+ "include"
+ ],
+ [
+ "▁qu",
+ "i"
+ ],
+ [
+ "▁q",
+ "ui"
+ ],
+ [
+ "v",
+ "a"
+ ],
+ [
+ "U",
+ "E"
+ ],
+ [
+ "▁ide",
+ "a"
+ ],
+ [
+ "▁id",
+ "ea"
+ ],
+ [
+ "▁",
+ "idea"
+ ],
+ [
+ ",",
+ "'"
+ ],
+ [
+ "▁requ",
+ "ired"
+ ],
+ [
+ "▁require",
+ "d"
+ ],
+ [
+ "▁",
+ "required"
+ ],
+ [
+ "▁he",
+ "art"
+ ],
+ [
+ "▁hear",
+ "t"
+ ],
+ [
+ "▁",
+ "heart"
+ ],
+ [
+ "ib",
+ "ility"
+ ],
+ [
+ "ibil",
+ "ity"
+ ],
+ [
+ "ict",
+ "ion"
+ ],
+ [
+ "i",
+ "ction"
+ ],
+ [
+ "Mod",
+ "el"
+ ],
+ [
+ "Mode",
+ "l"
+ ],
+ [
+ "Mo",
+ "del"
+ ],
+ [
+ "wr",
+ "ite"
+ ],
+ [
+ "writ",
+ "e"
+ ],
+ [
+ "w",
+ "rite"
+ ],
+ [
+ "▁cont",
+ "ent"
+ ],
+ [
+ "▁conten",
+ "t"
+ ],
+ [
+ "▁",
+ "content"
+ ],
+ [
+ "▁w",
+ "er"
+ ],
+ [
+ "▁we",
+ "r"
+ ],
+ [
+ "▁",
+ "wer"
+ ],
+ [
+ "▁h",
+ "ands"
+ ],
+ [
+ "▁hand",
+ "s"
+ ],
+ [
+ "▁han",
+ "ds"
+ ],
+ [
+ "ze",
+ "n"
+ ],
+ [
+ "z",
+ "en"
+ ],
+ [
+ "ch",
+ "ar"
+ ],
+ [
+ "cha",
+ "r"
+ ],
+ [
+ "c",
+ "har"
+ ],
+ [
+ "}^",
+ "{"
+ ],
+ [
+ "}",
+ "^{"
+ ],
+ [
+ "▁m",
+ "ass"
+ ],
+ [
+ "▁ma",
+ "ss"
+ ],
+ [
+ "▁mas",
+ "s"
+ ],
+ [
+ "▁",
+ "mass"
+ ],
+ [
+ "pl",
+ "y"
+ ],
+ [
+ "p",
+ "ly"
+ ],
+ [
+ "▁n",
+ "at"
+ ],
+ [
+ "▁na",
+ "t"
+ ],
+ [
+ "▁",
+ "nat"
+ ],
+ [
+ "re",
+ "l"
+ ],
+ [
+ "r",
+ "el"
+ ],
+ [
+ "▁d",
+ "at"
+ ],
+ [
+ "▁da",
+ "t"
+ ],
+ [
+ "▁",
+ "dat"
+ ],
+ [
+ "====",
+ "============"
+ ],
+ [
+ "========",
+ "========"
+ ],
+ [
+ "============",
+ "===="
+ ],
+ [
+ "im",
+ "al"
+ ],
+ [
+ "ima",
+ "l"
+ ],
+ [
+ "i",
+ "mal"
+ ],
+ [
+ "▁pro",
+ "bably"
+ ],
+ [
+ "▁prob",
+ "ably"
+ ],
+ [
+ "un",
+ "ch"
+ ],
+ [
+ "unc",
+ "h"
+ ],
+ [
+ "▁m",
+ "er"
+ ],
+ [
+ "▁me",
+ "r"
+ ],
+ [
+ "▁",
+ "mer"
+ ],
+ [
+ "il",
+ "ar"
+ ],
+ [
+ "ila",
+ "r"
+ ],
+ [
+ "i",
+ "lar"
+ ],
+ [
+ "ir",
+ "es"
+ ],
+ [
+ "ire",
+ "s"
+ ],
+ [
+ "i",
+ "res"
+ ],
+ [
+ "▁w",
+ "atch"
+ ],
+ [
+ "▁wat",
+ "ch"
+ ],
+ [
+ "▁",
+ "watch"
+ ],
+ [
+ "S",
+ "I"
+ ],
+ [
+ "▁c",
+ "ult"
+ ],
+ [
+ "▁cu",
+ "lt"
+ ],
+ [
+ "▁cul",
+ "t"
+ ],
+ [
+ "▁m",
+ "other"
+ ],
+ [
+ "▁mot",
+ "her"
+ ],
+ [
+ "▁mo",
+ "ther"
+ ],
+ [
+ "▁",
+ "mother"
+ ],
+ [
+ "▁govern",
+ "ment"
+ ],
+ [
+ "or",
+ "ding"
+ ],
+ [
+ "ord",
+ "ing"
+ ],
+ [
+ "▁(",
+ ")"
+ ],
+ [
+ "▁",
+ "()"
+ ],
+ [
+ "▁p",
+ "ri"
+ ],
+ [
+ "▁pr",
+ "i"
+ ],
+ [
+ "▁l",
+ "ink"
+ ],
+ [
+ "▁lin",
+ "k"
+ ],
+ [
+ "▁",
+ "link"
+ ],
+ [
+ "gr",
+ "oup"
+ ],
+ [
+ "gro",
+ "up"
+ ],
+ [
+ "g",
+ "roup"
+ ],
+ [
+ "O",
+ "L"
+ ],
+ [
+ "▁n",
+ "ear"
+ ],
+ [
+ "▁ne",
+ "ar"
+ ],
+ [
+ "▁S",
+ "er"
+ ],
+ [
+ "▁Se",
+ "r"
+ ],
+ [
+ "▁",
+ "Ser"
+ ],
+ [
+ "Se",
+ "r"
+ ],
+ [
+ "S",
+ "er"
+ ],
+ [
+ "it",
+ "o"
+ ],
+ [
+ "i",
+ "to"
+ ],
+ [
+ "▁value",
+ "s"
+ ],
+ [
+ "▁val",
+ "ues"
+ ],
+ [
+ "▁",
+ "values"
+ ],
+ [
+ "▁j",
+ "ava"
+ ],
+ [
+ "▁ja",
+ "va"
+ ],
+ [
+ "▁",
+ "java"
+ ],
+ [
+ "ful",
+ "ly"
+ ],
+ [
+ "full",
+ "y"
+ ],
+ [
+ "f",
+ "ully"
+ ],
+ [
+ "Co",
+ "unt"
+ ],
+ [
+ "C",
+ "ount"
+ ],
+ [
+ "++",
+ ")"
+ ],
+ [
+ "▁v",
+ "i"
+ ],
+ [
+ "▁",
+ "vi"
+ ],
+ [
+ "▁wh",
+ "ite"
+ ],
+ [
+ "▁",
+ "white"
+ ],
+ [
+ "ma",
+ "t"
+ ],
+ [
+ "m",
+ "at"
+ ],
+ [
+ "ct",
+ "x"
+ ],
+ [
+ "c",
+ "tx"
+ ],
+ [
+ "▁con",
+ "c"
+ ],
+ [
+ "▁co",
+ "nc"
+ ],
+ [
+ "▁",
+ "conc"
+ ],
+ [
+ "▁st",
+ "ay"
+ ],
+ [
+ "▁sta",
+ "y"
+ ],
+ [
+ "gi",
+ "ng"
+ ],
+ [
+ "gin",
+ "g"
+ ],
+ [
+ "g",
+ "ing"
+ ],
+ [
+ "▁c",
+ "lear"
+ ],
+ [
+ "▁cl",
+ "ear"
+ ],
+ [
+ "▁cle",
+ "ar"
+ ],
+ [
+ "▁",
+ "clear"
+ ],
+ [
+ "▁c",
+ "opy"
+ ],
+ [
+ "▁co",
+ "py"
+ ],
+ [
+ "▁cop",
+ "y"
+ ],
+ [
+ "▁",
+ "copy"
+ ],
+ [
+ "sel",
+ "ves"
+ ],
+ [
+ "▁prov",
+ "ide"
+ ],
+ [
+ "▁w",
+ "ords"
+ ],
+ [
+ "▁wor",
+ "ds"
+ ],
+ [
+ "▁word",
+ "s"
+ ],
+ [
+ "▁",
+ "words"
+ ],
+ [
+ "com",
+ "p"
+ ],
+ [
+ "co",
+ "mp"
+ ],
+ [
+ "c",
+ "omp"
+ ],
+ [
+ "ar",
+ "gs"
+ ],
+ [
+ "arg",
+ "s"
+ ],
+ [
+ "▁p",
+ "ick"
+ ],
+ [
+ "▁pi",
+ "ck"
+ ],
+ [
+ "▁pic",
+ "k"
+ ],
+ [
+ "▁",
+ "pick"
+ ],
+ [
+ "ul",
+ "y"
+ ],
+ [
+ "u",
+ "ly"
+ ],
+ [
+ "▁v",
+ "ari"
+ ],
+ [
+ "▁var",
+ "i"
+ ],
+ [
+ "▁va",
+ "ri"
+ ],
+ [
+ "▁",
+ "vari"
+ ],
+ [
+ "▁bel",
+ "ieve"
+ ],
+ [
+ "▁belie",
+ "ve"
+ ],
+ [
+ "▁C",
+ "o"
+ ],
+ [
+ "▁",
+ "Co"
+ ],
+ [
+ "Pro",
+ "perty"
+ ],
+ [
+ "Gr",
+ "oup"
+ ],
+ [
+ "G",
+ "roup"
+ ],
+ [
+ "▁t",
+ "en"
+ ],
+ [
+ "▁te",
+ "n"
+ ],
+ [
+ "▁",
+ "ten"
+ ],
+ [
+ "is",
+ "chen"
+ ],
+ [
+ "isch",
+ "en"
+ ],
+ [
+ "ische",
+ "n"
+ ],
+ [
+ "isc",
+ "hen"
+ ],
+ [
+ "i",
+ "schen"
+ ],
+ [
+ "et",
+ "urn"
+ ],
+ [
+ "e",
+ "turn"
+ ],
+ [
+ "iv",
+ "al"
+ ],
+ [
+ "iva",
+ "l"
+ ],
+ [
+ "i",
+ "val"
+ ],
+ [
+ "Sys",
+ "tem"
+ ],
+ [
+ "S",
+ "ystem"
+ ],
+ [
+ "C",
+ "L"
+ ],
+ [
+ "be",
+ "d"
+ ],
+ [
+ "b",
+ "ed"
+ ],
+ [
+ "▁t",
+ "otal"
+ ],
+ [
+ "▁to",
+ "tal"
+ ],
+ [
+ "▁tot",
+ "al"
+ ],
+ [
+ "▁",
+ "total"
+ ],
+ [
+ "▁is",
+ "t"
+ ],
+ [
+ "▁i",
+ "st"
+ ],
+ [
+ "▁",
+ "ist"
+ ],
+ [
+ "In",
+ "put"
+ ],
+ [
+ "um",
+ "ents"
+ ],
+ [
+ "ument",
+ "s"
+ ],
+ [
+ "umen",
+ "ts"
+ ],
+ [
+ "u",
+ "ments"
+ ],
+ [
+ "Man",
+ "ager"
+ ],
+ [
+ "ш",
+ "и"
+ ],
+ [
+ "▁w",
+ "in"
+ ],
+ [
+ "▁",
+ "win"
+ ],
+ [
+ "le",
+ "ep"
+ ],
+ [
+ "lee",
+ "p"
+ ],
+ [
+ "P",
+ "I"
+ ],
+ [
+ "но",
+ "го"
+ ],
+ [
+ "н",
+ "ого"
+ ],
+ [
+ "ru",
+ "ction"
+ ],
+ [
+ "ruct",
+ "ion"
+ ],
+ [
+ "r",
+ "uction"
+ ],
+ [
+ "▁in",
+ "te"
+ ],
+ [
+ "▁i",
+ "nte"
+ ],
+ [
+ "▁int",
+ "e"
+ ],
+ [
+ "▁",
+ "inte"
+ ],
+ [
+ "Ap",
+ "p"
+ ],
+ [
+ "A",
+ "pp"
+ ],
+ [
+ "av",
+ "or"
+ ],
+ [
+ "avo",
+ "r"
+ ],
+ [
+ "a",
+ "vor"
+ ],
+ [
+ "▁re",
+ "spect"
+ ],
+ [
+ "▁res",
+ "pect"
+ ],
+ [
+ "▁resp",
+ "ect"
+ ],
+ [
+ "▁",
+ "respect"
+ ],
+ [
+ "at",
+ "ors"
+ ],
+ [
+ "ator",
+ "s"
+ ],
+ [
+ "ato",
+ "rs"
+ ],
+ [
+ "▁c",
+ "omo"
+ ],
+ [
+ "▁com",
+ "o"
+ ],
+ [
+ "▁co",
+ "mo"
+ ],
+ [
+ "▁c",
+ "ut"
+ ],
+ [
+ "▁cu",
+ "t"
+ ],
+ [
+ "▁",
+ "cut"
+ ],
+ [
+ "F",
+ "A"
+ ],
+ [
+ "▁s",
+ "us"
+ ],
+ [
+ "▁su",
+ "s"
+ ],
+ [
+ "▁A",
+ "pp"
+ ],
+ [
+ "▁Ap",
+ "p"
+ ],
+ [
+ "▁",
+ "App"
+ ],
+ [
+ "re",
+ "ct"
+ ],
+ [
+ "rec",
+ "t"
+ ],
+ [
+ "r",
+ "ect"
+ ],
+ [
+ "F",
+ "I"
+ ],
+ [
+ "▁be",
+ "gan"
+ ],
+ [
+ "▁beg",
+ "an"
+ ],
+ [
+ "op",
+ "h"
+ ],
+ [
+ "o",
+ "ph"
+ ],
+ [
+ "▁s",
+ "ort"
+ ],
+ [
+ "▁so",
+ "rt"
+ ],
+ [
+ "▁sor",
+ "t"
+ ],
+ [
+ "▁",
+ "sort"
+ ],
+ [
+ "th",
+ "ough"
+ ],
+ [
+ "ј",
+ "е"
+ ],
+ [
+ "ic",
+ "ro"
+ ],
+ [
+ "i",
+ "cro"
+ ],
+ [
+ "Tr",
+ "ans"
+ ],
+ [
+ "Tra",
+ "ns"
+ ],
+ [
+ "л",
+ "і"
+ ],
+ [
+ "▁In",
+ "st"
+ ],
+ [
+ "▁Ins",
+ "t"
+ ],
+ [
+ "▁",
+ "Inst"
+ ],
+ [
+ "re",
+ "quest"
+ ],
+ [
+ "requ",
+ "est"
+ ],
+ [
+ "req",
+ "uest"
+ ],
+ [
+ "о",
+ "р"
+ ],
+ [
+ "▁rel",
+ "ations"
+ ],
+ [
+ "▁relation",
+ "s"
+ ],
+ [
+ "-",
+ "\\"
+ ],
+ [
+ "St",
+ "atus"
+ ],
+ [
+ "Stat",
+ "us"
+ ],
+ [
+ "ж",
+ "и"
+ ],
+ [
+ "▁f",
+ "ather"
+ ],
+ [
+ "▁fa",
+ "ther"
+ ],
+ [
+ "▁fat",
+ "her"
+ ],
+ [
+ "▁",
+ "father"
+ ],
+ [
+ "c",
+ "s"
+ ],
+ [
+ "▁s",
+ "ex"
+ ],
+ [
+ "▁se",
+ "x"
+ ],
+ [
+ "▁",
+ "sex"
+ ],
+ [
+ "is",
+ "ch"
+ ],
+ [
+ "isc",
+ "h"
+ ],
+ [
+ "i",
+ "sch"
+ ],
+ [
+ "v",
+ "o"
+ ],
+ [
+ "}_",
+ "{"
+ ],
+ [
+ "}",
+ "_{"
+ ],
+ [
+ "ave",
+ "n"
+ ],
+ [
+ "av",
+ "en"
+ ],
+ [
+ "a",
+ "ven"
+ ],
+ [
+ "▁N",
+ "e"
+ ],
+ [
+ "▁",
+ "Ne"
+ ],
+ [
+ "AT",
+ "E"
+ ],
+ [
+ "A",
+ "TE"
+ ],
+ [
+ "it",
+ "ten"
+ ],
+ [
+ "itt",
+ "en"
+ ],
+ [
+ "itte",
+ "n"
+ ],
+ [
+ "▁e",
+ "ss"
+ ],
+ [
+ "▁es",
+ "s"
+ ],
+ [
+ "▁",
+ "ess"
+ ],
+ [
+ "T",
+ "H"
+ ],
+ [
+ "ight",
+ "s"
+ ],
+ [
+ "igh",
+ "ts"
+ ],
+ [
+ "▁h",
+ "om"
+ ],
+ [
+ "▁ho",
+ "m"
+ ],
+ [
+ "▁",
+ "hom"
+ ],
+ [
+ "▁t",
+ "oday"
+ ],
+ [
+ "▁to",
+ "day"
+ ],
+ [
+ "▁tod",
+ "ay"
+ ],
+ [
+ "▁toda",
+ "y"
+ ],
+ [
+ "▁z",
+ "u"
+ ],
+ [
+ "▁",
+ "zu"
+ ],
+ [
+ "it",
+ "a"
+ ],
+ [
+ "i",
+ "ta"
+ ],
+ [
+ "▁is",
+ "n"
+ ],
+ [
+ "▁i",
+ "sn"
+ ],
+ [
+ "▁o",
+ "pt"
+ ],
+ [
+ "▁op",
+ "t"
+ ],
+ [
+ "▁",
+ "opt"
+ ],
+ [
+ "og",
+ "n"
+ ],
+ [
+ "o",
+ "gn"
+ ],
+ [
+ "é",
+ "r"
+ ],
+ [
+ "▁wh",
+ "ether"
+ ],
+ [
+ "▁whe",
+ "ther"
+ ],
+ [
+ "ix",
+ "ed"
+ ],
+ [
+ "ph",
+ "i"
+ ],
+ [
+ "p",
+ "hi"
+ ],
+ [
+ "id",
+ "ence"
+ ],
+ [
+ "iden",
+ "ce"
+ ],
+ [
+ "al",
+ "d"
+ ],
+ [
+ "a",
+ "ld"
+ ],
+ [
+ "Cl",
+ "ient"
+ ],
+ [
+ "A",
+ "t"
+ ],
+ [
+ "▁de",
+ "ath"
+ ],
+ [
+ "▁L",
+ "et"
+ ],
+ [
+ "▁Le",
+ "t"
+ ],
+ [
+ "▁",
+ "Let"
+ ],
+ [
+ "iu",
+ "s"
+ ],
+ [
+ "i",
+ "us"
+ ],
+ [
+ "г",
+ "и"
+ ],
+ [
+ "▁р",
+ "е"
+ ],
+ [
+ "▁",
+ "ре"
+ ],
+ [
+ "be",
+ "n"
+ ],
+ [
+ "b",
+ "en"
+ ],
+ [
+ ")",
+ "\r"
+ ],
+ [
+ "b",
+ "a"
+ ],
+ [
+ "><",
+ "/"
+ ],
+ [
+ ">",
+ ""
+ ],
+ [
+ "ave",
+ "l"
+ ],
+ [
+ "av",
+ "el"
+ ],
+ [
+ "a",
+ "vel"
+ ],
+ [
+ "▁m",
+ "iss"
+ ],
+ [
+ "▁mis",
+ "s"
+ ],
+ [
+ "▁mi",
+ "ss"
+ ],
+ [
+ "▁",
+ "miss"
+ ],
+ [
+ "▁n",
+ "ode"
+ ],
+ [
+ "▁no",
+ "de"
+ ],
+ [
+ "▁nod",
+ "e"
+ ],
+ [
+ "▁",
+ "node"
+ ],
+ [
+ "▁(",
+ "$"
+ ],
+ [
+ "▁",
+ "($"
+ ],
+ [
+ "▁col",
+ "or"
+ ],
+ [
+ "▁co",
+ "lor"
+ ],
+ [
+ "▁",
+ "color"
+ ],
+ [
+ "▁o",
+ "bt"
+ ],
+ [
+ "▁ob",
+ "t"
+ ],
+ [
+ "to",
+ "t"
+ ],
+ [
+ "t",
+ "ot"
+ ],
+ [
+ "▁п",
+ "ре"
+ ],
+ [
+ "▁пр",
+ "е"
+ ],
+ [
+ "▁",
+ "пре"
+ ],
+ [
+ "CO",
+ "N"
+ ],
+ [
+ "C",
+ "ON"
+ ],
+ [
+ "et",
+ "te"
+ ],
+ [
+ "ett",
+ "e"
+ ],
+ [
+ "▁G",
+ "o"
+ ],
+ [
+ "▁",
+ "Go"
+ ],
+ [
+ "F",
+ "l"
+ ],
+ [
+ "▁D",
+ "on"
+ ],
+ [
+ "▁Do",
+ "n"
+ ],
+ [
+ "▁",
+ "Don"
+ ],
+ [
+ "▁c",
+ "rit"
+ ],
+ [
+ "▁cr",
+ "it"
+ ],
+ [
+ "▁cri",
+ "t"
+ ],
+ [
+ "▁",
+ "crit"
+ ],
+ [
+ "▁r",
+ "i"
+ ],
+ [
+ "▁",
+ "ri"
+ ],
+ [
+ "pos",
+ "t"
+ ],
+ [
+ "po",
+ "st"
+ ],
+ [
+ "p",
+ "ost"
+ ],
+ [
+ "▁-",
+ ">"
+ ],
+ [
+ "▁",
+ "->"
+ ],
+ [
+ "▁J",
+ "ust"
+ ],
+ [
+ "▁Ju",
+ "st"
+ ],
+ [
+ "▁",
+ "Just"
+ ],
+ [
+ "Wh",
+ "at"
+ ],
+ [
+ "W",
+ "hat"
+ ],
+ [
+ "at",
+ "al"
+ ],
+ [
+ "ata",
+ "l"
+ ],
+ [
+ "a",
+ "tal"
+ ],
+ [
+ "▁M",
+ "in"
+ ],
+ [
+ "▁Mi",
+ "n"
+ ],
+ [
+ "▁",
+ "Min"
+ ],
+ [
+ "▁C",
+ "or"
+ ],
+ [
+ "▁Co",
+ "r"
+ ],
+ [
+ "▁",
+ "Cor"
+ ],
+ [
+ "▁d",
+ "ark"
+ ],
+ [
+ "▁dar",
+ "k"
+ ],
+ [
+ "▁",
+ "dark"
+ ],
+ [
+ "r",
+ "l"
+ ],
+ [
+ "▁l",
+ "arg"
+ ],
+ [
+ "▁la",
+ "rg"
+ ],
+ [
+ "▁",
+ "larg"
+ ],
+ [
+ "di",
+ "ng"
+ ],
+ [
+ "d",
+ "ing"
+ ],
+ [
+ "ó",
+ "n"
+ ],
+ [
+ "ou",
+ "ch"
+ ],
+ [
+ "o",
+ "uch"
+ ],
+ [
+ "▁u",
+ "m"
+ ],
+ [
+ "▁",
+ "um"
+ ],
+ [
+ "▁e",
+ "lect"
+ ],
+ [
+ "▁el",
+ "ect"
+ ],
+ [
+ "▁ele",
+ "ct"
+ ],
+ [
+ "▁",
+ "elect"
+ ],
+ [
+ "▁d",
+ "am"
+ ],
+ [
+ "▁da",
+ "m"
+ ],
+ [
+ "▁",
+ "dam"
+ ],
+ [
+ "▁ne",
+ "eds"
+ ],
+ [
+ "▁need",
+ "s"
+ ],
+ [
+ "▁m",
+ "atter"
+ ],
+ [
+ "▁mat",
+ "ter"
+ ],
+ [
+ "▁matt",
+ "er"
+ ],
+ [
+ "▁r",
+ "ather"
+ ],
+ [
+ "▁rat",
+ "her"
+ ],
+ [
+ "▁ra",
+ "ther"
+ ],
+ [
+ "fr",
+ "om"
+ ],
+ [
+ "f",
+ "rom"
+ ],
+ [
+ "ra",
+ "m"
+ ],
+ [
+ "r",
+ "am"
+ ],
+ [
+ "▁",
+ "і"
+ ],
+ [
+ "▁t",
+ "aken"
+ ],
+ [
+ "▁take",
+ "n"
+ ],
+ [
+ "▁tak",
+ "en"
+ ],
+ [
+ "▁ta",
+ "ken"
+ ],
+ [
+ "▁de",
+ "al"
+ ],
+ [
+ "▁per",
+ "iod"
+ ],
+ [
+ "▁",
+ "period"
+ ],
+ [
+ "▁M",
+ "on"
+ ],
+ [
+ "▁Mo",
+ "n"
+ ],
+ [
+ "▁",
+ "Mon"
+ ],
+ [
+ "▁",
+ "Л"
+ ],
+ [
+ "▁A",
+ "ug"
+ ],
+ [
+ "▁Au",
+ "g"
+ ],
+ [
+ "▁",
+ "Aug"
+ ],
+ [
+ "ru",
+ "n"
+ ],
+ [
+ "r",
+ "un"
+ ],
+ [
+ "m",
+ "m"
+ ],
+ [
+ "el",
+ "le"
+ ],
+ [
+ "ell",
+ "e"
+ ],
+ [
+ "e",
+ "lle"
+ ],
+ [
+ "▁ex",
+ "port"
+ ],
+ [
+ "▁exp",
+ "ort"
+ ],
+ [
+ "▁",
+ "export"
+ ],
+ [
+ "S",
+ "c"
+ ],
+ [
+ "vi",
+ "s"
+ ],
+ [
+ "v",
+ "is"
+ ],
+ [
+ "ab",
+ "or"
+ ],
+ [
+ "a",
+ "bor"
+ ],
+ [
+ "▁aut",
+ "hor"
+ ],
+ [
+ "▁auth",
+ "or"
+ ],
+ [
+ "▁",
+ "author"
+ ],
+ [
+ "è",
+ "re"
+ ],
+ [
+ "▁re",
+ "member"
+ ],
+ [
+ "▁rem",
+ "ember"
+ ],
+ [
+ "▁remem",
+ "ber"
+ ],
+ [
+ "▁re",
+ "du"
+ ],
+ [
+ "▁r",
+ "edu"
+ ],
+ [
+ "▁red",
+ "u"
+ ],
+ [
+ "▁",
+ "redu"
+ ],
+ [
+ "▁L",
+ "ist"
+ ],
+ [
+ "▁Li",
+ "st"
+ ],
+ [
+ "▁Lis",
+ "t"
+ ],
+ [
+ "▁",
+ "List"
+ ],
+ [
+ "▁f",
+ "ocus"
+ ],
+ [
+ "▁",
+ "focus"
+ ],
+ [
+ "▁char",
+ "acter"
+ ],
+ [
+ "▁",
+ "character"
+ ],
+ [
+ "Tab",
+ "le"
+ ],
+ [
+ "T",
+ "able"
+ ],
+ [
+ "▁individ",
+ "ual"
+ ],
+ [
+ "▁need",
+ "ed"
+ ],
+ [
+ "bu",
+ "m"
+ ],
+ [
+ "b",
+ "um"
+ ],
+ [
+ "▁st",
+ "yle"
+ ],
+ [
+ "▁sty",
+ "le"
+ ],
+ [
+ "▁",
+ "style"
+ ],
+ [
+ "in",
+ "ary"
+ ],
+ [
+ "ina",
+ "ry"
+ ],
+ [
+ "inar",
+ "y"
+ ],
+ [
+ "ers",
+ "ion"
+ ],
+ [
+ "ou",
+ "te"
+ ],
+ [
+ "out",
+ "e"
+ ],
+ [
+ "o",
+ "ute"
+ ],
+ [
+ "▁P",
+ "e"
+ ],
+ [
+ "▁",
+ "Pe"
+ ],
+ [
+ "▁h",
+ "on"
+ ],
+ [
+ "▁ho",
+ "n"
+ ],
+ [
+ "▁",
+ "hon"
+ ],
+ [
+ "mu",
+ "t"
+ ],
+ [
+ "m",
+ "ut"
+ ],
+ [
+ "se",
+ "e"
+ ],
+ [
+ "s",
+ "ee"
+ ],
+ [
+ "▁bec",
+ "ame"
+ ],
+ [
+ "▁d",
+ "ire"
+ ],
+ [
+ "▁di",
+ "re"
+ ],
+ [
+ "▁dir",
+ "e"
+ ],
+ [
+ "▁",
+ "dire"
+ ],
+ [
+ "▁d",
+ "ocument"
+ ],
+ [
+ "▁doc",
+ "ument"
+ ],
+ [
+ "▁",
+ "document"
+ ],
+ [
+ "se",
+ "c"
+ ],
+ [
+ "s",
+ "ec"
+ ],
+ [
+ "en",
+ "ing"
+ ],
+ [
+ "eni",
+ "ng"
+ ],
+ [
+ "e",
+ "ning"
+ ],
+ [
+ "▁vis",
+ "it"
+ ],
+ [
+ "▁",
+ "visit"
+ ],
+ [
+ "▁f",
+ "ac"
+ ],
+ [
+ "▁fa",
+ "c"
+ ],
+ [
+ "▁",
+ "fac"
+ ],
+ [
+ "t",
+ "x"
+ ],
+ [
+ "do",
+ "wn"
+ ],
+ [
+ "d",
+ "own"
+ ],
+ [
+ "pl",
+ "it"
+ ],
+ [
+ "p",
+ "lit"
+ ],
+ [
+ "▁ph",
+ "ys"
+ ],
+ [
+ "▁",
+ "phys"
+ ],
+ [
+ "it",
+ "ting"
+ ],
+ [
+ "itt",
+ "ing"
+ ],
+ [
+ "jo",
+ "y"
+ ],
+ [
+ "j",
+ "oy"
+ ],
+ [
+ "▁h",
+ "ig"
+ ],
+ [
+ "▁hi",
+ "g"
+ ],
+ [
+ "Th",
+ "is"
+ ],
+ [
+ "T",
+ "his"
+ ],
+ [
+ "A",
+ "d"
+ ],
+ [
+ "▁B",
+ "rit"
+ ],
+ [
+ "▁Br",
+ "it"
+ ],
+ [
+ "▁em",
+ "ploy"
+ ],
+ [
+ "▁r",
+ "é"
+ ],
+ [
+ "▁",
+ "ré"
+ ],
+ [
+ "▁",
+ "т"
+ ],
+ [
+ "l",
+ "ambda"
+ ],
+ [
+ "▁im",
+ "pro"
+ ],
+ [
+ "▁imp",
+ "ro"
+ ],
+ [
+ "▁B",
+ "o"
+ ],
+ [
+ "▁",
+ "Bo"
+ ],
+ [
+ "id",
+ "ing"
+ ],
+ [
+ "idi",
+ "ng"
+ ],
+ [
+ "i",
+ "ding"
+ ],
+ [
+ "▁on",
+ "line"
+ ],
+ [
+ "▁",
+ "online"
+ ],
+ [
+ "me",
+ "m"
+ ],
+ [
+ "m",
+ "em"
+ ],
+ [
+ "at",
+ "form"
+ ],
+ [
+ "▁W",
+ "ar"
+ ],
+ [
+ "▁Wa",
+ "r"
+ ],
+ [
+ "▁",
+ "War"
+ ],
+ [
+ "▁c",
+ "as"
+ ],
+ [
+ "▁ca",
+ "s"
+ ],
+ [
+ "▁",
+ "cas"
+ ],
+ [
+ "as",
+ "ure"
+ ],
+ [
+ "a",
+ "sure"
+ ],
+ [
+ "▁p",
+ "ur"
+ ],
+ [
+ "▁pu",
+ "r"
+ ],
+ [
+ "▁",
+ "pur"
+ ],
+ [
+ "me",
+ "di"
+ ],
+ [
+ "med",
+ "i"
+ ],
+ [
+ "m",
+ "edi"
+ ],
+ [
+ "Di",
+ "s"
+ ],
+ [
+ "D",
+ "is"
+ ],
+ [
+ "▁G",
+ "erm"
+ ],
+ [
+ "▁Ge",
+ "rm"
+ ],
+ [
+ "▁Ger",
+ "m"
+ ],
+ [
+ "p",
+ "c"
+ ],
+ [
+ "с",
+ "а"
+ ],
+ [
+ "▁friend",
+ "s"
+ ],
+ [
+ "▁M",
+ "c"
+ ],
+ [
+ "▁",
+ "Mc"
+ ],
+ [
+ "D",
+ "I"
+ ],
+ [
+ "▁pl",
+ "us"
+ ],
+ [
+ "▁",
+ "plus"
+ ],
+ [
+ "▁S",
+ "et"
+ ],
+ [
+ "▁Se",
+ "t"
+ ],
+ [
+ "▁",
+ "Set"
+ ],
+ [
+ "idd",
+ "le"
+ ],
+ [
+ "it",
+ "ut"
+ ],
+ [
+ "itu",
+ "t"
+ ],
+ [
+ "▁de",
+ "pend"
+ ],
+ [
+ "▁dep",
+ "end"
+ ],
+ [
+ "▁",
+ "depend"
+ ],
+ [
+ "re",
+ "st"
+ ],
+ [
+ "res",
+ "t"
+ ],
+ [
+ "r",
+ "est"
+ ],
+ [
+ "▁J",
+ "e"
+ ],
+ [
+ "▁",
+ "Je"
+ ],
+ [
+ "▁h",
+ "or"
+ ],
+ [
+ "▁ho",
+ "r"
+ ],
+ [
+ "▁",
+ "hor"
+ ],
+ [
+ "▁ent",
+ "ire"
+ ],
+ [
+ "Qu",
+ "ery"
+ ],
+ [
+ "Que",
+ "ry"
+ ],
+ [
+ "▁re",
+ "fer"
+ ],
+ [
+ "▁ref",
+ "er"
+ ],
+ [
+ "▁",
+ "refer"
+ ],
+ [
+ "▁h",
+ "ot"
+ ],
+ [
+ "▁ho",
+ "t"
+ ],
+ [
+ "▁",
+ "hot"
+ ],
+ [
+ "▁A",
+ "ust"
+ ],
+ [
+ "▁Aus",
+ "t"
+ ],
+ [
+ "▁Au",
+ "st"
+ ],
+ [
+ "▁com",
+ "mon"
+ ],
+ [
+ "▁comm",
+ "on"
+ ],
+ [
+ "▁",
+ "common"
+ ],
+ [
+ "ц",
+ "і"
+ ],
+ [
+ "▁p",
+ "ull"
+ ],
+ [
+ "▁pu",
+ "ll"
+ ],
+ [
+ "▁pul",
+ "l"
+ ],
+ [
+ "▁",
+ "pull"
+ ],
+ [
+ "▁A",
+ "dd"
+ ],
+ [
+ "▁Ad",
+ "d"
+ ],
+ [
+ "▁",
+ "Add"
+ ],
+ [
+ "▁se",
+ "ason"
+ ],
+ [
+ "▁sea",
+ "son"
+ ],
+ [
+ "▁seas",
+ "on"
+ ],
+ [
+ "▁",
+ "season"
+ ],
+ [
+ "▁in",
+ "vol"
+ ],
+ [
+ "▁inv",
+ "ol"
+ ],
+ [
+ "▁W",
+ "orld"
+ ],
+ [
+ "▁Wor",
+ "ld"
+ ],
+ [
+ "▁",
+ "World"
+ ],
+ [
+ "cl",
+ "ient"
+ ],
+ [
+ "cli",
+ "ent"
+ ],
+ [
+ "no",
+ "w"
+ ],
+ [
+ "n",
+ "ow"
+ ],
+ [
+ "tr",
+ "ue"
+ ],
+ [
+ "ap",
+ "pend"
+ ],
+ [
+ "app",
+ "end"
+ ],
+ [
+ "appe",
+ "nd"
+ ],
+ [
+ "appen",
+ "d"
+ ],
+ [
+ "it",
+ "ted"
+ ],
+ [
+ "itt",
+ "ed"
+ ],
+ [
+ "itte",
+ "d"
+ ],
+ [
+ "em",
+ "pt"
+ ],
+ [
+ "emp",
+ "t"
+ ],
+ [
+ ")",
+ "{"
+ ],
+ [
+ "//",
+ "/"
+ ],
+ [
+ "/",
+ "//"
+ ],
+ [
+ "▁p",
+ "rop"
+ ],
+ [
+ "▁pro",
+ "p"
+ ],
+ [
+ "▁pr",
+ "op"
+ ],
+ [
+ "▁",
+ "prop"
+ ],
+ [
+ "im",
+ "ate"
+ ],
+ [
+ "ima",
+ "te"
+ ],
+ [
+ "imat",
+ "e"
+ ],
+ [
+ "i",
+ "mate"
+ ],
+ [
+ "S",
+ "C"
+ ],
+ [
+ "▁h",
+ "ours"
+ ],
+ [
+ "▁hour",
+ "s"
+ ],
+ [
+ "▁ho",
+ "urs"
+ ],
+ [
+ "▁h",
+ "ope"
+ ],
+ [
+ "▁hop",
+ "e"
+ ],
+ [
+ "▁ho",
+ "pe"
+ ],
+ [
+ "an",
+ "dom"
+ ],
+ [
+ "and",
+ "om"
+ ],
+ [
+ "ando",
+ "m"
+ ],
+ [
+ "і",
+ "д"
+ ],
+ [
+ "ist",
+ "ic"
+ ],
+ [
+ "isti",
+ "c"
+ ],
+ [
+ "▁pro",
+ "perty"
+ ],
+ [
+ "▁proper",
+ "ty"
+ ],
+ [
+ "▁",
+ "property"
+ ],
+ [
+ "s",
+ "g"
+ ],
+ [
+ ">",
+ "("
+ ],
+ [
+ "▁w",
+ "rite"
+ ],
+ [
+ "▁wr",
+ "ite"
+ ],
+ [
+ "▁writ",
+ "e"
+ ],
+ [
+ "▁",
+ "write"
+ ],
+ [
+ "mar",
+ "k"
+ ],
+ [
+ "m",
+ "ark"
+ ],
+ [
+ "fin",
+ "d"
+ ],
+ [
+ "fi",
+ "nd"
+ ],
+ [
+ "f",
+ "ind"
+ ],
+ [
+ "▁person",
+ "al"
+ ],
+ [
+ "▁pers",
+ "onal"
+ ],
+ [
+ "▁persona",
+ "l"
+ ],
+ [
+ "▁",
+ "personal"
+ ],
+ [
+ "]",
+ "["
+ ],
+ [
+ "ro",
+ "wn"
+ ],
+ [
+ "row",
+ "n"
+ ],
+ [
+ "r",
+ "own"
+ ],
+ [
+ "P",
+ "h"
+ ],
+ [
+ "▁f",
+ "oot"
+ ],
+ [
+ "▁fo",
+ "ot"
+ ],
+ [
+ "▁foo",
+ "t"
+ ],
+ [
+ "▁",
+ "foot"
+ ],
+ [
+ "▁re",
+ "search"
+ ],
+ [
+ "▁res",
+ "earch"
+ ],
+ [
+ "iron",
+ "ment"
+ ],
+ [
+ "▁n",
+ "om"
+ ],
+ [
+ "▁no",
+ "m"
+ ],
+ [
+ "▁",
+ "nom"
+ ],
+ [
+ "▁in",
+ "stance"
+ ],
+ [
+ "▁inst",
+ "ance"
+ ],
+ [
+ "▁",
+ "instance"
+ ],
+ [
+ "▁h",
+ "eld"
+ ],
+ [
+ "▁he",
+ "ld"
+ ],
+ [
+ "▁hel",
+ "d"
+ ],
+ [
+ "▁",
+ "held"
+ ],
+ [
+ "D",
+ "e"
+ ],
+ [
+ "▁mem",
+ "bers"
+ ],
+ [
+ "▁member",
+ "s"
+ ],
+ [
+ "▁",
+ "members"
+ ],
+ [
+ "▁f",
+ "ire"
+ ],
+ [
+ "▁fi",
+ "re"
+ ],
+ [
+ "▁fir",
+ "e"
+ ],
+ [
+ "▁",
+ "fire"
+ ],
+ [
+ "▁hist",
+ "ory"
+ ],
+ [
+ "▁histor",
+ "y"
+ ],
+ [
+ "▁hi",
+ "story"
+ ],
+ [
+ "▁",
+ "history"
+ ],
+ [
+ "▁m",
+ "ap"
+ ],
+ [
+ "▁ma",
+ "p"
+ ],
+ [
+ "▁",
+ "map"
+ ],
+ [
+ "▁dis",
+ "cuss"
+ ],
+ [
+ "▁disc",
+ "uss"
+ ],
+ [
+ "▁e",
+ "spec"
+ ],
+ [
+ "▁es",
+ "pec"
+ ],
+ [
+ "▁esp",
+ "ec"
+ ],
+ [
+ "▁",
+ "espec"
+ ],
+ [
+ "▁t",
+ "aking"
+ ],
+ [
+ "▁tak",
+ "ing"
+ ],
+ [
+ "▁ta",
+ "king"
+ ],
+ [
+ "▁s",
+ "ervices"
+ ],
+ [
+ "▁serv",
+ "ices"
+ ],
+ [
+ "▁service",
+ "s"
+ ],
+ [
+ "▁",
+ "services"
+ ],
+ [
+ "▁ind",
+ "ust"
+ ],
+ [
+ "▁indu",
+ "st"
+ ],
+ [
+ "▁",
+ "indust"
+ ],
+ [
+ "ig",
+ "en"
+ ],
+ [
+ "ige",
+ "n"
+ ],
+ [
+ "i",
+ "gen"
+ ],
+ [
+ "▁A",
+ "ss"
+ ],
+ [
+ "▁As",
+ "s"
+ ],
+ [
+ "▁",
+ "Ass"
+ ],
+ [
+ "▁e",
+ "xpected"
+ ],
+ [
+ "▁ex",
+ "pected"
+ ],
+ [
+ "▁expect",
+ "ed"
+ ],
+ [
+ "▁",
+ "expected"
+ ],
+ [
+ "▁wur",
+ "de"
+ ],
+ [
+ "di",
+ "r"
+ ],
+ [
+ "d",
+ "ir"
+ ],
+ [
+ "▁a",
+ "mong"
+ ],
+ [
+ "▁am",
+ "ong"
+ ],
+ [
+ "▁s",
+ "ugg"
+ ],
+ [
+ "▁su",
+ "gg"
+ ],
+ [
+ "▁sug",
+ "g"
+ ],
+ [
+ "re",
+ "c"
+ ],
+ [
+ "r",
+ "ec"
+ ],
+ [
+ "In",
+ "ter"
+ ],
+ [
+ "Int",
+ "er"
+ ],
+ [
+ "bl",
+ "ock"
+ ],
+ [
+ "blo",
+ "ck"
+ ],
+ [
+ "b",
+ "lock"
+ ],
+ [
+ "▁R",
+ "ep"
+ ],
+ [
+ "▁Re",
+ "p"
+ ],
+ [
+ "▁",
+ "Rep"
+ ],
+ [
+ "▁p",
+ "ain"
+ ],
+ [
+ "▁pa",
+ "in"
+ ],
+ [
+ "▁f",
+ "ive"
+ ],
+ [
+ "▁fi",
+ "ve"
+ ],
+ [
+ "▁",
+ "five"
+ ],
+ [
+ "▁f",
+ "und"
+ ],
+ [
+ "▁fun",
+ "d"
+ ],
+ [
+ "▁fu",
+ "nd"
+ ],
+ [
+ "▁",
+ "fund"
+ ],
+ [
+ "ri",
+ "d"
+ ],
+ [
+ "r",
+ "id"
+ ],
+ [
+ "ar",
+ "row"
+ ],
+ [
+ "arr",
+ "ow"
+ ],
+ [
+ "▁t",
+ "reat"
+ ],
+ [
+ "▁tre",
+ "at"
+ ],
+ [
+ "▁he",
+ "ard"
+ ],
+ [
+ "▁hear",
+ "d"
+ ],
+ [
+ "▁de",
+ "term"
+ ],
+ [
+ "▁det",
+ "erm"
+ ],
+ [
+ "▁deter",
+ "m"
+ ],
+ [
+ "ic",
+ "ult"
+ ],
+ [
+ "▁s",
+ "ense"
+ ],
+ [
+ "▁sens",
+ "e"
+ ],
+ [
+ "▁sen",
+ "se"
+ ],
+ [
+ "es",
+ "e"
+ ],
+ [
+ "e",
+ "se"
+ ],
+ [
+ "F",
+ "un"
+ ],
+ [
+ "▁month",
+ "s"
+ ],
+ [
+ "▁mont",
+ "hs"
+ ],
+ [
+ "js",
+ "on"
+ ],
+ [
+ "j",
+ "son"
+ ],
+ [
+ ",",
+ "”"
+ ],
+ [
+ "T",
+ "I"
+ ],
+ [
+ "or",
+ "age"
+ ],
+ [
+ "ora",
+ "ge"
+ ],
+ [
+ "o",
+ "rage"
+ ],
+ [
+ "▁",
+ "У"
+ ],
+ [
+ "▁every",
+ "one"
+ ],
+ [
+ "▁c",
+ "los"
+ ],
+ [
+ "▁cl",
+ "os"
+ ],
+ [
+ "▁clo",
+ "s"
+ ],
+ [
+ "▁",
+ "clos"
+ ],
+ [
+ "ie",
+ "rs"
+ ],
+ [
+ "ier",
+ "s"
+ ],
+ [
+ "i",
+ "ers"
+ ],
+ [
+ "air",
+ "s"
+ ],
+ [
+ "ai",
+ "rs"
+ ],
+ [
+ "a",
+ "irs"
+ ],
+ [
+ "def",
+ "ine"
+ ],
+ [
+ "I",
+ "f"
+ ],
+ [
+ "os",
+ "p"
+ ],
+ [
+ "o",
+ "sp"
+ ],
+ [
+ "▁w",
+ "onder"
+ ],
+ [
+ "▁won",
+ "der"
+ ],
+ [
+ "▁wo",
+ "nder"
+ ],
+ [
+ "N",
+ "A"
+ ],
+ [
+ "qu",
+ "ery"
+ ],
+ [
+ "que",
+ "ry"
+ ],
+ [
+ "quer",
+ "y"
+ ],
+ [
+ "p",
+ "g"
+ ],
+ [
+ "it",
+ "es"
+ ],
+ [
+ "ite",
+ "s"
+ ],
+ [
+ "i",
+ "tes"
+ ],
+ [
+ "▁m",
+ "aterial"
+ ],
+ [
+ "▁mat",
+ "erial"
+ ],
+ [
+ "▁mate",
+ "rial"
+ ],
+ [
+ "▁mater",
+ "ial"
+ ],
+ [
+ "▁",
+ "material"
+ ],
+ [
+ "y",
+ "d"
+ ],
+ [
+ "Re",
+ "ad"
+ ],
+ [
+ "R",
+ "ead"
+ ],
+ [
+ "ht",
+ "ml"
+ ],
+ [
+ "h",
+ "tml"
+ ],
+ [
+ "T",
+ "E"
+ ],
+ [
+ "P",
+ "r"
+ ],
+ [
+ "^{",
+ "\\"
+ ],
+ [
+ "^",
+ "{\\"
+ ],
+ [
+ "▁g",
+ "ave"
+ ],
+ [
+ "▁ga",
+ "ve"
+ ],
+ [
+ "▁I",
+ "S"
+ ],
+ [
+ "▁",
+ "IS"
+ ],
+ [
+ "▁s",
+ "uggest"
+ ],
+ [
+ "▁sugg",
+ "est"
+ ],
+ [
+ "▁sug",
+ "gest"
+ ],
+ [
+ "Over",
+ "ride"
+ ],
+ [
+ "ro",
+ "du"
+ ],
+ [
+ "rod",
+ "u"
+ ],
+ [
+ "Fr",
+ "om"
+ ],
+ [
+ "F",
+ "rom"
+ ],
+ [
+ "▁E",
+ "urope"
+ ],
+ [
+ "▁Europ",
+ "e"
+ ],
+ [
+ "▁Euro",
+ "pe"
+ ],
+ [
+ "▁",
+ "Europe"
+ ],
+ [
+ "P",
+ "O"
+ ],
+ [
+ "▁s",
+ "oon"
+ ],
+ [
+ "▁so",
+ "on"
+ ],
+ [
+ "ho",
+ "st"
+ ],
+ [
+ "hos",
+ "t"
+ ],
+ [
+ "h",
+ "ost"
+ ],
+ [
+ "▁B",
+ "er"
+ ],
+ [
+ "▁Be",
+ "r"
+ ],
+ [
+ "▁",
+ "Ber"
+ ],
+ [
+ "..",
+ ".."
+ ],
+ [
+ "...",
+ "."
+ ],
+ [
+ ".",
+ "..."
+ ],
+ [
+ "▁H",
+ "ar"
+ ],
+ [
+ "▁Ha",
+ "r"
+ ],
+ [
+ "▁",
+ "Har"
+ ],
+ [
+ "▁e",
+ "nergy"
+ ],
+ [
+ "▁ener",
+ "gy"
+ ],
+ [
+ "▁energ",
+ "y"
+ ],
+ [
+ "▁",
+ "energy"
+ ],
+ [
+ ">",
+ "<"
+ ],
+ [
+ "ave",
+ "s"
+ ],
+ [
+ "av",
+ "es"
+ ],
+ [
+ "a",
+ "ves"
+ ],
+ [
+ "▁e",
+ "asy"
+ ],
+ [
+ "▁eas",
+ "y"
+ ],
+ [
+ "▁b",
+ "re"
+ ],
+ [
+ "▁br",
+ "e"
+ ],
+ [
+ "▁",
+ "bre"
+ ],
+ [
+ "fr",
+ "ame"
+ ],
+ [
+ "▁g",
+ "round"
+ ],
+ [
+ "▁gr",
+ "ound"
+ ],
+ [
+ "▁gro",
+ "und"
+ ],
+ [
+ "▁",
+ "ground"
+ ],
+ [
+ "wi",
+ "th"
+ ],
+ [
+ "w",
+ "ith"
+ ],
+ [
+ "▁in",
+ "side"
+ ],
+ [
+ "▁ins",
+ "ide"
+ ],
+ [
+ "ie",
+ "f"
+ ],
+ [
+ "i",
+ "ef"
+ ],
+ [
+ "▁m",
+ "o"
+ ],
+ [
+ "▁",
+ "mo"
+ ],
+ [
+ "p",
+ "m"
+ ],
+ [
+ "pa",
+ "n"
+ ],
+ [
+ "p",
+ "an"
+ ],
+ [
+ "ig",
+ "r"
+ ],
+ [
+ "i",
+ "gr"
+ ],
+ [
+ "▁o",
+ "m"
+ ],
+ [
+ "▁",
+ "om"
+ ],
+ [
+ "ne",
+ "xt"
+ ],
+ [
+ "nex",
+ "t"
+ ],
+ [
+ "n",
+ "ext"
+ ],
+ [
+ "om",
+ "et"
+ ],
+ [
+ "ome",
+ "t"
+ ],
+ [
+ "o",
+ "met"
+ ],
+ [
+ "▁st",
+ "atus"
+ ],
+ [
+ "▁stat",
+ "us"
+ ],
+ [
+ "▁",
+ "status"
+ ],
+ [
+ "▁}",
+ "\r"
+ ],
+ [
+ "▁",
+ "}\r"
+ ],
+ [
+ "▁mus",
+ "ic"
+ ],
+ [
+ "or",
+ "a"
+ ],
+ [
+ "o",
+ "ra"
+ ],
+ [
+ "il",
+ "es"
+ ],
+ [
+ "ile",
+ "s"
+ ],
+ [
+ "i",
+ "les"
+ ],
+ [
+ "k",
+ "i"
+ ],
+ [
+ "▁e",
+ "sc"
+ ],
+ [
+ "▁es",
+ "c"
+ ],
+ [
+ "▁",
+ "esc"
+ ],
+ [
+ "▁b",
+ "es"
+ ],
+ [
+ "▁be",
+ "s"
+ ],
+ [
+ "▁",
+ "bes"
+ ],
+ [
+ "▁D",
+ "is"
+ ],
+ [
+ "▁Di",
+ "s"
+ ],
+ [
+ "▁",
+ "Dis"
+ ],
+ [
+ "▁h",
+ "ost"
+ ],
+ [
+ "▁ho",
+ "st"
+ ],
+ [
+ "▁",
+ "host"
+ ],
+ [
+ "▁c",
+ "omes"
+ ],
+ [
+ "▁com",
+ "es"
+ ],
+ [
+ "▁co",
+ "mes"
+ ],
+ [
+ "▁come",
+ "s"
+ ],
+ [
+ "▁",
+ "comes"
+ ],
+ [
+ "us",
+ "ed"
+ ],
+ [
+ "use",
+ "d"
+ ],
+ [
+ "u",
+ "sed"
+ ],
+ [
+ "▁f",
+ "uture"
+ ],
+ [
+ "▁fut",
+ "ure"
+ ],
+ [
+ "▁",
+ "future"
+ ],
+ [
+ "lic",
+ "k"
+ ],
+ [
+ "li",
+ "ck"
+ ],
+ [
+ "l",
+ "ick"
+ ],
+ [
+ "ai",
+ "d"
+ ],
+ [
+ "a",
+ "id"
+ ],
+ [
+ "▁com",
+ "pet"
+ ],
+ [
+ "▁comp",
+ "et"
+ ],
+ [
+ "▁",
+ "compet"
+ ],
+ [
+ "▁v",
+ "oice"
+ ],
+ [
+ "▁vo",
+ "ice"
+ ],
+ [
+ "▁",
+ "voice"
+ ],
+ [
+ "▁l",
+ "oad"
+ ],
+ [
+ "▁lo",
+ "ad"
+ ],
+ [
+ "▁",
+ "load"
+ ],
+ [
+ "ev",
+ "el"
+ ],
+ [
+ "eve",
+ "l"
+ ],
+ [
+ "e",
+ "vel"
+ ],
+ [
+ "▁n",
+ "eg"
+ ],
+ [
+ "▁ne",
+ "g"
+ ],
+ [
+ "▁",
+ "neg"
+ ],
+ [
+ "▁com",
+ "mand"
+ ],
+ [
+ "▁comm",
+ "and"
+ ],
+ [
+ "▁",
+ "command"
+ ],
+ [
+ "▁f",
+ "ür"
+ ],
+ [
+ "▁p",
+ "ie"
+ ],
+ [
+ "▁pi",
+ "e"
+ ],
+ [
+ "▁",
+ "pie"
+ ],
+ [
+ "▁qu",
+ "ite"
+ ],
+ [
+ "▁qui",
+ "te"
+ ],
+ [
+ "▁quit",
+ "e"
+ ],
+ [
+ "▁b",
+ "lo"
+ ],
+ [
+ "▁bl",
+ "o"
+ ],
+ [
+ "▁",
+ "blo"
+ ],
+ [
+ "ag",
+ "n"
+ ],
+ [
+ "a",
+ "gn"
+ ],
+ [
+ "il",
+ "on"
+ ],
+ [
+ "ilo",
+ "n"
+ ],
+ [
+ "i",
+ "lon"
+ ],
+ [
+ "▁cl",
+ "aim"
+ ],
+ [
+ "▁",
+ "claim"
+ ],
+ [
+ "▁t",
+ "each"
+ ],
+ [
+ "▁te",
+ "ach"
+ ],
+ [
+ "▁tea",
+ "ch"
+ ],
+ [
+ "▁pre",
+ "vious"
+ ],
+ [
+ "▁prev",
+ "ious"
+ ],
+ [
+ "▁",
+ "previous"
+ ],
+ [
+ "▁s",
+ "ite"
+ ],
+ [
+ "▁sit",
+ "e"
+ ],
+ [
+ "▁si",
+ "te"
+ ],
+ [
+ "▁",
+ "site"
+ ],
+ [
+ "co",
+ "lor"
+ ],
+ [
+ "col",
+ "or"
+ ],
+ [
+ "colo",
+ "r"
+ ],
+ [
+ "at",
+ "tr"
+ ],
+ [
+ "att",
+ "r"
+ ],
+ [
+ "▁ac",
+ "cept"
+ ],
+ [
+ "▁",
+ "accept"
+ ],
+ [
+ "▁ex",
+ "act"
+ ],
+ [
+ ")",
+ "}"
+ ],
+ [
+ "af",
+ "t"
+ ],
+ [
+ "a",
+ "ft"
+ ],
+ [
+ "rol",
+ "ler"
+ ],
+ [
+ "roll",
+ "er"
+ ],
+ [
+ "о",
+ "н"
+ ],
+ [
+ "o",
+ "o"
+ ],
+ [
+ "Dat",
+ "e"
+ ],
+ [
+ "Da",
+ "te"
+ ],
+ [
+ "D",
+ "ate"
+ ],
+ [
+ "▁o",
+ "u"
+ ],
+ [
+ "▁",
+ "ou"
+ ],
+ [
+ "s",
+ "y"
+ ],
+ [
+ "▁pre",
+ "tty"
+ ],
+ [
+ "▁pret",
+ "ty"
+ ],
+ [
+ "▁im",
+ "age"
+ ],
+ [
+ "▁imag",
+ "e"
+ ],
+ [
+ "▁",
+ "image"
+ ],
+ [
+ "B",
+ "U"
+ ],
+ [
+ "▁term",
+ "s"
+ ],
+ [
+ "▁ter",
+ "ms"
+ ],
+ [
+ "▁s",
+ "earch"
+ ],
+ [
+ "▁se",
+ "arch"
+ ],
+ [
+ "▁sear",
+ "ch"
+ ],
+ [
+ "▁",
+ "search"
+ ],
+ [
+ "▁",
+ "è"
+ ],
+ [
+ "▁V",
+ "al"
+ ],
+ [
+ "▁Va",
+ "l"
+ ],
+ [
+ "▁",
+ "Val"
+ ],
+ [
+ "▁",
+ "‘"
+ ],
+ [
+ "▁D",
+ "av"
+ ],
+ [
+ "▁Da",
+ "v"
+ ],
+ [
+ "M",
+ "S"
+ ],
+ [
+ "sr",
+ "c"
+ ],
+ [
+ "s",
+ "rc"
+ ],
+ [
+ "ma",
+ "r"
+ ],
+ [
+ "m",
+ "ar"
+ ],
+ [
+ "in",
+ "cip"
+ ],
+ [
+ "inc",
+ "ip"
+ ],
+ [
+ "▁could",
+ "n"
+ ],
+ [
+ "ad",
+ "os"
+ ],
+ [
+ "ado",
+ "s"
+ ],
+ [
+ "▁d",
+ "ro"
+ ],
+ [
+ "▁dr",
+ "o"
+ ],
+ [
+ "▁",
+ "dro"
+ ],
+ [
+ "be",
+ "ta"
+ ],
+ [
+ "bet",
+ "a"
+ ],
+ [
+ "b",
+ "eta"
+ ],
+ [
+ "im",
+ "um"
+ ],
+ [
+ "▁min",
+ "utes"
+ ],
+ [
+ "▁minute",
+ "s"
+ ],
+ [
+ "▁minut",
+ "es"
+ ],
+ [
+ "▁g",
+ "rand"
+ ],
+ [
+ "▁gr",
+ "and"
+ ],
+ [
+ "▁gran",
+ "d"
+ ],
+ [
+ "▁gra",
+ "nd"
+ ],
+ [
+ "▁",
+ "grand"
+ ],
+ [
+ "▁",
+ "»"
+ ],
+ [
+ "▁O",
+ "ur"
+ ],
+ [
+ "▁",
+ "Our"
+ ],
+ [
+ "St",
+ "r"
+ ],
+ [
+ "S",
+ "tr"
+ ],
+ [
+ "VE",
+ "R"
+ ],
+ [
+ "V",
+ "ER"
+ ],
+ [
+ "ma",
+ "z"
+ ],
+ [
+ "m",
+ "az"
+ ],
+ [
+ "▁or",
+ "iginal"
+ ],
+ [
+ "▁orig",
+ "inal"
+ ],
+ [
+ "▁origin",
+ "al"
+ ],
+ [
+ "▁",
+ "original"
+ ],
+ [
+ "in",
+ "i"
+ ],
+ [
+ "i",
+ "ni"
+ ],
+ [
+ "▁c",
+ "oll"
+ ],
+ [
+ "▁col",
+ "l"
+ ],
+ [
+ "▁co",
+ "ll"
+ ],
+ [
+ "▁",
+ "coll"
+ ],
+ [
+ "lo",
+ "at"
+ ],
+ [
+ "▁o",
+ "s"
+ ],
+ [
+ "▁",
+ "os"
+ ],
+ [
+ "})",
+ ";"
+ ],
+ [
+ "}",
+ ");"
+ ],
+ [
+ "sum",
+ "mary"
+ ],
+ [
+ "▁w",
+ "all"
+ ],
+ [
+ "▁wa",
+ "ll"
+ ],
+ [
+ "▁wal",
+ "l"
+ ],
+ [
+ "▁",
+ "wall"
+ ],
+ [
+ "Col",
+ "or"
+ ],
+ [
+ "Co",
+ "lor"
+ ],
+ [
+ "▁v",
+ "ers"
+ ],
+ [
+ "▁ver",
+ "s"
+ ],
+ [
+ "▁ve",
+ "rs"
+ ],
+ [
+ "▁",
+ "vers"
+ ],
+ [
+ "▁d",
+ "ella"
+ ],
+ [
+ "▁de",
+ "lla"
+ ],
+ [
+ "▁del",
+ "la"
+ ],
+ [
+ "▁dell",
+ "a"
+ ],
+ [
+ "▁\"",
+ "\"\""
+ ],
+ [
+ "▁\"\"",
+ "\""
+ ],
+ [
+ "▁",
+ "\"\"\""
+ ],
+ [
+ "math",
+ "bf"
+ ],
+ [
+ "ze",
+ "r"
+ ],
+ [
+ "z",
+ "er"
+ ],
+ [
+ "au",
+ "r"
+ ],
+ [
+ "a",
+ "ur"
+ ],
+ [
+ "▁tr",
+ "ack"
+ ],
+ [
+ "▁tra",
+ "ck"
+ ],
+ [
+ "▁",
+ "track"
+ ],
+ [
+ "▁ass",
+ "oci"
+ ],
+ [
+ "▁",
+ "associ"
+ ],
+ [
+ "▁s",
+ "uff"
+ ],
+ [
+ "▁su",
+ "ff"
+ ],
+ [
+ "▁in",
+ "de"
+ ],
+ [
+ "▁i",
+ "nde"
+ ],
+ [
+ "▁ind",
+ "e"
+ ],
+ [
+ "▁",
+ "inde"
+ ],
+ [
+ "ag",
+ "ue"
+ ],
+ [
+ "agu",
+ "e"
+ ],
+ [
+ "a",
+ "gue"
+ ],
+ [
+ "▁A",
+ "pr"
+ ],
+ [
+ "▁Ap",
+ "r"
+ ],
+ [
+ "▁",
+ "Apr"
+ ],
+ [
+ "L",
+ "e"
+ ],
+ [
+ "ro",
+ "ups"
+ ],
+ [
+ "rou",
+ "ps"
+ ],
+ [
+ "roup",
+ "s"
+ ],
+ [
+ "bo",
+ "ard"
+ ],
+ [
+ "b",
+ "oard"
+ ],
+ [
+ "▁att",
+ "ack"
+ ],
+ [
+ "▁s",
+ "eries"
+ ],
+ [
+ "▁se",
+ "ries"
+ ],
+ [
+ "▁ser",
+ "ies"
+ ],
+ [
+ "▁serie",
+ "s"
+ ],
+ [
+ "▁",
+ "series"
+ ],
+ [
+ "▁in",
+ "stead"
+ ],
+ [
+ "▁inst",
+ "ead"
+ ],
+ [
+ "ha",
+ "m"
+ ],
+ [
+ "h",
+ "am"
+ ],
+ [
+ "bo",
+ "ok"
+ ],
+ [
+ "b",
+ "ook"
+ ],
+ [
+ "▁s",
+ "ix"
+ ],
+ [
+ "▁si",
+ "x"
+ ],
+ [
+ "▁",
+ "six"
+ ],
+ [
+ "▁R",
+ "ec"
+ ],
+ [
+ "▁Re",
+ "c"
+ ],
+ [
+ "▁",
+ "Rec"
+ ],
+ [
+ "▁c",
+ "oming"
+ ],
+ [
+ "▁com",
+ "ing"
+ ],
+ [
+ "▁co",
+ "ming"
+ ],
+ [
+ "▁",
+ "coming"
+ ],
+ [
+ "ur",
+ "t"
+ ],
+ [
+ "u",
+ "rt"
+ ],
+ [
+ "▁gl",
+ "obal"
+ ],
+ [
+ "▁glob",
+ "al"
+ ],
+ [
+ "▁glo",
+ "bal"
+ ],
+ [
+ "▁",
+ "global"
+ ],
+ [
+ "▁ne",
+ "cess"
+ ],
+ [
+ "▁neces",
+ "s"
+ ],
+ [
+ "▁",
+ "necess"
+ ],
+ [
+ "le",
+ "ge"
+ ],
+ [
+ "leg",
+ "e"
+ ],
+ [
+ "Po",
+ "s"
+ ],
+ [
+ "P",
+ "os"
+ ],
+ [
+ "▁le",
+ "ave"
+ ],
+ [
+ "▁",
+ "leave"
+ ],
+ [
+ "▁p",
+ "od"
+ ],
+ [
+ "▁po",
+ "d"
+ ],
+ [
+ "▁",
+ "pod"
+ ],
+ [
+ "ateg",
+ "ory"
+ ],
+ [
+ "ategor",
+ "y"
+ ],
+ [
+ "u",
+ "z"
+ ],
+ [
+ "▁de",
+ "ep"
+ ],
+ [
+ "▁",
+ "deep"
+ ],
+ [
+ "▁k",
+ "m"
+ ],
+ [
+ "▁",
+ "km"
+ ],
+ [
+ "▁out",
+ "side"
+ ],
+ [
+ "▁outs",
+ "ide"
+ ],
+ [
+ "ha",
+ "s"
+ ],
+ [
+ "h",
+ "as"
+ ],
+ [
+ "opt",
+ "ions"
+ ],
+ [
+ "option",
+ "s"
+ ],
+ [
+ "o",
+ "ptions"
+ ],
+ [
+ "▁S",
+ "m"
+ ],
+ [
+ "▁",
+ "Sm"
+ ],
+ [
+ "Su",
+ "b"
+ ],
+ [
+ "S",
+ "ub"
+ ],
+ [
+ "ro",
+ "ws"
+ ],
+ [
+ "row",
+ "s"
+ ],
+ [
+ "r",
+ "ows"
+ ],
+ [
+ "▁в",
+ "и"
+ ],
+ [
+ "▁",
+ "ви"
+ ],
+ [
+ "▁St",
+ "ates"
+ ],
+ [
+ "▁State",
+ "s"
+ ],
+ [
+ "▁Stat",
+ "es"
+ ],
+ [
+ "▁Sta",
+ "tes"
+ ],
+ [
+ "▁",
+ "States"
+ ],
+ [
+ "▁wr",
+ "ong"
+ ],
+ [
+ "▁how",
+ "ever"
+ ],
+ [
+ "▁s",
+ "em"
+ ],
+ [
+ "▁se",
+ "m"
+ ],
+ [
+ "▁",
+ "sem"
+ ],
+ [
+ "▁c",
+ "atch"
+ ],
+ [
+ "▁cat",
+ "ch"
+ ],
+ [
+ "▁",
+ "catch"
+ ],
+ [
+ "\")",
+ ","
+ ],
+ [
+ "\"",
+ "),"
+ ],
+ [
+ "mod",
+ "el"
+ ],
+ [
+ "mode",
+ "l"
+ ],
+ [
+ "mo",
+ "del"
+ ],
+ [
+ "▁h",
+ "ttp"
+ ],
+ [
+ "▁htt",
+ "p"
+ ],
+ [
+ "▁",
+ "http"
+ ],
+ [
+ "▁o",
+ "ption"
+ ],
+ [
+ "▁opt",
+ "ion"
+ ],
+ [
+ "▁",
+ "option"
+ ],
+ [
+ "ri",
+ "e"
+ ],
+ [
+ "r",
+ "ie"
+ ],
+ [
+ "▁с",
+ "та"
+ ],
+ [
+ "▁ст",
+ "а"
+ ],
+ [
+ "▁",
+ "ста"
+ ],
+ [
+ "▁ä",
+ "r"
+ ],
+ [
+ "▁",
+ "är"
+ ],
+ [
+ "▁en",
+ "joy"
+ ],
+ [
+ "▁enjo",
+ "y"
+ ],
+ [
+ "n",
+ "u"
+ ],
+ [
+ "▁p",
+ "as"
+ ],
+ [
+ "▁pa",
+ "s"
+ ],
+ [
+ "▁",
+ "pas"
+ ],
+ [
+ "▁a",
+ "mount"
+ ],
+ [
+ "▁am",
+ "ount"
+ ],
+ [
+ "▁",
+ "amount"
+ ],
+ [
+ "▁res",
+ "pons"
+ ],
+ [
+ "▁respon",
+ "s"
+ ],
+ [
+ "▁resp",
+ "ons"
+ ],
+ [
+ "▁",
+ "respons"
+ ],
+ [
+ "▁In",
+ "tern"
+ ],
+ [
+ "▁Inter",
+ "n"
+ ],
+ [
+ "▁Int",
+ "ern"
+ ],
+ [
+ "▁",
+ "Intern"
+ ],
+ [
+ "▁my",
+ "self"
+ ],
+ [
+ "▁o",
+ "pp"
+ ],
+ [
+ "▁op",
+ "p"
+ ],
+ [
+ "▁",
+ "opp"
+ ],
+ [
+ "▁S",
+ "im"
+ ],
+ [
+ "▁Si",
+ "m"
+ ],
+ [
+ "▁",
+ "Sim"
+ ],
+ [
+ "▁s",
+ "ens"
+ ],
+ [
+ "▁se",
+ "ns"
+ ],
+ [
+ "▁sen",
+ "s"
+ ],
+ [
+ "E",
+ "d"
+ ],
+ [
+ "▁(",
+ "\\"
+ ],
+ [
+ "▁",
+ "(\\"
+ ],
+ [
+ "▁stud",
+ "ents"
+ ],
+ [
+ "▁student",
+ "s"
+ ],
+ [
+ "но",
+ "в"
+ ],
+ [
+ "н",
+ "ов"
+ ],
+ [
+ "▁point",
+ "s"
+ ],
+ [
+ "▁",
+ "points"
+ ],
+ [
+ "ar",
+ "ning"
+ ],
+ [
+ "arn",
+ "ing"
+ ],
+ [
+ "U",
+ "P"
+ ],
+ [
+ "el",
+ "ling"
+ ],
+ [
+ "ell",
+ "ing"
+ ],
+ [
+ "elli",
+ "ng"
+ ],
+ [
+ "▁c",
+ "annot"
+ ],
+ [
+ "▁can",
+ "not"
+ ],
+ [
+ "B",
+ "e"
+ ],
+ [
+ "▁l",
+ "ength"
+ ],
+ [
+ "▁le",
+ "ngth"
+ ],
+ [
+ "▁",
+ "length"
+ ],
+ [
+ "nu",
+ "ll"
+ ],
+ [
+ "n",
+ "ull"
+ ],
+ [
+ "ui",
+ "nt"
+ ],
+ [
+ "u",
+ "int"
+ ],
+ [
+ "wi",
+ "se"
+ ],
+ [
+ "w",
+ "ise"
+ ],
+ [
+ "▁d",
+ "ouble"
+ ],
+ [
+ "▁dou",
+ "ble"
+ ],
+ [
+ "▁doub",
+ "le"
+ ],
+ [
+ "▁",
+ "double"
+ ],
+ [
+ "ig",
+ "e"
+ ],
+ [
+ "i",
+ "ge"
+ ],
+ [
+ "is",
+ "ta"
+ ],
+ [
+ "ist",
+ "a"
+ ],
+ [
+ "i",
+ "sta"
+ ],
+ [
+ "▁est",
+ "ab"
+ ],
+ [
+ "▁es",
+ "tab"
+ ],
+ [
+ "▁esta",
+ "b"
+ ],
+ [
+ "an",
+ "ch"
+ ],
+ [
+ "anc",
+ "h"
+ ],
+ [
+ "▁a",
+ "go"
+ ],
+ [
+ "▁ag",
+ "o"
+ ],
+ [
+ "▁",
+ "ago"
+ ],
+ [
+ "▁b",
+ "ound"
+ ],
+ [
+ "▁bo",
+ "und"
+ ],
+ [
+ "▁bou",
+ "nd"
+ ],
+ [
+ "▁",
+ "bound"
+ ],
+ [
+ "▁f",
+ "a"
+ ],
+ [
+ "▁",
+ "fa"
+ ],
+ [
+ "▁c",
+ "lean"
+ ],
+ [
+ "▁cle",
+ "an"
+ ],
+ [
+ "▁",
+ "clean"
+ ],
+ [
+ "▁sim",
+ "ple"
+ ],
+ [
+ "▁simpl",
+ "e"
+ ],
+ [
+ "▁",
+ "simple"
+ ],
+ [
+ "m",
+ "i"
+ ],
+ [
+ "####",
+ "####"
+ ],
+ [
+ "if",
+ "ier"
+ ],
+ [
+ "ifi",
+ "er"
+ ],
+ [
+ "▁Gener",
+ "al"
+ ],
+ [
+ "▁Gen",
+ "eral"
+ ],
+ [
+ "▁Gene",
+ "ral"
+ ],
+ [
+ "▁",
+ "General"
+ ],
+ [
+ "▁se",
+ "emed"
+ ],
+ [
+ "▁see",
+ "med"
+ ],
+ [
+ "▁seem",
+ "ed"
+ ],
+ [
+ "en",
+ "a"
+ ],
+ [
+ "e",
+ "na"
+ ],
+ [
+ "▁a",
+ "ge"
+ ],
+ [
+ "▁ag",
+ "e"
+ ],
+ [
+ "▁",
+ "age"
+ ],
+ [
+ "но",
+ "й"
+ ],
+ [
+ "end",
+ "if"
+ ],
+ [
+ "A",
+ "A"
+ ],
+ [
+ "▁c",
+ "aus"
+ ],
+ [
+ "▁ca",
+ "us"
+ ],
+ [
+ "▁e",
+ "duc"
+ ],
+ [
+ "▁ed",
+ "uc"
+ ],
+ [
+ "▁",
+ "educ"
+ ],
+ [
+ "▁c",
+ "ell"
+ ],
+ [
+ "▁ce",
+ "ll"
+ ],
+ [
+ "▁cel",
+ "l"
+ ],
+ [
+ "▁",
+ "cell"
+ ],
+ [
+ "Ge",
+ "ner"
+ ],
+ [
+ "Gen",
+ "er"
+ ],
+ [
+ "G",
+ "ener"
+ ],
+ [
+ "sp",
+ "ace"
+ ],
+ [
+ "s",
+ "pace"
+ ],
+ [
+ "▁Y",
+ "our"
+ ],
+ [
+ "▁You",
+ "r"
+ ],
+ [
+ "▁",
+ "Your"
+ ],
+ [
+ "▁be",
+ "aut"
+ ],
+ [
+ "g",
+ "t"
+ ],
+ [
+ "▁l",
+ "imit"
+ ],
+ [
+ "▁li",
+ "mit"
+ ],
+ [
+ "▁lim",
+ "it"
+ ],
+ [
+ "▁",
+ "limit"
+ ],
+ [
+ "▁d",
+ "ate"
+ ],
+ [
+ "▁da",
+ "te"
+ ],
+ [
+ "▁dat",
+ "e"
+ ],
+ [
+ "▁",
+ "date"
+ ],
+ [
+ "Ut",
+ "il"
+ ],
+ [
+ "U",
+ "til"
+ ],
+ [
+ "▁N",
+ "ational"
+ ],
+ [
+ "▁Nat",
+ "ional"
+ ],
+ [
+ "▁Nation",
+ "al"
+ ],
+ [
+ "▁",
+ "National"
+ ],
+ [
+ "ow",
+ "s"
+ ],
+ [
+ "o",
+ "ws"
+ ],
+ [
+ "pa",
+ "t"
+ ],
+ [
+ "p",
+ "at"
+ ],
+ [
+ "qu",
+ "ad"
+ ],
+ [
+ "▁o",
+ "k"
+ ],
+ [
+ "▁",
+ "ok"
+ ],
+ [
+ "▁",
+ "И"
+ ],
+ [
+ "ar",
+ "th"
+ ],
+ [
+ "art",
+ "h"
+ ],
+ [
+ "ha",
+ "t"
+ ],
+ [
+ "h",
+ "at"
+ ],
+ [
+ "▁comm",
+ "unity"
+ ],
+ [
+ "▁commun",
+ "ity"
+ ],
+ [
+ "ou",
+ "l"
+ ],
+ [
+ "o",
+ "ul"
+ ],
+ [
+ "▁e",
+ "conom"
+ ],
+ [
+ "▁ec",
+ "onom"
+ ],
+ [
+ "▁",
+ "econom"
+ ],
+ [
+ "Com",
+ "ponent"
+ ],
+ [
+ "bo",
+ "r"
+ ],
+ [
+ "b",
+ "or"
+ ],
+ [
+ "us",
+ "ion"
+ ],
+ [
+ "▁be",
+ "low"
+ ],
+ [
+ "▁bel",
+ "ow"
+ ],
+ [
+ "ear",
+ "ch"
+ ],
+ [
+ "e",
+ "arch"
+ ],
+ [
+ "or",
+ "es"
+ ],
+ [
+ "ore",
+ "s"
+ ],
+ [
+ "o",
+ "res"
+ ],
+ [
+ "ba",
+ "n"
+ ],
+ [
+ "b",
+ "an"
+ ],
+ [
+ "▁Aug",
+ "ust"
+ ],
+ [
+ "▁fur",
+ "ther"
+ ],
+ [
+ "sig",
+ "ma"
+ ],
+ [
+ "s",
+ "igma"
+ ],
+ [
+ "▁h",
+ "a"
+ ],
+ [
+ "▁",
+ "ha"
+ ],
+ [
+ "j",
+ "i"
+ ],
+ [
+ "▁com",
+ "put"
+ ],
+ [
+ "▁comp",
+ "ut"
+ ],
+ [
+ "▁",
+ "comput"
+ ],
+ [
+ "г",
+ "ра"
+ ],
+ [
+ "▁N",
+ "one"
+ ],
+ [
+ "▁No",
+ "ne"
+ ],
+ [
+ "▁Non",
+ "e"
+ ],
+ [
+ "▁",
+ "None"
+ ],
+ [
+ "▁t",
+ "er"
+ ],
+ [
+ "▁te",
+ "r"
+ ],
+ [
+ "▁",
+ "ter"
+ ],
+ [
+ "▁any",
+ "one"
+ ],
+ [
+ "▁t",
+ "ask"
+ ],
+ [
+ "▁ta",
+ "sk"
+ ],
+ [
+ "▁",
+ "task"
+ ],
+ [
+ "en",
+ "te"
+ ],
+ [
+ "ent",
+ "e"
+ ],
+ [
+ "e",
+ "nte"
+ ],
+ [
+ "pos",
+ "ition"
+ ],
+ [
+ "pp",
+ "ed"
+ ],
+ [
+ "ppe",
+ "d"
+ ],
+ [
+ "p",
+ "ped"
+ ],
+ [
+ "▁a",
+ "us"
+ ],
+ [
+ "▁au",
+ "s"
+ ],
+ [
+ "▁",
+ "aus"
+ ],
+ [
+ "Att",
+ "ribute"
+ ],
+ [
+ "Attrib",
+ "ute"
+ ],
+ [
+ "re",
+ "q"
+ ],
+ [
+ "r",
+ "eq"
+ ],
+ [
+ "ad",
+ "dr"
+ ],
+ [
+ "add",
+ "r"
+ ],
+ [
+ "li",
+ "ght"
+ ],
+ [
+ "lig",
+ "ht"
+ ],
+ [
+ "l",
+ "ight"
+ ],
+ [
+ "ш",
+ "е"
+ ],
+ [
+ "▁a",
+ "rm"
+ ],
+ [
+ "▁ar",
+ "m"
+ ],
+ [
+ "▁",
+ "arm"
+ ],
+ [
+ "co",
+ "ver"
+ ],
+ [
+ "cov",
+ "er"
+ ],
+ [
+ "c",
+ "over"
+ ],
+ [
+ "up",
+ "port"
+ ],
+ [
+ "upp",
+ "ort"
+ ],
+ [
+ "▁G",
+ "l"
+ ],
+ [
+ "▁",
+ "Gl"
+ ],
+ [
+ "▁S",
+ "an"
+ ],
+ [
+ "▁Sa",
+ "n"
+ ],
+ [
+ "▁",
+ "San"
+ ],
+ [
+ "▁wr",
+ "iting"
+ ],
+ [
+ "▁writ",
+ "ing"
+ ],
+ [
+ "▁",
+ "writing"
+ ],
+ [
+ "▁l",
+ "ost"
+ ],
+ [
+ "▁lo",
+ "st"
+ ],
+ [
+ "▁los",
+ "t"
+ ],
+ [
+ "▁M",
+ "ark"
+ ],
+ [
+ "▁Mar",
+ "k"
+ ],
+ [
+ "▁",
+ "Mark"
+ ],
+ [
+ "▁g",
+ "re"
+ ],
+ [
+ "▁gr",
+ "e"
+ ],
+ [
+ "▁",
+ "gre"
+ ],
+ [
+ "TY",
+ "PE"
+ ],
+ [
+ "T",
+ "YPE"
+ ],
+ [
+ "▁S",
+ "outh"
+ ],
+ [
+ "▁So",
+ "uth"
+ ],
+ [
+ "▁Sou",
+ "th"
+ ],
+ [
+ "▁Sout",
+ "h"
+ ],
+ [
+ "▁",
+ "South"
+ ],
+ [
+ "▁per",
+ "fect"
+ ],
+ [
+ "▁perf",
+ "ect"
+ ],
+ [
+ "▁pack",
+ "age"
+ ],
+ [
+ "▁",
+ "package"
+ ],
+ [
+ "▁in",
+ "fl"
+ ],
+ [
+ "▁inf",
+ "l"
+ ],
+ [
+ "▁",
+ "infl"
+ ],
+ [
+ "ha",
+ "ps"
+ ],
+ [
+ "h",
+ "aps"
+ ],
+ [
+ "▁A",
+ "ng"
+ ],
+ [
+ "▁An",
+ "g"
+ ],
+ [
+ "▁",
+ "Ang"
+ ],
+ [
+ "res",
+ "pon"
+ ],
+ [
+ "resp",
+ "on"
+ ],
+ [
+ "ri",
+ "s"
+ ],
+ [
+ "r",
+ "is"
+ ],
+ [
+ "pt",
+ "ember"
+ ],
+ [
+ "pte",
+ "mber"
+ ],
+ [
+ "▁build",
+ "ing"
+ ],
+ [
+ "▁",
+ "building"
+ ],
+ [
+ "VA",
+ "L"
+ ],
+ [
+ "V",
+ "AL"
+ ],
+ [
+ "fr",
+ "ee"
+ ],
+ [
+ "fre",
+ "e"
+ ],
+ [
+ "f",
+ "ree"
+ ],
+ [
+ "▁c",
+ "e"
+ ],
+ [
+ "▁",
+ "ce"
+ ],
+ [
+ "H",
+ "T"
+ ],
+ [
+ "▁F",
+ "rom"
+ ],
+ [
+ "▁Fr",
+ "om"
+ ],
+ [
+ "▁Fro",
+ "m"
+ ],
+ [
+ "▁",
+ "From"
+ ],
+ [
+ "d",
+ "s"
+ ],
+ [
+ "ro",
+ "y"
+ ],
+ [
+ "r",
+ "oy"
+ ],
+ [
+ "ach",
+ "ine"
+ ],
+ [
+ "achi",
+ "ne"
+ ],
+ [
+ "no",
+ "wn"
+ ],
+ [
+ "now",
+ "n"
+ ],
+ [
+ "n",
+ "own"
+ ],
+ [
+ "▁sa",
+ "ying"
+ ],
+ [
+ "▁say",
+ "ing"
+ ],
+ [
+ "▁б",
+ "ы"
+ ],
+ [
+ "▁",
+ "бы"
+ ],
+ [
+ "o",
+ "e"
+ ],
+ [
+ "Re",
+ "f"
+ ],
+ [
+ "R",
+ "ef"
+ ],
+ [
+ "▁net",
+ "work"
+ ],
+ [
+ "▁",
+ "network"
+ ],
+ [
+ "par",
+ "ent"
+ ],
+ [
+ "pa",
+ "rent"
+ ],
+ [
+ "pare",
+ "nt"
+ ],
+ [
+ "paren",
+ "t"
+ ],
+ [
+ "p",
+ "arent"
+ ],
+ [
+ "ug",
+ "e"
+ ],
+ [
+ "u",
+ "ge"
+ ],
+ [
+ "▁sim",
+ "ilar"
+ ],
+ [
+ ">",
+ "\r"
+ ],
+ [
+ "Build",
+ "er"
+ ],
+ [
+ "B",
+ "uilder"
+ ],
+ [
+ "▁l",
+ "iving"
+ ],
+ [
+ "▁li",
+ "ving"
+ ],
+ [
+ "▁liv",
+ "ing"
+ ],
+ [
+ "▁contin",
+ "ue"
+ ],
+ [
+ "▁continu",
+ "e"
+ ],
+ [
+ "▁",
+ "continue"
+ ],
+ [
+ "an",
+ "ger"
+ ],
+ [
+ "ang",
+ "er"
+ ],
+ [
+ "ange",
+ "r"
+ ],
+ [
+ "▁R",
+ "ed"
+ ],
+ [
+ "▁Re",
+ "d"
+ ],
+ [
+ "▁",
+ "Red"
+ ],
+ [
+ "▁h",
+ "air"
+ ],
+ [
+ "▁ha",
+ "ir"
+ ],
+ [
+ "an",
+ "ced"
+ ],
+ [
+ "ance",
+ "d"
+ ],
+ [
+ "anc",
+ "ed"
+ ],
+ [
+ "ia",
+ "ns"
+ ],
+ [
+ "ian",
+ "s"
+ ],
+ [
+ "i",
+ "ans"
+ ],
+ [
+ "▁d",
+ "ead"
+ ],
+ [
+ "▁de",
+ "ad"
+ ],
+ [
+ "▁",
+ "dead"
+ ],
+ [
+ "▁bo",
+ "olean"
+ ],
+ [
+ "▁",
+ "boolean"
+ ],
+ [
+ "ic",
+ "ation"
+ ],
+ [
+ "▁д",
+ "е"
+ ],
+ [
+ "▁",
+ "де"
+ ],
+ [
+ "▁cl",
+ "ient"
+ ],
+ [
+ "▁",
+ "client"
+ ],
+ [
+ "uc",
+ "t"
+ ],
+ [
+ "u",
+ "ct"
+ ],
+ [
+ "▁",
+ "•"
+ ],
+ [
+ "S",
+ "P"
+ ],
+ [
+ "ol",
+ "der"
+ ],
+ [
+ "old",
+ "er"
+ ],
+ [
+ "п",
+ "е"
+ ],
+ [
+ "ud",
+ "io"
+ ],
+ [
+ "udi",
+ "o"
+ ],
+ [
+ "▁d",
+ "eg"
+ ],
+ [
+ "▁de",
+ "g"
+ ],
+ [
+ "▁",
+ "deg"
+ ],
+ [
+ "as",
+ "ing"
+ ],
+ [
+ "asi",
+ "ng"
+ ],
+ [
+ "a",
+ "sing"
+ ],
+ [
+ "▁st",
+ "ep"
+ ],
+ [
+ "▁ste",
+ "p"
+ ],
+ [
+ "▁",
+ "step"
+ ],
+ [
+ "▁p",
+ "ers"
+ ],
+ [
+ "▁per",
+ "s"
+ ],
+ [
+ "▁pe",
+ "rs"
+ ],
+ [
+ "▁",
+ "pers"
+ ],
+ [
+ "ç",
+ "ão"
+ ],
+ [
+ "ob",
+ "j"
+ ],
+ [
+ "o",
+ "z"
+ ],
+ [
+ "ul",
+ "a"
+ ],
+ [
+ "u",
+ "la"
+ ],
+ [
+ "▁r",
+ "ound"
+ ],
+ [
+ "▁ro",
+ "und"
+ ],
+ [
+ "▁rou",
+ "nd"
+ ],
+ [
+ "▁",
+ "round"
+ ],
+ [
+ "▁u",
+ "pon"
+ ],
+ [
+ "▁up",
+ "on"
+ ],
+ [
+ "▁re",
+ "source"
+ ],
+ [
+ "▁res",
+ "ource"
+ ],
+ [
+ "▁",
+ "resource"
+ ],
+ [
+ "▁val",
+ "id"
+ ],
+ [
+ "▁",
+ "valid"
+ ],
+ [
+ "▁I",
+ "I"
+ ],
+ [
+ "▁",
+ "II"
+ ],
+ [
+ "bu",
+ "g"
+ ],
+ [
+ "b",
+ "ug"
+ ],
+ [
+ "st",
+ "d"
+ ],
+ [
+ "s",
+ "td"
+ ],
+ [
+ "▁a",
+ "ng"
+ ],
+ [
+ "▁an",
+ "g"
+ ],
+ [
+ "▁",
+ "ang"
+ ],
+ [
+ "sp",
+ "an"
+ ],
+ [
+ "s",
+ "pan"
+ ],
+ [
+ "po",
+ "l"
+ ],
+ [
+ "p",
+ "ol"
+ ],
+ [
+ "ial",
+ "og"
+ ],
+ [
+ "ia",
+ "log"
+ ],
+ [
+ "▁p",
+ "hot"
+ ],
+ [
+ "▁ph",
+ "ot"
+ ],
+ [
+ "?",
+ "'"
+ ],
+ [
+ "D",
+ "B"
+ ],
+ [
+ "▁F",
+ "in"
+ ],
+ [
+ "▁Fi",
+ "n"
+ ],
+ [
+ "▁",
+ "Fin"
+ ],
+ [
+ "V",
+ "E"
+ ],
+ [
+ "E",
+ "m"
+ ],
+ [
+ "▁c",
+ "am"
+ ],
+ [
+ "▁ca",
+ "m"
+ ],
+ [
+ "▁",
+ "cam"
+ ],
+ [
+ "tar",
+ "get"
+ ],
+ [
+ "t",
+ "arget"
+ ],
+ [
+ "pe",
+ "cted"
+ ],
+ [
+ "pect",
+ "ed"
+ ],
+ [
+ "pec",
+ "ted"
+ ],
+ [
+ "He",
+ "l"
+ ],
+ [
+ "H",
+ "el"
+ ],
+ [
+ "▁u",
+ "t"
+ ],
+ [
+ "▁",
+ "ut"
+ ],
+ [
+ "▁T",
+ "est"
+ ],
+ [
+ "▁Te",
+ "st"
+ ],
+ [
+ "▁Tes",
+ "t"
+ ],
+ [
+ "▁",
+ "Test"
+ ],
+ [
+ "▁t",
+ "own"
+ ],
+ [
+ "▁to",
+ "wn"
+ ],
+ [
+ "▁tow",
+ "n"
+ ],
+ [
+ "▁",
+ "town"
+ ],
+ [
+ "al",
+ "ign"
+ ],
+ [
+ "ali",
+ "gn"
+ ],
+ [
+ "▁we",
+ "bs"
+ ],
+ [
+ "▁web",
+ "s"
+ ],
+ [
+ "in",
+ "ner"
+ ],
+ [
+ "inn",
+ "er"
+ ],
+ [
+ "au",
+ "gh"
+ ],
+ [
+ "aug",
+ "h"
+ ],
+ [
+ "a",
+ "ugh"
+ ],
+ [
+ "▁ex",
+ "cept"
+ ],
+ [
+ "▁",
+ "except"
+ ],
+ [
+ "▁init",
+ "ial"
+ ],
+ [
+ "▁initi",
+ "al"
+ ],
+ [
+ "▁",
+ "initial"
+ ],
+ [
+ "en",
+ "ty"
+ ],
+ [
+ "ent",
+ "y"
+ ],
+ [
+ "lic",
+ "h"
+ ],
+ [
+ "li",
+ "ch"
+ ],
+ [
+ "l",
+ "ich"
+ ],
+ [
+ "▁A",
+ "ut"
+ ],
+ [
+ "▁Au",
+ "t"
+ ],
+ [
+ "▁",
+ "Aut"
+ ],
+ [
+ "to",
+ "p"
+ ],
+ [
+ "t",
+ "op"
+ ],
+ [
+ "▁f",
+ "ail"
+ ],
+ [
+ "▁fa",
+ "il"
+ ],
+ [
+ "▁",
+ "fail"
+ ],
+ [
+ "on",
+ "a"
+ ],
+ [
+ "o",
+ "na"
+ ],
+ [
+ "▁ben",
+ "ef"
+ ],
+ [
+ "an",
+ "ks"
+ ],
+ [
+ "ank",
+ "s"
+ ],
+ [
+ "is",
+ "che"
+ ],
+ [
+ "isch",
+ "e"
+ ],
+ [
+ "isc",
+ "he"
+ ],
+ [
+ "i",
+ "sche"
+ ],
+ [
+ ".",
+ "*"
+ ],
+ [
+ "▁sign",
+ "ific"
+ ],
+ [
+ "▁cont",
+ "act"
+ ],
+ [
+ "▁",
+ "contact"
+ ],
+ [
+ "Re",
+ "c"
+ ],
+ [
+ "R",
+ "ec"
+ ],
+ [
+ "ar",
+ "io"
+ ],
+ [
+ "ari",
+ "o"
+ ],
+ [
+ "a",
+ "rio"
+ ],
+ [
+ "ot",
+ "tom"
+ ],
+ [
+ "ott",
+ "om"
+ ],
+ [
+ "otto",
+ "m"
+ ],
+ [
+ "▁rel",
+ "ationship"
+ ],
+ [
+ "▁relations",
+ "hip"
+ ],
+ [
+ "▁relation",
+ "ship"
+ ],
+ [
+ "])",
+ ";"
+ ],
+ [
+ "]",
+ ");"
+ ],
+ [
+ "▁Н",
+ "а"
+ ],
+ [
+ "▁",
+ "На"
+ ],
+ [
+ "He",
+ "ad"
+ ],
+ [
+ "H",
+ "ead"
+ ],
+ [
+ "form",
+ "at"
+ ],
+ [
+ "for",
+ "mat"
+ ],
+ [
+ "▁é",
+ "t"
+ ],
+ [
+ "▁",
+ "ét"
+ ],
+ [
+ "▁M",
+ "ore"
+ ],
+ [
+ "▁Mor",
+ "e"
+ ],
+ [
+ "▁Mo",
+ "re"
+ ],
+ [
+ "▁",
+ "More"
+ ],
+ [
+ "act",
+ "ory"
+ ],
+ [
+ "actor",
+ "y"
+ ],
+ [
+ "port",
+ "un"
+ ],
+ [
+ "+",
+ "\\"
+ ],
+ [
+ "▁sim",
+ "ply"
+ ],
+ [
+ "▁simpl",
+ "y"
+ ],
+ [
+ "▁e",
+ "p"
+ ],
+ [
+ "▁",
+ "ep"
+ ],
+ [
+ "▁R",
+ "uss"
+ ],
+ [
+ "▁Ru",
+ "ss"
+ ],
+ [
+ "▁Rus",
+ "s"
+ ],
+ [
+ "n",
+ "í"
+ ],
+ [
+ "u",
+ "a"
+ ],
+ [
+ "er",
+ "c"
+ ],
+ [
+ "e",
+ "rc"
+ ],
+ [
+ "▁long",
+ "er"
+ ],
+ [
+ "▁lon",
+ "ger"
+ ],
+ [
+ "in",
+ "ition"
+ ],
+ [
+ "init",
+ "ion"
+ ],
+ [
+ "ect",
+ "or"
+ ],
+ [
+ "ec",
+ "tor"
+ ],
+ [
+ "e",
+ "ctor"
+ ],
+ [
+ "apt",
+ "ion"
+ ],
+ [
+ "a",
+ "ption"
+ ],
+ [
+ "▁prof",
+ "ess"
+ ],
+ [
+ "▁profes",
+ "s"
+ ],
+ [
+ "▁M",
+ "us"
+ ],
+ [
+ "▁Mu",
+ "s"
+ ],
+ [
+ "▁",
+ "Mus"
+ ],
+ [
+ "il",
+ "ities"
+ ],
+ [
+ "ili",
+ "ties"
+ ],
+ [
+ "è",
+ "s"
+ ],
+ [
+ "▁A",
+ "ct"
+ ],
+ [
+ "▁Ac",
+ "t"
+ ],
+ [
+ "▁",
+ "Act"
+ ],
+ [
+ "off",
+ "set"
+ ],
+ [
+ "offs",
+ "et"
+ ],
+ [
+ "▁i",
+ "ll"
+ ],
+ [
+ "▁il",
+ "l"
+ ],
+ [
+ "▁",
+ "ill"
+ ],
+ [
+ "ba",
+ "nd"
+ ],
+ [
+ "ban",
+ "d"
+ ],
+ [
+ "b",
+ "and"
+ ],
+ [
+ "▁A",
+ "g"
+ ],
+ [
+ "▁",
+ "Ag"
+ ],
+ [
+ "▁П",
+ "о"
+ ],
+ [
+ "▁",
+ "По"
+ ],
+ [
+ "б",
+ "и"
+ ],
+ [
+ "cont",
+ "ent"
+ ],
+ [
+ "ic",
+ "on"
+ ],
+ [
+ "ico",
+ "n"
+ ],
+ [
+ "i",
+ "con"
+ ],
+ [
+ "▁work",
+ "s"
+ ],
+ [
+ "▁wor",
+ "ks"
+ ],
+ [
+ "▁",
+ "works"
+ ],
+ [
+ "yn",
+ "am"
+ ],
+ [
+ "yna",
+ "m"
+ ],
+ [
+ "y",
+ "nam"
+ ],
+ [
+ "pl",
+ "ement"
+ ],
+ [
+ "ple",
+ "ment"
+ ],
+ [
+ "p",
+ "lement"
+ ],
+ [
+ "Res",
+ "ource"
+ ],
+ [
+ "Re",
+ "source"
+ ],
+ [
+ "Act",
+ "ion"
+ ],
+ [
+ "A",
+ "ction"
+ ],
+ [
+ "▁diff",
+ "icult"
+ ],
+ [
+ "▁W",
+ "est"
+ ],
+ [
+ "▁We",
+ "st"
+ ],
+ [
+ "▁Wes",
+ "t"
+ ],
+ [
+ "▁",
+ "West"
+ ],
+ [
+ "▁v",
+ "ideo"
+ ],
+ [
+ "▁vide",
+ "o"
+ ],
+ [
+ "▁",
+ "video"
+ ],
+ [
+ "▁T",
+ "HE"
+ ],
+ [
+ "▁TH",
+ "E"
+ ],
+ [
+ "▁",
+ "THE"
+ ],
+ [
+ "▁de",
+ "cl"
+ ],
+ [
+ "▁dec",
+ "l"
+ ],
+ [
+ "▁",
+ "decl"
+ ],
+ [
+ "on",
+ "don"
+ ],
+ [
+ "ond",
+ "on"
+ ],
+ [
+ "ondo",
+ "n"
+ ],
+ [
+ "de",
+ "d"
+ ],
+ [
+ "d",
+ "ed"
+ ],
+ [
+ "}{",
+ "\\"
+ ],
+ [
+ "}",
+ "{\\"
+ ],
+ [
+ "oc",
+ "r"
+ ],
+ [
+ "o",
+ "cr"
+ ],
+ [
+ "▁C",
+ "ity"
+ ],
+ [
+ "▁Cit",
+ "y"
+ ],
+ [
+ "▁Ci",
+ "ty"
+ ],
+ [
+ "▁",
+ "City"
+ ],
+ [
+ "▁",
+ "я"
+ ],
+ [
+ "ue",
+ "r"
+ ],
+ [
+ "u",
+ "er"
+ ],
+ [
+ "c",
+ "z"
+ ],
+ [
+ "▁im",
+ "ag"
+ ],
+ [
+ "▁i",
+ "mag"
+ ],
+ [
+ "▁",
+ "imag"
+ ],
+ [
+ "c",
+ "r"
+ ],
+ [
+ "et",
+ "e"
+ ],
+ [
+ "e",
+ "te"
+ ],
+ [
+ "id",
+ "get"
+ ],
+ [
+ "idge",
+ "t"
+ ],
+ [
+ "▁M",
+ "od"
+ ],
+ [
+ "▁Mo",
+ "d"
+ ],
+ [
+ "▁",
+ "Mod"
+ ],
+ [
+ "▁for",
+ "ward"
+ ],
+ [
+ "▁",
+ "forward"
+ ],
+ [
+ "▁p",
+ "ict"
+ ],
+ [
+ "▁pi",
+ "ct"
+ ],
+ [
+ "▁pic",
+ "t"
+ ],
+ [
+ "or",
+ "ge"
+ ],
+ [
+ "org",
+ "e"
+ ],
+ [
+ "▁sub",
+ "ject"
+ ],
+ [
+ "▁",
+ "subject"
+ ],
+ [
+ "up",
+ "date"
+ ],
+ [
+ "at",
+ "tle"
+ ],
+ [
+ "att",
+ "le"
+ ],
+ [
+ "s",
+ "a"
+ ],
+ [
+ "▁A",
+ "nt"
+ ],
+ [
+ "▁An",
+ "t"
+ ],
+ [
+ "▁",
+ "Ant"
+ ],
+ [
+ "▁r",
+ "unning"
+ ],
+ [
+ "▁run",
+ "ning"
+ ],
+ [
+ "▁",
+ "running"
+ ],
+ [
+ "▁s",
+ "al"
+ ],
+ [
+ "▁sa",
+ "l"
+ ],
+ [
+ "▁",
+ "sal"
+ ],
+ [
+ "con",
+ "ne"
+ ],
+ [
+ "conn",
+ "e"
+ ],
+ [
+ "c",
+ "onne"
+ ],
+ [
+ "▁out",
+ "put"
+ ],
+ [
+ "▁",
+ "output"
+ ],
+ [
+ "ad",
+ "ata"
+ ],
+ [
+ "ada",
+ "ta"
+ ],
+ [
+ "a",
+ "data"
+ ],
+ [
+ "M",
+ "L"
+ ],
+ [
+ "Che",
+ "ck"
+ ],
+ [
+ "C",
+ "heck"
+ ],
+ [
+ "led",
+ "ge"
+ ],
+ [
+ "l",
+ "edge"
+ ],
+ [
+ "▁p",
+ "aper"
+ ],
+ [
+ "▁pa",
+ "per"
+ ],
+ [
+ "▁pap",
+ "er"
+ ],
+ [
+ "▁",
+ "paper"
+ ],
+ [
+ "param",
+ "s"
+ ],
+ [
+ "par",
+ "ams"
+ ],
+ [
+ "para",
+ "ms"
+ ],
+ [
+ "av",
+ "y"
+ ],
+ [
+ "a",
+ "vy"
+ ],
+ [
+ "▁a",
+ "f"
+ ],
+ [
+ "▁",
+ "af"
+ ],
+ [
+ "▁e",
+ "ine"
+ ],
+ [
+ "▁ein",
+ "e"
+ ],
+ [
+ "▁j",
+ "our"
+ ],
+ [
+ "▁jo",
+ "ur"
+ ],
+ [
+ "▁jou",
+ "r"
+ ],
+ [
+ "▁",
+ "jour"
+ ],
+ [
+ "A",
+ "Y"
+ ],
+ [
+ "▁it",
+ "self"
+ ],
+ [
+ "▁its",
+ "elf"
+ ],
+ [
+ "▁S",
+ "tr"
+ ],
+ [
+ "▁St",
+ "r"
+ ],
+ [
+ "▁",
+ "Str"
+ ],
+ [
+ "st",
+ "yle"
+ ],
+ [
+ "sty",
+ "le"
+ ],
+ [
+ "Th",
+ "at"
+ ],
+ [
+ "T",
+ "hat"
+ ],
+ [
+ "▁m",
+ "illion"
+ ],
+ [
+ "▁mill",
+ "ion"
+ ],
+ [
+ "▁l",
+ "anguage"
+ ],
+ [
+ "▁",
+ "language"
+ ],
+ [
+ "O",
+ "S"
+ ],
+ [
+ "vi",
+ "ng"
+ ],
+ [
+ "vin",
+ "g"
+ ],
+ [
+ "v",
+ "ing"
+ ],
+ [
+ "▁м",
+ "а"
+ ],
+ [
+ "▁",
+ "ма"
+ ],
+ [
+ "▁т",
+ "о"
+ ],
+ [
+ "▁",
+ "то"
+ ],
+ [
+ ")",
+ "("
+ ],
+ [
+ "▁b",
+ "uy"
+ ],
+ [
+ "▁bu",
+ "y"
+ ],
+ [
+ ".",
+ "/"
+ ],
+ [
+ "▁.",
+ ".."
+ ],
+ [
+ "▁..",
+ "."
+ ],
+ [
+ "▁",
+ "..."
+ ],
+ [
+ "▁t",
+ "ried"
+ ],
+ [
+ "▁tr",
+ "ied"
+ ],
+ [
+ "▁tri",
+ "ed"
+ ],
+ [
+ "▁com",
+ "pl"
+ ],
+ [
+ "▁comp",
+ "l"
+ ],
+ [
+ "▁act",
+ "iv"
+ ],
+ [
+ "▁",
+ "activ"
+ ],
+ [
+ "ap",
+ "ped"
+ ],
+ [
+ "app",
+ "ed"
+ ],
+ [
+ "appe",
+ "d"
+ ],
+ [
+ "a",
+ "pped"
+ ],
+ [
+ "But",
+ "ton"
+ ],
+ [
+ "B",
+ "utton"
+ ],
+ [
+ "To",
+ "ken"
+ ],
+ [
+ "Tok",
+ "en"
+ ],
+ [
+ "T",
+ "oken"
+ ],
+ [
+ "▁prov",
+ "ided"
+ ],
+ [
+ "▁provide",
+ "d"
+ ],
+ [
+ "ib",
+ "er"
+ ],
+ [
+ "ibe",
+ "r"
+ ],
+ [
+ "i",
+ "ber"
+ ],
+ [
+ "▁c",
+ "reated"
+ ],
+ [
+ "▁cre",
+ "ated"
+ ],
+ [
+ "▁create",
+ "d"
+ ],
+ [
+ "▁creat",
+ "ed"
+ ],
+ [
+ "▁",
+ "created"
+ ],
+ [
+ "cur",
+ "ity"
+ ],
+ [
+ "c",
+ "urity"
+ ],
+ [
+ "En",
+ "d"
+ ],
+ [
+ "E",
+ "nd"
+ ],
+ [
+ "a",
+ "ł"
+ ],
+ [
+ "us",
+ "ter"
+ ],
+ [
+ "ust",
+ "er"
+ ],
+ [
+ "u",
+ "ster"
+ ],
+ [
+ "iz",
+ "ing"
+ ],
+ [
+ "izi",
+ "ng"
+ ],
+ [
+ "i",
+ "zing"
+ ],
+ [
+ "om",
+ "b"
+ ],
+ [
+ "o",
+ "mb"
+ ],
+ [
+ "▁s",
+ "ich"
+ ],
+ [
+ "▁si",
+ "ch"
+ ],
+ [
+ "▁com",
+ "pon"
+ ],
+ [
+ "▁comp",
+ "on"
+ ],
+ [
+ "▁S",
+ "ee"
+ ],
+ [
+ "▁Se",
+ "e"
+ ],
+ [
+ "▁",
+ "See"
+ ],
+ [
+ "▁u",
+ "int"
+ ],
+ [
+ "▁ui",
+ "nt"
+ ],
+ [
+ "▁",
+ "uint"
+ ],
+ [
+ "▁l",
+ "abel"
+ ],
+ [
+ "▁la",
+ "bel"
+ ],
+ [
+ "▁lab",
+ "el"
+ ],
+ [
+ "▁",
+ "label"
+ ],
+ [
+ "vo",
+ "l"
+ ],
+ [
+ "v",
+ "ol"
+ ],
+ [
+ "ó",
+ "w"
+ ],
+ [
+ "oc",
+ "ol"
+ ],
+ [
+ "oco",
+ "l"
+ ],
+ [
+ "o",
+ "col"
+ ],
+ [
+ "▁re",
+ "ceived"
+ ],
+ [
+ "▁rece",
+ "ived"
+ ],
+ [
+ "▁receive",
+ "d"
+ ],
+ [
+ "▁in",
+ "tern"
+ ],
+ [
+ "▁int",
+ "ern"
+ ],
+ [
+ "▁inter",
+ "n"
+ ],
+ [
+ "▁inte",
+ "rn"
+ ],
+ [
+ "▁",
+ "intern"
+ ],
+ [
+ "ц",
+ "е"
+ ],
+ [
+ "R",
+ "un"
+ ],
+ [
+ "▁r",
+ "oad"
+ ],
+ [
+ "▁ro",
+ "ad"
+ ],
+ [
+ "▁",
+ "road"
+ ],
+ [
+ "▁O",
+ "ct"
+ ],
+ [
+ "▁",
+ "Oct"
+ ],
+ [
+ "▁C",
+ "omp"
+ ],
+ [
+ "▁Com",
+ "p"
+ ],
+ [
+ "▁Co",
+ "mp"
+ ],
+ [
+ "▁",
+ "Comp"
+ ],
+ [
+ "▁stud",
+ "y"
+ ],
+ [
+ "▁т",
+ "е"
+ ],
+ [
+ "▁",
+ "те"
+ ],
+ [
+ "Ac",
+ "t"
+ ],
+ [
+ "A",
+ "ct"
+ ],
+ [
+ "▁t",
+ "our"
+ ],
+ [
+ "▁to",
+ "ur"
+ ],
+ [
+ "▁tou",
+ "r"
+ ],
+ [
+ "▁St",
+ "ate"
+ ],
+ [
+ "▁Stat",
+ "e"
+ ],
+ [
+ "▁Sta",
+ "te"
+ ],
+ [
+ "▁",
+ "State"
+ ],
+ [
+ "▁ad",
+ "ded"
+ ],
+ [
+ "▁add",
+ "ed"
+ ],
+ [
+ "▁",
+ "added"
+ ],
+ [
+ "htt",
+ "ps"
+ ],
+ [
+ "http",
+ "s"
+ ],
+ [
+ "st",
+ "ream"
+ ],
+ [
+ "stre",
+ "am"
+ ],
+ [
+ "▁l",
+ "ower"
+ ],
+ [
+ "▁lo",
+ "wer"
+ ],
+ [
+ "▁low",
+ "er"
+ ],
+ [
+ "▁",
+ "lower"
+ ],
+ [
+ "▁b",
+ "ox"
+ ],
+ [
+ "▁bo",
+ "x"
+ ],
+ [
+ "▁",
+ "box"
+ ],
+ [
+ "▁S",
+ "k"
+ ],
+ [
+ "▁",
+ "Sk"
+ ],
+ [
+ "▁them",
+ "selves"
+ ],
+ [
+ "▁c",
+ "ross"
+ ],
+ [
+ "▁cr",
+ "oss"
+ ],
+ [
+ "▁cro",
+ "ss"
+ ],
+ [
+ "▁",
+ "cross"
+ ],
+ [
+ "▁e",
+ "cho"
+ ],
+ [
+ "▁ec",
+ "ho"
+ ],
+ [
+ "▁",
+ "echo"
+ ],
+ [
+ "▁dev",
+ "ice"
+ ],
+ [
+ "▁",
+ "device"
+ ],
+ [
+ "pos",
+ "e"
+ ],
+ [
+ "po",
+ "se"
+ ],
+ [
+ "p",
+ "ose"
+ ],
+ [
+ "▁g",
+ "ames"
+ ],
+ [
+ "▁game",
+ "s"
+ ],
+ [
+ "▁gam",
+ "es"
+ ],
+ [
+ "▁ga",
+ "mes"
+ ],
+ [
+ "P",
+ "L"
+ ],
+ [
+ "W",
+ "indow"
+ ],
+ [
+ "is",
+ "es"
+ ],
+ [
+ "ise",
+ "s"
+ ],
+ [
+ "i",
+ "ses"
+ ],
+ [
+ "ti",
+ "tle"
+ ],
+ [
+ "tit",
+ "le"
+ ],
+ [
+ "t",
+ "itle"
+ ],
+ [
+ "St",
+ "ream"
+ ],
+ [
+ "z",
+ "t"
+ ],
+ [
+ "▁S",
+ "w"
+ ],
+ [
+ "▁",
+ "Sw"
+ ],
+ [
+ "▁r",
+ "ole"
+ ],
+ [
+ "▁ro",
+ "le"
+ ],
+ [
+ "▁",
+ "role"
+ ],
+ [
+ "ia",
+ "nt"
+ ],
+ [
+ "ian",
+ "t"
+ ],
+ [
+ "i",
+ "ant"
+ ],
+ [
+ "k",
+ "u"
+ ],
+ [
+ "se",
+ "qu"
+ ],
+ [
+ "seq",
+ "u"
+ ],
+ [
+ "s",
+ "equ"
+ ],
+ [
+ "▁l",
+ "ate"
+ ],
+ [
+ "▁la",
+ "te"
+ ],
+ [
+ "▁lat",
+ "e"
+ ],
+ [
+ "▁",
+ "late"
+ ],
+ [
+ "▁s",
+ "old"
+ ],
+ [
+ "▁so",
+ "ld"
+ ],
+ [
+ "▁sol",
+ "d"
+ ],
+ [
+ "р",
+ "я"
+ ],
+ [
+ "Com",
+ "m"
+ ],
+ [
+ "Co",
+ "mm"
+ ],
+ [
+ "C",
+ "omm"
+ ],
+ [
+ "▁en",
+ "tre"
+ ],
+ [
+ "▁ent",
+ "re"
+ ],
+ [
+ "▁entr",
+ "e"
+ ],
+ [
+ "▁",
+ "entre"
+ ],
+ [
+ "▁d",
+ "og"
+ ],
+ [
+ "▁do",
+ "g"
+ ],
+ [
+ "▁",
+ "dog"
+ ],
+ [
+ "dev",
+ "ice"
+ ],
+ [
+ "P",
+ "ar"
+ ],
+ [
+ "▁like",
+ "ly"
+ ],
+ [
+ "▁lik",
+ "ely"
+ ],
+ [
+ "▁",
+ "likely"
+ ],
+ [
+ "^{",
+ "-"
+ ],
+ [
+ "^",
+ "{-"
+ ],
+ [
+ "▁l",
+ "en"
+ ],
+ [
+ "▁le",
+ "n"
+ ],
+ [
+ "▁",
+ "len"
+ ],
+ [
+ "▁P",
+ "aul"
+ ],
+ [
+ "▁Pa",
+ "ul"
+ ],
+ [
+ "▁",
+ "Paul"
+ ],
+ [
+ "▁t",
+ "ool"
+ ],
+ [
+ "▁to",
+ "ol"
+ ],
+ [
+ "▁too",
+ "l"
+ ],
+ [
+ "▁",
+ "tool"
+ ],
+ [
+ "Of",
+ "f"
+ ],
+ [
+ "O",
+ "ff"
+ ],
+ [
+ "▁f",
+ "amil"
+ ],
+ [
+ "▁fam",
+ "il"
+ ],
+ [
+ "▁fa",
+ "mil"
+ ],
+ [
+ "▁d",
+ "raw"
+ ],
+ [
+ "▁dr",
+ "aw"
+ ],
+ [
+ "▁",
+ "draw"
+ ],
+ [
+ "ap",
+ "ping"
+ ],
+ [
+ "app",
+ "ing"
+ ],
+ [
+ "a",
+ "pping"
+ ],
+ [
+ "▁ev",
+ "ents"
+ ],
+ [
+ "▁even",
+ "ts"
+ ],
+ [
+ "▁event",
+ "s"
+ ],
+ [
+ "▁",
+ "events"
+ ],
+ [
+ "cre",
+ "t"
+ ],
+ [
+ "cr",
+ "et"
+ ],
+ [
+ "c",
+ "ret"
+ ],
+ [
+ "rou",
+ "ght"
+ ],
+ [
+ "rough",
+ "t"
+ ],
+ [
+ "r",
+ "ought"
+ ],
+ [
+ "Cont",
+ "ent"
+ ],
+ [
+ "▁soft",
+ "ware"
+ ],
+ [
+ "ri",
+ "a"
+ ],
+ [
+ "r",
+ "ia"
+ ],
+ [
+ "ms",
+ "g"
+ ],
+ [
+ "m",
+ "sg"
+ ],
+ [
+ "ga",
+ "mma"
+ ],
+ [
+ "g",
+ "amma"
+ ],
+ [
+ "▁h",
+ "ear"
+ ],
+ [
+ "▁he",
+ "ar"
+ ],
+ [
+ "Op",
+ "er"
+ ],
+ [
+ "O",
+ "per"
+ ],
+ [
+ "▁your",
+ "self"
+ ],
+ [
+ "▁yours",
+ "elf"
+ ],
+ [
+ "▁l",
+ "iter"
+ ],
+ [
+ "▁li",
+ "ter"
+ ],
+ [
+ "▁lit",
+ "er"
+ ],
+ [
+ "▁",
+ "liter"
+ ],
+ [
+ "em",
+ "p"
+ ],
+ [
+ "e",
+ "mp"
+ ],
+ [
+ "▁se",
+ "par"
+ ],
+ [
+ "▁sep",
+ "ar"
+ ],
+ [
+ "▁",
+ "separ"
+ ],
+ [
+ "▁",
+ "З"
+ ],
+ [
+ "▁t",
+ "itle"
+ ],
+ [
+ "▁tit",
+ "le"
+ ],
+ [
+ "▁ti",
+ "tle"
+ ],
+ [
+ "▁",
+ "title"
+ ],
+ [
+ "M",
+ "ethod"
+ ],
+ [
+ "math",
+ "rm"
+ ],
+ [
+ "▁s",
+ "low"
+ ],
+ [
+ "▁sl",
+ "ow"
+ ],
+ [
+ "▁R",
+ "om"
+ ],
+ [
+ "▁Ro",
+ "m"
+ ],
+ [
+ "▁",
+ "Rom"
+ ],
+ [
+ "!",
+ "!"
+ ],
+ [
+ "▁t",
+ "ax"
+ ],
+ [
+ "▁ta",
+ "x"
+ ],
+ [
+ "▁",
+ "tax"
+ ],
+ [
+ "ск",
+ "а"
+ ],
+ [
+ "с",
+ "ка"
+ ],
+ [
+ "empl",
+ "ate"
+ ],
+ [
+ "emp",
+ "late"
+ ],
+ [
+ "o",
+ "i"
+ ],
+ [
+ "▁A",
+ "rt"
+ ],
+ [
+ "▁Ar",
+ "t"
+ ],
+ [
+ "▁",
+ "Art"
+ ],
+ [
+ "f",
+ "alse"
+ ],
+ [
+ "ast",
+ "ic"
+ ],
+ [
+ "ст",
+ "ь"
+ ],
+ [
+ "с",
+ "ть"
+ ],
+ [
+ "oc",
+ "ket"
+ ],
+ [
+ "ock",
+ "et"
+ ],
+ [
+ "▁e",
+ "ns"
+ ],
+ [
+ "▁en",
+ "s"
+ ],
+ [
+ "▁",
+ "ens"
+ ],
+ [
+ "T",
+ "O"
+ ],
+ [
+ "am",
+ "ente"
+ ],
+ [
+ "ame",
+ "nte"
+ ],
+ [
+ "ament",
+ "e"
+ ],
+ [
+ "amen",
+ "te"
+ ],
+ [
+ "a",
+ "mente"
+ ],
+ [
+ "lo",
+ "cal"
+ ],
+ [
+ "loc",
+ "al"
+ ],
+ [
+ "l",
+ "ocal"
+ ],
+ [
+ "ch",
+ "ie"
+ ],
+ [
+ "chi",
+ "e"
+ ],
+ [
+ "▁p",
+ "an"
+ ],
+ [
+ "▁pa",
+ "n"
+ ],
+ [
+ "▁",
+ "pan"
+ ],
+ [
+ "ни",
+ "й"
+ ],
+ [
+ "ch",
+ "ema"
+ ],
+ [
+ "che",
+ "ma"
+ ],
+ [
+ "chem",
+ "a"
+ ],
+ [
+ "▁N",
+ "orth"
+ ],
+ [
+ "▁Nor",
+ "th"
+ ],
+ [
+ "▁Nort",
+ "h"
+ ],
+ [
+ "з",
+ "о"
+ ],
+ [
+ "▁>",
+ "="
+ ],
+ [
+ "▁",
+ ">="
+ ],
+ [
+ "A",
+ "ut"
+ ],
+ [
+ "▁d",
+ "ig"
+ ],
+ [
+ "▁di",
+ "g"
+ ],
+ [
+ "▁",
+ "dig"
+ ],
+ [
+ "▁se",
+ "ems"
+ ],
+ [
+ "▁see",
+ "ms"
+ ],
+ [
+ "▁seem",
+ "s"
+ ],
+ [
+ "▁mor",
+ "ning"
+ ],
+ [
+ "so",
+ "le"
+ ],
+ [
+ "sol",
+ "e"
+ ],
+ [
+ "s",
+ "ole"
+ ],
+ [
+ "um",
+ "er"
+ ],
+ [
+ "ume",
+ "r"
+ ],
+ [
+ "u",
+ "mer"
+ ],
+ [
+ "del",
+ "ta"
+ ],
+ [
+ "d",
+ "elta"
+ ],
+ [
+ "it",
+ "é"
+ ],
+ [
+ "i",
+ "té"
+ ],
+ [
+ "ab",
+ "ase"
+ ],
+ [
+ "aba",
+ "se"
+ ],
+ [
+ "a",
+ "base"
+ ],
+ [
+ "ra",
+ "f"
+ ],
+ [
+ "r",
+ "af"
+ ],
+ [
+ "▁ob",
+ "serv"
+ ],
+ [
+ "▁obs",
+ "erv"
+ ],
+ [
+ "▁",
+ "observ"
+ ],
+ [
+ "▁E",
+ "st"
+ ],
+ [
+ "▁Es",
+ "t"
+ ],
+ [
+ "▁",
+ "Est"
+ ],
+ [
+ "▁s",
+ "eg"
+ ],
+ [
+ "▁se",
+ "g"
+ ],
+ [
+ "▁",
+ "seg"
+ ],
+ [
+ "▁[",
+ "]"
+ ],
+ [
+ "▁",
+ "[]"
+ ],
+ [
+ "▁P",
+ "res"
+ ],
+ [
+ "▁Pr",
+ "es"
+ ],
+ [
+ "▁Pre",
+ "s"
+ ],
+ [
+ "▁",
+ "Pres"
+ ],
+ [
+ "if",
+ "ul"
+ ],
+ [
+ "i",
+ "ful"
+ ],
+ [
+ "pu",
+ "sh"
+ ],
+ [
+ "pus",
+ "h"
+ ],
+ [
+ "p",
+ "ush"
+ ],
+ [
+ "▁O",
+ "ff"
+ ],
+ [
+ "▁Of",
+ "f"
+ ],
+ [
+ "▁",
+ "Off"
+ ],
+ [
+ "ip",
+ "e"
+ ],
+ [
+ "i",
+ "pe"
+ ],
+ [
+ "at",
+ "i"
+ ],
+ [
+ "a",
+ "ti"
+ ],
+ [
+ "▁d",
+ "im"
+ ],
+ [
+ "▁di",
+ "m"
+ ],
+ [
+ "▁",
+ "dim"
+ ],
+ [
+ "ce",
+ "ed"
+ ],
+ [
+ "c",
+ "eed"
+ ],
+ [
+ "En",
+ "t"
+ ],
+ [
+ "E",
+ "nt"
+ ],
+ [
+ "__",
+ "__"
+ ],
+ [
+ "___",
+ "_"
+ ],
+ [
+ "_",
+ "___"
+ ],
+ [
+ "en",
+ "try"
+ ],
+ [
+ "ent",
+ "ry"
+ ],
+ [
+ "entr",
+ "y"
+ ],
+ [
+ "▁f",
+ "ight"
+ ],
+ [
+ "▁fig",
+ "ht"
+ ],
+ [
+ "▁fi",
+ "ght"
+ ],
+ [
+ "▁c",
+ "red"
+ ],
+ [
+ "▁cre",
+ "d"
+ ],
+ [
+ "▁cr",
+ "ed"
+ ],
+ [
+ "▁",
+ "cred"
+ ],
+ [
+ "▁O",
+ "R"
+ ],
+ [
+ "▁",
+ "OR"
+ ],
+ [
+ "▁D",
+ "ep"
+ ],
+ [
+ "▁De",
+ "p"
+ ],
+ [
+ "▁",
+ "Dep"
+ ],
+ [
+ "$",
+ "{"
+ ],
+ [
+ "ле",
+ "н"
+ ],
+ [
+ "л",
+ "ен"
+ ],
+ [
+ "Creat",
+ "e"
+ ],
+ [
+ "C",
+ "reate"
+ ],
+ [
+ "▁Apr",
+ "il"
+ ],
+ [
+ "▁Ap",
+ "ril"
+ ],
+ [
+ "min",
+ "istr"
+ ],
+ [
+ "F",
+ "L"
+ ],
+ [
+ "▁A",
+ "p"
+ ],
+ [
+ "▁",
+ "Ap"
+ ],
+ [
+ "▁H",
+ "ere"
+ ],
+ [
+ "▁He",
+ "re"
+ ],
+ [
+ "▁Her",
+ "e"
+ ],
+ [
+ "▁",
+ "Here"
+ ],
+ [
+ "priv",
+ "ate"
+ ],
+ [
+ "p",
+ "rivate"
+ ],
+ [
+ "In",
+ "stance"
+ ],
+ [
+ "Inst",
+ "ance"
+ ],
+ [
+ "ie",
+ "m"
+ ],
+ [
+ "i",
+ "em"
+ ],
+ [
+ "▁off",
+ "ice"
+ ],
+ [
+ "▁offic",
+ "e"
+ ],
+ [
+ "▁th",
+ "ird"
+ ],
+ [
+ "▁",
+ "third"
+ ],
+ [
+ "▁up",
+ "date"
+ ],
+ [
+ "▁",
+ "update"
+ ],
+ [
+ "Lin",
+ "e"
+ ],
+ [
+ "Li",
+ "ne"
+ ],
+ [
+ "L",
+ "ine"
+ ],
+ [
+ "ta",
+ "g"
+ ],
+ [
+ "t",
+ "ag"
+ ],
+ [
+ "▁e",
+ "specially"
+ ],
+ [
+ "▁espec",
+ "ially"
+ ],
+ [
+ "▁especial",
+ "ly"
+ ],
+ [
+ "▁",
+ "especially"
+ ],
+ [
+ "▁го",
+ "да"
+ ],
+ [
+ "▁год",
+ "а"
+ ],
+ [
+ "▁c",
+ "u"
+ ],
+ [
+ "▁",
+ "cu"
+ ],
+ [
+ "▁k",
+ "ill"
+ ],
+ [
+ "▁kil",
+ "l"
+ ],
+ [
+ "▁ki",
+ "ll"
+ ],
+ [
+ "▁",
+ "kill"
+ ],
+ [
+ "au",
+ "ght"
+ ],
+ [
+ "augh",
+ "t"
+ ],
+ [
+ "aug",
+ "ht"
+ ],
+ [
+ "▁s",
+ "we"
+ ],
+ [
+ "▁sw",
+ "e"
+ ],
+ [
+ "Option",
+ "s"
+ ],
+ [
+ "Opt",
+ "ions"
+ ],
+ [
+ "O",
+ "ptions"
+ ],
+ [
+ "I",
+ "M"
+ ],
+ [
+ "C",
+ "C"
+ ],
+ [
+ "▁com",
+ "pan"
+ ],
+ [
+ "▁comp",
+ "an"
+ ],
+ [
+ "ju",
+ "st"
+ ],
+ [
+ "j",
+ "ust"
+ ],
+ [
+ "▁Wh",
+ "ile"
+ ],
+ [
+ "▁",
+ "While"
+ ],
+ [
+ "iz",
+ "er"
+ ],
+ [
+ "ize",
+ "r"
+ ],
+ [
+ "i",
+ "zer"
+ ],
+ [
+ "▁м",
+ "о"
+ ],
+ [
+ "▁",
+ "мо"
+ ],
+ [
+ "к",
+ "е"
+ ],
+ [
+ "▁a",
+ "uto"
+ ],
+ [
+ "▁aut",
+ "o"
+ ],
+ [
+ "▁au",
+ "to"
+ ],
+ [
+ "▁",
+ "auto"
+ ],
+ [
+ "▁b",
+ "and"
+ ],
+ [
+ "▁ban",
+ "d"
+ ],
+ [
+ "▁ba",
+ "nd"
+ ],
+ [
+ "▁",
+ "band"
+ ],
+ [
+ "ме",
+ "н"
+ ],
+ [
+ "м",
+ "ен"
+ ],
+ [
+ "ique",
+ "s"
+ ],
+ [
+ "iqu",
+ "es"
+ ],
+ [
+ "iq",
+ "ues"
+ ],
+ [
+ "i",
+ "ques"
+ ],
+ [
+ "▁p",
+ "le"
+ ],
+ [
+ "▁pl",
+ "e"
+ ],
+ [
+ "▁",
+ "ple"
+ ],
+ [
+ "N",
+ "O"
+ ],
+ [
+ "▁O",
+ "F"
+ ],
+ [
+ "▁",
+ "OF"
+ ],
+ [
+ "▁s",
+ "ong"
+ ],
+ [
+ "▁so",
+ "ng"
+ ],
+ [
+ "▁son",
+ "g"
+ ],
+ [
+ "▁A",
+ "cc"
+ ],
+ [
+ "▁Ac",
+ "c"
+ ],
+ [
+ "▁",
+ "Acc"
+ ],
+ [
+ "EX",
+ "T"
+ ],
+ [
+ "E",
+ "XT"
+ ],
+ [
+ "en",
+ "sor"
+ ],
+ [
+ "ens",
+ "or"
+ ],
+ [
+ "enso",
+ "r"
+ ],
+ [
+ "in",
+ "ing"
+ ],
+ [
+ "ini",
+ "ng"
+ ],
+ [
+ "i",
+ "ning"
+ ],
+ [
+ "▁l",
+ "at"
+ ],
+ [
+ "▁la",
+ "t"
+ ],
+ [
+ "▁",
+ "lat"
+ ],
+ [
+ "bi",
+ "g"
+ ],
+ [
+ "b",
+ "ig"
+ ],
+ [
+ "▁K",
+ "ing"
+ ],
+ [
+ "▁Ki",
+ "ng"
+ ],
+ [
+ "▁Kin",
+ "g"
+ ],
+ [
+ "▁",
+ "King"
+ ],
+ [
+ "oc",
+ "h"
+ ],
+ [
+ "o",
+ "ch"
+ ],
+ [
+ "s",
+ "i"
+ ],
+ [
+ "▁H",
+ "ist"
+ ],
+ [
+ "▁His",
+ "t"
+ ],
+ [
+ "▁Hi",
+ "st"
+ ],
+ [
+ "▁",
+ "Hist"
+ ],
+ [
+ "▁qu",
+ "ality"
+ ],
+ [
+ "▁qual",
+ "ity"
+ ],
+ [
+ "▁",
+ "quality"
+ ],
+ [
+ "mod",
+ "e"
+ ],
+ [
+ "mo",
+ "de"
+ ],
+ [
+ "m",
+ "ode"
+ ],
+ [
+ "▁op",
+ "portun"
+ ],
+ [
+ "▁would",
+ "n"
+ ],
+ [
+ ":*",
+ "*"
+ ],
+ [
+ ":",
+ "**"
+ ],
+ [
+ "out",
+ "put"
+ ],
+ [
+ "▁fe",
+ "et"
+ ],
+ [
+ "▁fee",
+ "t"
+ ],
+ [
+ "▁m",
+ "is"
+ ],
+ [
+ "▁mi",
+ "s"
+ ],
+ [
+ "d",
+ "f"
+ ],
+ [
+ "ag",
+ "ing"
+ ],
+ [
+ "agi",
+ "ng"
+ ],
+ [
+ "a",
+ "ging"
+ ],
+ [
+ "▁м",
+ "е"
+ ],
+ [
+ "▁",
+ "ме"
+ ],
+ [
+ "▁t",
+ "ro"
+ ],
+ [
+ "▁tr",
+ "o"
+ ],
+ [
+ "▁d",
+ "efined"
+ ],
+ [
+ "▁def",
+ "ined"
+ ],
+ [
+ "▁define",
+ "d"
+ ],
+ [
+ "▁defin",
+ "ed"
+ ],
+ [
+ "▁",
+ "defined"
+ ],
+ [
+ "▁re",
+ "view"
+ ],
+ [
+ "▁rev",
+ "iew"
+ ],
+ [
+ "▁",
+ "review"
+ ],
+ [
+ "▁F",
+ "il"
+ ],
+ [
+ "▁Fi",
+ "l"
+ ],
+ [
+ "▁",
+ "Fil"
+ ],
+ [
+ ">",
+ ">"
+ ],
+ [
+ "▁pr",
+ "incip"
+ ],
+ [
+ "▁prin",
+ "cip"
+ ],
+ [
+ "Bas",
+ "e"
+ ],
+ [
+ "B",
+ "ase"
+ ],
+ [
+ "di",
+ "ct"
+ ],
+ [
+ "d",
+ "ict"
+ ],
+ [
+ "ve",
+ "rage"
+ ],
+ [
+ "ver",
+ "age"
+ ],
+ [
+ "ic",
+ "ient"
+ ],
+ [
+ "ici",
+ "ent"
+ ],
+ [
+ "I",
+ "F"
+ ],
+ [
+ "▁h",
+ "it"
+ ],
+ [
+ "▁hi",
+ "t"
+ ],
+ [
+ "▁",
+ "hit"
+ ],
+ [
+ "Pag",
+ "e"
+ ],
+ [
+ "P",
+ "age"
+ ],
+ [
+ "▁p",
+ "erm"
+ ],
+ [
+ "▁per",
+ "m"
+ ],
+ [
+ "▁pe",
+ "rm"
+ ],
+ [
+ "▁",
+ "perm"
+ ],
+ [
+ "ce",
+ "l"
+ ],
+ [
+ "c",
+ "el"
+ ],
+ [
+ "í",
+ "t"
+ ],
+ [
+ "▁ex",
+ "press"
+ ],
+ [
+ "▁exp",
+ "ress"
+ ],
+ [
+ "▁expr",
+ "ess"
+ ],
+ [
+ "▁",
+ "express"
+ ],
+ [
+ "▁ind",
+ "ic"
+ ],
+ [
+ "▁Se",
+ "ptember"
+ ],
+ [
+ "▁Sept",
+ "ember"
+ ],
+ [
+ "im",
+ "age"
+ ],
+ [
+ "ima",
+ "ge"
+ ],
+ [
+ "imag",
+ "e"
+ ],
+ [
+ "▁product",
+ "s"
+ ],
+ [
+ "▁",
+ "products"
+ ],
+ [
+ "▁m",
+ "edia"
+ ],
+ [
+ "▁med",
+ "ia"
+ ],
+ [
+ "▁medi",
+ "a"
+ ],
+ [
+ "▁",
+ "media"
+ ],
+ [
+ "ch",
+ "ange"
+ ],
+ [
+ "chan",
+ "ge"
+ ],
+ [
+ "ig",
+ "ger"
+ ],
+ [
+ "igg",
+ "er"
+ ],
+ [
+ "▁s",
+ "end"
+ ],
+ [
+ "▁se",
+ "nd"
+ ],
+ [
+ "▁sen",
+ "d"
+ ],
+ [
+ "▁",
+ "send"
+ ],
+ [
+ "la",
+ "st"
+ ],
+ [
+ "las",
+ "t"
+ ],
+ [
+ "l",
+ "ast"
+ ],
+ [
+ "min",
+ "g"
+ ],
+ [
+ "mi",
+ "ng"
+ ],
+ [
+ "m",
+ "ing"
+ ],
+ [
+ "p",
+ "a"
+ ],
+ [
+ "ua",
+ "ry"
+ ],
+ [
+ "uar",
+ "y"
+ ],
+ [
+ "u",
+ "ary"
+ ],
+ [
+ "▁spe",
+ "ak"
+ ],
+ [
+ "ны",
+ "й"
+ ],
+ [
+ "щ",
+ "е"
+ ],
+ [
+ "ys",
+ "is"
+ ],
+ [
+ "y",
+ "sis"
+ ],
+ [
+ "ly",
+ "ing"
+ ],
+ [
+ "l",
+ "ying"
+ ],
+ [
+ "▁",
+ "ч"
+ ],
+ [
+ "li",
+ "ke"
+ ],
+ [
+ "lik",
+ "e"
+ ],
+ [
+ "l",
+ "ike"
+ ],
+ [
+ "р",
+ "ы"
+ ],
+ [
+ "в",
+ "і"
+ ],
+ [
+ "▁M",
+ "ich"
+ ],
+ [
+ "▁Mic",
+ "h"
+ ],
+ [
+ "▁Mi",
+ "ch"
+ ],
+ [
+ "M",
+ "O"
+ ],
+ [
+ "▁J",
+ "ah"
+ ],
+ [
+ "▁Ja",
+ "h"
+ ],
+ [
+ "ens",
+ "ive"
+ ],
+ [
+ "▁sh",
+ "are"
+ ],
+ [
+ "▁shar",
+ "e"
+ ],
+ [
+ "▁sha",
+ "re"
+ ],
+ [
+ "▁",
+ "share"
+ ],
+ [
+ "▁develop",
+ "ment"
+ ],
+ [
+ "C",
+ "P"
+ ],
+ [
+ "sp",
+ "ec"
+ ],
+ [
+ "spe",
+ "c"
+ ],
+ [
+ "s",
+ "pec"
+ ],
+ [
+ "▁f",
+ "ast"
+ ],
+ [
+ "▁fa",
+ "st"
+ ],
+ [
+ "▁",
+ "fast"
+ ],
+ [
+ "he",
+ "t"
+ ],
+ [
+ "h",
+ "et"
+ ],
+ [
+ "H",
+ "O"
+ ],
+ [
+ "▁part",
+ "icip"
+ ],
+ [
+ "▁partic",
+ "ip"
+ ],
+ [
+ "▁parti",
+ "cip"
+ ],
+ [
+ "Bl",
+ "ock"
+ ],
+ [
+ "Blo",
+ "ck"
+ ],
+ [
+ "B",
+ "lock"
+ ],
+ [
+ "▁vi",
+ "ol"
+ ],
+ [
+ "▁fr",
+ "ame"
+ ],
+ [
+ "▁fra",
+ "me"
+ ],
+ [
+ "▁fram",
+ "e"
+ ],
+ [
+ "▁",
+ "frame"
+ ],
+ [
+ "▁qu",
+ "al"
+ ],
+ [
+ "▁q",
+ "ual"
+ ],
+ [
+ "▁",
+ "qual"
+ ],
+ [
+ "tr",
+ "e"
+ ],
+ [
+ "t",
+ "re"
+ ],
+ [
+ "▁",
+ "Ф"
+ ],
+ [
+ "▁to",
+ "ward"
+ ],
+ [
+ "▁tow",
+ "ard"
+ ],
+ [
+ "f",
+ "g"
+ ],
+ [
+ "Bo",
+ "x"
+ ],
+ [
+ "B",
+ "ox"
+ ],
+ [
+ "Col",
+ "umn"
+ ],
+ [
+ "▁mil",
+ "it"
+ ],
+ [
+ "▁mi",
+ "lit"
+ ],
+ [
+ "▁M",
+ "arch"
+ ],
+ [
+ "▁Mar",
+ "ch"
+ ],
+ [
+ "▁Marc",
+ "h"
+ ],
+ [
+ "▁var",
+ "ious"
+ ],
+ [
+ "▁vari",
+ "ous"
+ ],
+ [
+ "pa",
+ "ss"
+ ],
+ [
+ "pas",
+ "s"
+ ],
+ [
+ "p",
+ "ass"
+ ],
+ [
+ "▁P",
+ "ark"
+ ],
+ [
+ "▁Par",
+ "k"
+ ],
+ [
+ "▁B",
+ "en"
+ ],
+ [
+ "▁Be",
+ "n"
+ ],
+ [
+ "▁",
+ "Ben"
+ ],
+ [
+ "Fr",
+ "ame"
+ ],
+ [
+ "▁n",
+ "ormal"
+ ],
+ [
+ "▁nor",
+ "mal"
+ ],
+ [
+ "▁norm",
+ "al"
+ ],
+ [
+ "▁",
+ "normal"
+ ],
+ [
+ "op",
+ "en"
+ ],
+ [
+ "ope",
+ "n"
+ ],
+ [
+ "o",
+ "pen"
+ ],
+ [
+ "p",
+ "x"
+ ],
+ [
+ "▁ph",
+ "one"
+ ],
+ [
+ "▁",
+ "phone"
+ ],
+ [
+ "▁E",
+ "ven"
+ ],
+ [
+ "▁Ev",
+ "en"
+ ],
+ [
+ "▁Eve",
+ "n"
+ ],
+ [
+ "▁",
+ "Even"
+ ],
+ [
+ "▁m",
+ "a"
+ ],
+ [
+ "▁",
+ "ma"
+ ],
+ [
+ "ibr",
+ "ary"
+ ],
+ [
+ "St",
+ "art"
+ ],
+ [
+ "Star",
+ "t"
+ ],
+ [
+ "id",
+ "den"
+ ],
+ [
+ "idd",
+ "en"
+ ],
+ [
+ "rh",
+ "o"
+ ],
+ [
+ "r",
+ "ho"
+ ],
+ [
+ "gr",
+ "aph"
+ ],
+ [
+ "gra",
+ "ph"
+ ],
+ [
+ "g",
+ "raph"
+ ],
+ [
+ "ac",
+ "ing"
+ ],
+ [
+ "aci",
+ "ng"
+ ],
+ [
+ "a",
+ "cing"
+ ],
+ [
+ "'",
+ "."
+ ],
+ [
+ "ar",
+ "ter"
+ ],
+ [
+ "art",
+ "er"
+ ],
+ [
+ "arte",
+ "r"
+ ],
+ [
+ "me",
+ "s"
+ ],
+ [
+ "m",
+ "es"
+ ],
+ [
+ "in",
+ "st"
+ ],
+ [
+ "ins",
+ "t"
+ ],
+ [
+ "▁i",
+ "r"
+ ],
+ [
+ "▁",
+ "ir"
+ ],
+ [
+ "act",
+ "ive"
+ ],
+ [
+ "activ",
+ "e"
+ ],
+ [
+ "▁f",
+ "em"
+ ],
+ [
+ "▁fe",
+ "m"
+ ],
+ [
+ "▁",
+ "fem"
+ ],
+ [
+ "▁m",
+ "oved"
+ ],
+ [
+ "▁mov",
+ "ed"
+ ],
+ [
+ "▁move",
+ "d"
+ ],
+ [
+ "▁mo",
+ "ved"
+ ],
+ [
+ "▁st",
+ "ore"
+ ],
+ [
+ "▁stor",
+ "e"
+ ],
+ [
+ "▁sto",
+ "re"
+ ],
+ [
+ "▁",
+ "store"
+ ],
+ [
+ "▁p",
+ "rice"
+ ],
+ [
+ "▁pr",
+ "ice"
+ ],
+ [
+ "▁pri",
+ "ce"
+ ],
+ [
+ "▁",
+ "price"
+ ],
+ [
+ "\")",
+ "."
+ ],
+ [
+ "\"",
+ ")."
+ ],
+ [
+ "ber",
+ "g"
+ ],
+ [
+ "be",
+ "rg"
+ ],
+ [
+ "b",
+ "erg"
+ ],
+ [
+ "▁n",
+ "ov"
+ ],
+ [
+ "▁no",
+ "v"
+ ],
+ [
+ "▁",
+ "nov"
+ ],
+ [
+ "▁c",
+ "ard"
+ ],
+ [
+ "▁car",
+ "d"
+ ],
+ [
+ "▁ca",
+ "rd"
+ ],
+ [
+ "▁",
+ "card"
+ ],
+ [
+ "el",
+ "low"
+ ],
+ [
+ "ell",
+ "ow"
+ ],
+ [
+ "ello",
+ "w"
+ ],
+ [
+ "▁part",
+ "y"
+ ],
+ [
+ "▁par",
+ "ty"
+ ],
+ [
+ "▁",
+ "party"
+ ],
+ [
+ "▁M",
+ "or"
+ ],
+ [
+ "▁Mo",
+ "r"
+ ],
+ [
+ "ae",
+ "l"
+ ],
+ [
+ "a",
+ "el"
+ ],
+ [
+ "▁per",
+ "cent"
+ ],
+ [
+ "▁",
+ "percent"
+ ],
+ [
+ "▁tr",
+ "aining"
+ ],
+ [
+ "▁tra",
+ "ining"
+ ],
+ [
+ "▁train",
+ "ing"
+ ],
+ [
+ "▁",
+ "training"
+ ],
+ [
+ "▁in",
+ "g"
+ ],
+ [
+ "▁i",
+ "ng"
+ ],
+ [
+ "▁",
+ "ing"
+ ],
+ [
+ "im",
+ "er"
+ ],
+ [
+ "ime",
+ "r"
+ ],
+ [
+ "i",
+ "mer"
+ ],
+ [
+ "▁S",
+ "am"
+ ],
+ [
+ "▁Sa",
+ "m"
+ ],
+ [
+ "▁",
+ "Sam"
+ ],
+ [
+ "Def",
+ "ault"
+ ],
+ [
+ "▁f",
+ "uck"
+ ],
+ [
+ "▁fu",
+ "ck"
+ ],
+ [
+ "▁com",
+ "plete"
+ ],
+ [
+ "▁comp",
+ "lete"
+ ],
+ [
+ "▁complet",
+ "e"
+ ],
+ [
+ "▁compl",
+ "ete"
+ ],
+ [
+ "▁",
+ "complete"
+ ],
+ [
+ "ui",
+ "d"
+ ],
+ [
+ "u",
+ "id"
+ ],
+ [
+ "▁det",
+ "ails"
+ ],
+ [
+ "▁detail",
+ "s"
+ ],
+ [
+ "▁",
+ "details"
+ ],
+ [
+ "▁l",
+ "ed"
+ ],
+ [
+ "▁le",
+ "d"
+ ],
+ [
+ "▁",
+ "led"
+ ],
+ [
+ "Po",
+ "int"
+ ],
+ [
+ "P",
+ "oint"
+ ],
+ [
+ "▁C",
+ "ount"
+ ],
+ [
+ "▁Co",
+ "unt"
+ ],
+ [
+ "▁Coun",
+ "t"
+ ],
+ [
+ "▁Cou",
+ "nt"
+ ],
+ [
+ "▁",
+ "Count"
+ ],
+ [
+ "▁reg",
+ "ard"
+ ],
+ [
+ "z",
+ "o"
+ ],
+ [
+ "▁B",
+ "ro"
+ ],
+ [
+ "▁Br",
+ "o"
+ ],
+ [
+ "▁",
+ "Bro"
+ ],
+ [
+ "▁rec",
+ "ogn"
+ ],
+ [
+ "▁",
+ "recogn"
+ ],
+ [
+ "▁H",
+ "ol"
+ ],
+ [
+ "▁Ho",
+ "l"
+ ],
+ [
+ "▁",
+ "Hol"
+ ],
+ [
+ "U",
+ "M"
+ ],
+ [
+ "el",
+ "ement"
+ ],
+ [
+ "ele",
+ "ment"
+ ],
+ [
+ "elem",
+ "ent"
+ ],
+ [
+ "e",
+ "lement"
+ ],
+ [
+ "Mod",
+ "e"
+ ],
+ [
+ "Mo",
+ "de"
+ ],
+ [
+ "M",
+ "ode"
+ ],
+ [
+ "▁ex",
+ "am"
+ ],
+ [
+ "▁E",
+ "X"
+ ],
+ [
+ "▁",
+ "EX"
+ ],
+ [
+ "Im",
+ "age"
+ ],
+ [
+ "ver",
+ "se"
+ ],
+ [
+ "vers",
+ "e"
+ ],
+ [
+ "ri",
+ "ter"
+ ],
+ [
+ "rit",
+ "er"
+ ],
+ [
+ "rite",
+ "r"
+ ],
+ [
+ "r",
+ "iter"
+ ],
+ [
+ "so",
+ "ft"
+ ],
+ [
+ "s",
+ "oft"
+ ],
+ [
+ "▁int",
+ "rodu"
+ ],
+ [
+ "▁intro",
+ "du"
+ ],
+ [
+ "▁sur",
+ "pr"
+ ],
+ [
+ "Buf",
+ "fer"
+ ],
+ [
+ "Buff",
+ "er"
+ ],
+ [
+ "B",
+ "uffer"
+ ],
+ [
+ "le",
+ "ctor"
+ ],
+ [
+ "lect",
+ "or"
+ ],
+ [
+ "l",
+ "ector"
+ ],
+ [
+ "ar",
+ "en"
+ ],
+ [
+ "are",
+ "n"
+ ],
+ [
+ "a",
+ "ren"
+ ],
+ [
+ "an",
+ "ged"
+ ],
+ [
+ "ang",
+ "ed"
+ ],
+ [
+ "ange",
+ "d"
+ ],
+ [
+ "▁P",
+ "at"
+ ],
+ [
+ "▁Pa",
+ "t"
+ ],
+ [
+ "▁",
+ "Pat"
+ ],
+ [
+ "▁P",
+ "al"
+ ],
+ [
+ "▁Pa",
+ "l"
+ ],
+ [
+ "▁",
+ "Pal"
+ ],
+ [
+ "▁con",
+ "tr"
+ ],
+ [
+ "▁cont",
+ "r"
+ ],
+ [
+ "▁",
+ "contr"
+ ],
+ [
+ "Hand",
+ "ler"
+ ],
+ [
+ "Handle",
+ "r"
+ ],
+ [
+ "▁fe",
+ "atures"
+ ],
+ [
+ "▁feature",
+ "s"
+ ],
+ [
+ "▁feat",
+ "ures"
+ ],
+ [
+ "▁",
+ "features"
+ ],
+ [
+ "ip",
+ "le"
+ ],
+ [
+ "i",
+ "ple"
+ ],
+ [
+ "▁C",
+ "ON"
+ ],
+ [
+ "▁CO",
+ "N"
+ ],
+ [
+ "▁",
+ "CON"
+ ],
+ [
+ "Fi",
+ "l"
+ ],
+ [
+ "F",
+ "il"
+ ],
+ [
+ "▁P",
+ "ort"
+ ],
+ [
+ "▁Po",
+ "rt"
+ ],
+ [
+ "▁Por",
+ "t"
+ ],
+ [
+ "▁",
+ "Port"
+ ],
+ [
+ "▁th",
+ "inking"
+ ],
+ [
+ "▁think",
+ "ing"
+ ],
+ [
+ "▁thin",
+ "king"
+ ],
+ [
+ "do",
+ "c"
+ ],
+ [
+ "d",
+ "oc"
+ ],
+ [
+ "we",
+ "r"
+ ],
+ [
+ "w",
+ "er"
+ ],
+ [
+ "▁work",
+ "ed"
+ ],
+ [
+ "▁wor",
+ "ked"
+ ],
+ [
+ "P",
+ "C"
+ ],
+ [
+ "c",
+ "m"
+ ],
+ [
+ "da",
+ "t"
+ ],
+ [
+ "d",
+ "at"
+ ],
+ [
+ "PR",
+ "O"
+ ],
+ [
+ "P",
+ "RO"
+ ],
+ [
+ "▁E",
+ "very"
+ ],
+ [
+ "▁Ev",
+ "ery"
+ ],
+ [
+ "▁Ever",
+ "y"
+ ],
+ [
+ "▁Eve",
+ "ry"
+ ],
+ [
+ "▁",
+ "Every"
+ ],
+ [
+ "▁e",
+ "ra"
+ ],
+ [
+ "▁er",
+ "a"
+ ],
+ [
+ "▁",
+ "era"
+ ],
+ [
+ "▁F",
+ "irst"
+ ],
+ [
+ "▁",
+ "First"
+ ],
+ [
+ "g",
+ "n"
+ ],
+ [
+ "▁im",
+ "medi"
+ ],
+ [
+ "▁imm",
+ "edi"
+ ],
+ [
+ "ov",
+ "ember"
+ ],
+ [
+ "ove",
+ "mber"
+ ],
+ [
+ "ap",
+ "an"
+ ],
+ [
+ "apa",
+ "n"
+ ],
+ [
+ "a",
+ "pan"
+ ],
+ [
+ "▁ex",
+ "tra"
+ ],
+ [
+ "▁ext",
+ "ra"
+ ],
+ [
+ "▁extr",
+ "a"
+ ],
+ [
+ "▁",
+ "extra"
+ ],
+ [
+ "▁s",
+ "ection"
+ ],
+ [
+ "▁se",
+ "ction"
+ ],
+ [
+ "▁sect",
+ "ion"
+ ],
+ [
+ "▁",
+ "section"
+ ],
+ [
+ "▁J",
+ "une"
+ ],
+ [
+ "▁Jun",
+ "e"
+ ],
+ [
+ "▁Ju",
+ "ne"
+ ],
+ [
+ "▁v",
+ "ia"
+ ],
+ [
+ "▁vi",
+ "a"
+ ],
+ [
+ "▁",
+ "via"
+ ],
+ [
+ "▁g",
+ "one"
+ ],
+ [
+ "▁go",
+ "ne"
+ ],
+ [
+ "com",
+ "e"
+ ],
+ [
+ "co",
+ "me"
+ ],
+ [
+ "c",
+ "ome"
+ ],
+ [
+ "▁s",
+ "tri"
+ ],
+ [
+ "▁st",
+ "ri"
+ ],
+ [
+ "▁str",
+ "i"
+ ],
+ [
+ "▁",
+ "stri"
+ ],
+ [
+ "^",
+ "\\"
+ ],
+ [
+ "ant",
+ "ly"
+ ],
+ [
+ "▁ar",
+ "ch"
+ ],
+ [
+ "▁arc",
+ "h"
+ ],
+ [
+ "▁",
+ "arch"
+ ],
+ [
+ "S",
+ "ource"
+ ],
+ [
+ "▁con",
+ "v"
+ ],
+ [
+ "▁co",
+ "nv"
+ ],
+ [
+ "▁",
+ "conv"
+ ],
+ [
+ "▁L",
+ "ondon"
+ ],
+ [
+ "▁Lond",
+ "on"
+ ],
+ [
+ "▁",
+ "London"
+ ],
+ [
+ "Num",
+ "ber"
+ ],
+ [
+ "N",
+ "umber"
+ ],
+ [
+ "▁quest",
+ "ions"
+ ],
+ [
+ "▁question",
+ "s"
+ ],
+ [
+ "an",
+ "did"
+ ],
+ [
+ "and",
+ "id"
+ ],
+ [
+ "▁play",
+ "ed"
+ ],
+ [
+ "en",
+ "v"
+ ],
+ [
+ "e",
+ "nv"
+ ],
+ [
+ "▁Sch",
+ "ool"
+ ],
+ [
+ "▁nat",
+ "ural"
+ ],
+ [
+ "▁natur",
+ "al"
+ ],
+ [
+ "▁",
+ "natural"
+ ],
+ [
+ "ca",
+ "n"
+ ],
+ [
+ "c",
+ "an"
+ ],
+ [
+ "▁ne",
+ "ws"
+ ],
+ [
+ "▁new",
+ "s"
+ ],
+ [
+ "▁",
+ "news"
+ ],
+ [
+ "D",
+ "R"
+ ],
+ [
+ "▁c",
+ "hall"
+ ],
+ [
+ "▁ch",
+ "all"
+ ],
+ [
+ "▁cha",
+ "ll"
+ ],
+ [
+ "▁S",
+ "oc"
+ ],
+ [
+ "▁So",
+ "c"
+ ],
+ [
+ "▁",
+ "э"
+ ],
+ [
+ "▁att",
+ "empt"
+ ],
+ [
+ "*",
+ "}"
+ ],
+ [
+ "N",
+ "ull"
+ ],
+ [
+ "ro",
+ "te"
+ ],
+ [
+ "rot",
+ "e"
+ ],
+ [
+ "r",
+ "ote"
+ ],
+ [
+ "▁b",
+ "i"
+ ],
+ [
+ "▁",
+ "bi"
+ ],
+ [
+ "▁wr",
+ "itten"
+ ],
+ [
+ "▁writ",
+ "ten"
+ ],
+ [
+ "▁",
+ "written"
+ ],
+ [
+ "▁bl",
+ "ood"
+ ],
+ [
+ "▁blo",
+ "od"
+ ],
+ [
+ "▁happ",
+ "ened"
+ ],
+ [
+ "▁happen",
+ "ed"
+ ],
+ [
+ "▁c",
+ "ause"
+ ],
+ [
+ "▁caus",
+ "e"
+ ],
+ [
+ "▁ca",
+ "use"
+ ],
+ [
+ "as",
+ "hing"
+ ],
+ [
+ "ash",
+ "ing"
+ ],
+ [
+ "ashi",
+ "ng"
+ ],
+ [
+ "▁Will",
+ "iam"
+ ],
+ [
+ "ad",
+ "em"
+ ],
+ [
+ "ade",
+ "m"
+ ],
+ [
+ "a",
+ "dem"
+ ],
+ [
+ "▁b",
+ "rought"
+ ],
+ [
+ "▁br",
+ "ought"
+ ],
+ [
+ "▁dis",
+ "play"
+ ],
+ [
+ "▁displ",
+ "ay"
+ ],
+ [
+ "▁disp",
+ "lay"
+ ],
+ [
+ "▁",
+ "display"
+ ],
+ [
+ "im",
+ "a"
+ ],
+ [
+ "i",
+ "ma"
+ ],
+ [
+ "▁fin",
+ "ally"
+ ],
+ [
+ "▁final",
+ "ly"
+ ],
+ [
+ "ta",
+ "b"
+ ],
+ [
+ "t",
+ "ab"
+ ],
+ [
+ "▁return",
+ "ed"
+ ],
+ [
+ "ны",
+ "х"
+ ],
+ [
+ "ni",
+ "e"
+ ],
+ [
+ "n",
+ "ie"
+ ],
+ [
+ "▁",
+ "q"
+ ],
+ [
+ "▁h",
+ "ers"
+ ],
+ [
+ "▁he",
+ "rs"
+ ],
+ [
+ "▁her",
+ "s"
+ ],
+ [
+ "▁P",
+ "re"
+ ],
+ [
+ "▁Pr",
+ "e"
+ ],
+ [
+ "▁",
+ "Pre"
+ ],
+ [
+ "▁d",
+ "ou"
+ ],
+ [
+ "▁do",
+ "u"
+ ],
+ [
+ "buf",
+ "fer"
+ ],
+ [
+ "buff",
+ "er"
+ ],
+ [
+ "b",
+ "uffer"
+ ],
+ [
+ "▁eff",
+ "ort"
+ ],
+ [
+ "ain",
+ "e"
+ ],
+ [
+ "ai",
+ "ne"
+ ],
+ [
+ "a",
+ "ine"
+ ],
+ [
+ "x",
+ "y"
+ ],
+ [
+ "▁his",
+ "tor"
+ ],
+ [
+ "▁hist",
+ "or"
+ ],
+ [
+ "en",
+ "u"
+ ],
+ [
+ "e",
+ "nu"
+ ],
+ [
+ "▁ar",
+ "riv"
+ ],
+ [
+ "▁arr",
+ "iv"
+ ],
+ [
+ "▁D",
+ "em"
+ ],
+ [
+ "▁De",
+ "m"
+ ],
+ [
+ "▁",
+ "Dem"
+ ],
+ [
+ "▁f",
+ "avor"
+ ],
+ [
+ "▁fa",
+ "vor"
+ ],
+ [
+ "▁fav",
+ "or"
+ ],
+ [
+ "▁hand",
+ "le"
+ ],
+ [
+ "▁",
+ "handle"
+ ],
+ [
+ "SE",
+ "T"
+ ],
+ [
+ "S",
+ "ET"
+ ],
+ [
+ "▁P",
+ "ublic"
+ ],
+ [
+ "▁Pub",
+ "lic"
+ ],
+ [
+ "▁Pu",
+ "blic"
+ ],
+ [
+ "▁",
+ "Public"
+ ],
+ [
+ "ru",
+ "pt"
+ ],
+ [
+ "rup",
+ "t"
+ ],
+ [
+ "r",
+ "upt"
+ ],
+ [
+ "▁u",
+ "r"
+ ],
+ [
+ "▁",
+ "ur"
+ ],
+ [
+ "▁for",
+ "ce"
+ ],
+ [
+ "▁",
+ "force"
+ ],
+ [
+ "▁é",
+ "s"
+ ],
+ [
+ "▁",
+ "és"
+ ],
+ [
+ "ub",
+ "e"
+ ],
+ [
+ "u",
+ "be"
+ ],
+ [
+ "Pr",
+ "e"
+ ],
+ [
+ "P",
+ "re"
+ ],
+ [
+ "р",
+ "і"
+ ],
+ [
+ "in",
+ "y"
+ ],
+ [
+ "i",
+ "ny"
+ ],
+ [
+ "th",
+ "eta"
+ ],
+ [
+ "the",
+ "ta"
+ ],
+ [
+ "is",
+ "f"
+ ],
+ [
+ "i",
+ "sf"
+ ],
+ [
+ "▁n",
+ "ational"
+ ],
+ [
+ "▁nat",
+ "ional"
+ ],
+ [
+ "▁nation",
+ "al"
+ ],
+ [
+ "Equ",
+ "al"
+ ],
+ [
+ "Eq",
+ "ual"
+ ],
+ [
+ "E",
+ "qual"
+ ],
+ [
+ "ren",
+ "ch"
+ ],
+ [
+ "▁w",
+ "ife"
+ ],
+ [
+ "▁c",
+ "apt"
+ ],
+ [
+ "▁cap",
+ "t"
+ ],
+ [
+ "▁ca",
+ "pt"
+ ],
+ [
+ "▁In",
+ "ter"
+ ],
+ [
+ "▁Int",
+ "er"
+ ],
+ [
+ "▁",
+ "Inter"
+ ],
+ [
+ "ta",
+ "u"
+ ],
+ [
+ "t",
+ "au"
+ ],
+ [
+ "▁s",
+ "leep"
+ ],
+ [
+ "▁sle",
+ "ep"
+ ],
+ [
+ "▁",
+ "sleep"
+ ],
+ [
+ "../",
+ "../"
+ ],
+ [
+ "▁iss",
+ "ue"
+ ],
+ [
+ "▁",
+ "issue"
+ ],
+ [
+ "▁m",
+ "ember"
+ ],
+ [
+ "▁me",
+ "mber"
+ ],
+ [
+ "▁mem",
+ "ber"
+ ],
+ [
+ "▁",
+ "member"
+ ],
+ [
+ "▁a",
+ "wait"
+ ],
+ [
+ "▁aw",
+ "ait"
+ ],
+ [
+ "▁",
+ "await"
+ ],
+ [
+ "▁D",
+ "an"
+ ],
+ [
+ "▁Da",
+ "n"
+ ],
+ [
+ "▁",
+ "Dan"
+ ],
+ [
+ "z",
+ "i"
+ ],
+ [
+ "in",
+ "ate"
+ ],
+ [
+ "ina",
+ "te"
+ ],
+ [
+ "i",
+ "nate"
+ ],
+ [
+ "▁s",
+ "ym"
+ ],
+ [
+ "▁sy",
+ "m"
+ ],
+ [
+ "▁",
+ "sym"
+ ],
+ [
+ "ch",
+ "an"
+ ],
+ [
+ "cha",
+ "n"
+ ],
+ [
+ "c",
+ "han"
+ ],
+ [
+ "▁J",
+ "ack"
+ ],
+ [
+ "▁Jac",
+ "k"
+ ],
+ [
+ "▁Ja",
+ "ck"
+ ],
+ [
+ "▁",
+ "Jack"
+ ],
+ [
+ "▁Eng",
+ "lish"
+ ],
+ [
+ "▁",
+ "English"
+ ],
+ [
+ "▁s",
+ "z"
+ ],
+ [
+ "▁",
+ "sz"
+ ],
+ [
+ "rib",
+ "utes"
+ ],
+ [
+ "ribut",
+ "es"
+ ],
+ [
+ "ribute",
+ "s"
+ ],
+ [
+ "ribu",
+ "tes"
+ ],
+ [
+ "▁i",
+ "gn"
+ ],
+ [
+ "▁ig",
+ "n"
+ ],
+ [
+ "▁",
+ "ign"
+ ],
+ [
+ "á",
+ "l"
+ ],
+ [
+ "▁app",
+ "ear"
+ ],
+ [
+ "▁appe",
+ "ar"
+ ],
+ [
+ "ra",
+ "d"
+ ],
+ [
+ "r",
+ "ad"
+ ],
+ [
+ "id",
+ "ge"
+ ],
+ [
+ "▁co",
+ "uple"
+ ],
+ [
+ "▁cou",
+ "ple"
+ ],
+ [
+ "▁coup",
+ "le"
+ ],
+ [
+ "▁s",
+ "hip"
+ ],
+ [
+ "▁sh",
+ "ip"
+ ],
+ [
+ "▁",
+ "ship"
+ ],
+ [
+ "li",
+ "g"
+ ],
+ [
+ "l",
+ "ig"
+ ],
+ [
+ "we",
+ "b"
+ ],
+ [
+ "w",
+ "eb"
+ ],
+ [
+ "▁us",
+ "ually"
+ ],
+ [
+ "▁usual",
+ "ly"
+ ],
+ [
+ "▁re",
+ "ady"
+ ],
+ [
+ "▁read",
+ "y"
+ ],
+ [
+ "▁",
+ "ready"
+ ],
+ [
+ "▁v",
+ "ill"
+ ],
+ [
+ "▁vi",
+ "ll"
+ ],
+ [
+ "▁vil",
+ "l"
+ ],
+ [
+ "▁W",
+ "hy"
+ ],
+ [
+ "▁Wh",
+ "y"
+ ],
+ [
+ "▁",
+ "Why"
+ ],
+ [
+ "eb",
+ "ru"
+ ],
+ [
+ "e",
+ "bru"
+ ],
+ [
+ "▁g",
+ "rad"
+ ],
+ [
+ "▁gr",
+ "ad"
+ ],
+ [
+ "▁gra",
+ "d"
+ ],
+ [
+ "▁",
+ "grad"
+ ],
+ [
+ "or",
+ "ds"
+ ],
+ [
+ "ord",
+ "s"
+ ],
+ [
+ "▁in",
+ "f"
+ ],
+ [
+ "▁i",
+ "nf"
+ ],
+ [
+ "▁",
+ "inf"
+ ],
+ [
+ "▁l",
+ "oss"
+ ],
+ [
+ "▁lo",
+ "ss"
+ ],
+ [
+ "▁los",
+ "s"
+ ],
+ [
+ "▁",
+ "loss"
+ ],
+ [
+ "▁o",
+ "d"
+ ],
+ [
+ "▁",
+ "od"
+ ],
+ [
+ "▁Ph",
+ "il"
+ ],
+ [
+ "▁",
+ "Phil"
+ ],
+ [
+ "ser",
+ "ver"
+ ],
+ [
+ "serv",
+ "er"
+ ],
+ [
+ "serve",
+ "r"
+ ],
+ [
+ "▁U",
+ "p"
+ ],
+ [
+ "▁",
+ "Up"
+ ],
+ [
+ "▁b",
+ "uff"
+ ],
+ [
+ "▁bu",
+ "ff"
+ ],
+ [
+ "▁buf",
+ "f"
+ ],
+ [
+ "▁",
+ "buff"
+ ],
+ [
+ "▁fil",
+ "ename"
+ ],
+ [
+ "▁file",
+ "name"
+ ],
+ [
+ "▁",
+ "filename"
+ ],
+ [
+ "AB",
+ "LE"
+ ],
+ [
+ "it",
+ "ing"
+ ],
+ [
+ "iti",
+ "ng"
+ ],
+ [
+ "i",
+ "ting"
+ ],
+ [
+ "ef",
+ "ore"
+ ],
+ [
+ "e",
+ "fore"
+ ],
+ [
+ "()",
+ "->"
+ ],
+ [
+ "(",
+ ")->"
+ ],
+ [
+ "▁cond",
+ "itions"
+ ],
+ [
+ "▁condition",
+ "s"
+ ],
+ [
+ "▁",
+ "conditions"
+ ],
+ [
+ "v",
+ "m"
+ ],
+ [
+ "el",
+ "d"
+ ],
+ [
+ "e",
+ "ld"
+ ],
+ [
+ "it",
+ "z"
+ ],
+ [
+ "i",
+ "tz"
+ ],
+ [
+ "▁Tr",
+ "ans"
+ ],
+ [
+ "▁Tra",
+ "ns"
+ ],
+ [
+ "▁",
+ "Trans"
+ ],
+ [
+ "▁w",
+ "eight"
+ ],
+ [
+ "▁we",
+ "ight"
+ ],
+ [
+ "▁weigh",
+ "t"
+ ],
+ [
+ "▁",
+ "weight"
+ ],
+ [
+ "▁high",
+ "er"
+ ],
+ [
+ "▁hig",
+ "her"
+ ],
+ [
+ "▁r",
+ "ate"
+ ],
+ [
+ "▁rat",
+ "e"
+ ],
+ [
+ "▁ra",
+ "te"
+ ],
+ [
+ "▁",
+ "rate"
+ ],
+ [
+ "▁acc",
+ "om"
+ ],
+ [
+ "▁ac",
+ "com"
+ ],
+ [
+ "vi",
+ "der"
+ ],
+ [
+ "vid",
+ "er"
+ ],
+ [
+ "v",
+ "ider"
+ ],
+ [
+ "O",
+ "M"
+ ],
+ [
+ "▁w",
+ "ays"
+ ],
+ [
+ "▁way",
+ "s"
+ ],
+ [
+ "▁wa",
+ "ys"
+ ],
+ [
+ "▁",
+ "ways"
+ ],
+ [
+ "com",
+ "ing"
+ ],
+ [
+ "co",
+ "ming"
+ ],
+ [
+ "c",
+ "oming"
+ ],
+ [
+ "▁l",
+ "ock"
+ ],
+ [
+ "▁loc",
+ "k"
+ ],
+ [
+ "▁lo",
+ "ck"
+ ],
+ [
+ "▁",
+ "lock"
+ ],
+ [
+ "▁e",
+ "tc"
+ ],
+ [
+ "▁et",
+ "c"
+ ],
+ [
+ "▁",
+ "etc"
+ ],
+ [
+ "▁a",
+ "vec"
+ ],
+ [
+ "▁av",
+ "ec"
+ ],
+ [
+ "▁ave",
+ "c"
+ ],
+ [
+ "▁t",
+ "akes"
+ ],
+ [
+ "▁take",
+ "s"
+ ],
+ [
+ "▁tak",
+ "es"
+ ],
+ [
+ "▁ta",
+ "kes"
+ ],
+ [
+ "▁C",
+ "har"
+ ],
+ [
+ "▁Ch",
+ "ar"
+ ],
+ [
+ "▁Cha",
+ "r"
+ ],
+ [
+ "▁",
+ "Char"
+ ],
+ [
+ "▁N",
+ "ovember"
+ ],
+ [
+ "▁Nov",
+ "ember"
+ ],
+ [
+ "m",
+ "ethod"
+ ],
+ [
+ "▁A",
+ "ustral"
+ ],
+ [
+ "▁Aust",
+ "ral"
+ ],
+ [
+ "▁",
+ "Austral"
+ ],
+ [
+ "▁Amer",
+ "ica"
+ ],
+ [
+ "▁",
+ "America"
+ ],
+ [
+ "lo",
+ "ng"
+ ],
+ [
+ "lon",
+ "g"
+ ],
+ [
+ "l",
+ "ong"
+ ],
+ [
+ "ce",
+ "mber"
+ ],
+ [
+ "c",
+ "ember"
+ ],
+ [
+ "▁polit",
+ "ical"
+ ],
+ [
+ "fl",
+ "ow"
+ ],
+ [
+ "f",
+ "low"
+ ],
+ [
+ "▁may",
+ "be"
+ ],
+ [
+ "▁",
+ "maybe"
+ ],
+ [
+ "▁a",
+ "mb"
+ ],
+ [
+ "▁am",
+ "b"
+ ],
+ [
+ "▁",
+ "amb"
+ ],
+ [
+ "La",
+ "yout"
+ ],
+ [
+ "L",
+ "ayout"
+ ],
+ [
+ "il",
+ "ed"
+ ],
+ [
+ "ile",
+ "d"
+ ],
+ [
+ "i",
+ "led"
+ ],
+ [
+ "om",
+ "en"
+ ],
+ [
+ "ome",
+ "n"
+ ],
+ [
+ "o",
+ "men"
+ ],
+ [
+ "ol",
+ "a"
+ ],
+ [
+ "o",
+ "la"
+ ],
+ [
+ "ic",
+ "ip"
+ ],
+ [
+ "ici",
+ "p"
+ ],
+ [
+ "i",
+ "cip"
+ ],
+ [
+ "part",
+ "ial"
+ ],
+ [
+ "Tr",
+ "ue"
+ ],
+ [
+ "▁f",
+ "loor"
+ ],
+ [
+ "▁fl",
+ "oor"
+ ],
+ [
+ "▁flo",
+ "or"
+ ],
+ [
+ "▁",
+ "floor"
+ ],
+ [
+ "▁D",
+ "ef"
+ ],
+ [
+ "▁De",
+ "f"
+ ],
+ [
+ "▁",
+ "Def"
+ ],
+ [
+ "▁conc",
+ "ern"
+ ],
+ [
+ "▁conce",
+ "rn"
+ ],
+ [
+ "▁concer",
+ "n"
+ ],
+ [
+ "y",
+ "r"
+ ],
+ [
+ "▁sh",
+ "ows"
+ ],
+ [
+ "▁show",
+ "s"
+ ],
+ [
+ "i",
+ "h"
+ ],
+ [
+ "▁an",
+ "swer"
+ ],
+ [
+ "▁answ",
+ "er"
+ ],
+ [
+ "▁ans",
+ "wer"
+ ],
+ [
+ "▁",
+ "answer"
+ ],
+ [
+ "ac",
+ "c"
+ ],
+ [
+ "a",
+ "cc"
+ ],
+ [
+ "▁b",
+ "all"
+ ],
+ [
+ "▁bal",
+ "l"
+ ],
+ [
+ "▁ba",
+ "ll"
+ ],
+ [
+ "▁",
+ "ball"
+ ],
+ [
+ "▁R",
+ "ev"
+ ],
+ [
+ "▁Re",
+ "v"
+ ],
+ [
+ "▁",
+ "Rev"
+ ],
+ [
+ "▁s",
+ "un"
+ ],
+ [
+ "▁su",
+ "n"
+ ],
+ [
+ "▁",
+ "sun"
+ ],
+ [
+ "▁quick",
+ "ly"
+ ],
+ [
+ "▁s",
+ "omet"
+ ],
+ [
+ "▁so",
+ "met"
+ ],
+ [
+ "▁some",
+ "t"
+ ],
+ [
+ "▁som",
+ "et"
+ ],
+ [
+ "ment",
+ "e"
+ ],
+ [
+ "me",
+ "nte"
+ ],
+ [
+ "men",
+ "te"
+ ],
+ [
+ "m",
+ "ente"
+ ],
+ [
+ "▁M",
+ "al"
+ ],
+ [
+ "▁Ma",
+ "l"
+ ],
+ [
+ "▁",
+ "Mal"
+ ],
+ [
+ "und",
+ "red"
+ ],
+ [
+ "▁iss",
+ "ues"
+ ],
+ [
+ "▁issue",
+ "s"
+ ],
+ [
+ "▁",
+ "issues"
+ ],
+ [
+ "ec",
+ "ause"
+ ],
+ [
+ "eca",
+ "use"
+ ],
+ [
+ "pe",
+ "s"
+ ],
+ [
+ "p",
+ "es"
+ ],
+ [
+ "▁p",
+ "layer"
+ ],
+ [
+ "▁pl",
+ "ayer"
+ ],
+ [
+ "▁play",
+ "er"
+ ],
+ [
+ "▁",
+ "player"
+ ],
+ [
+ "▁par",
+ "ents"
+ ],
+ [
+ "▁parent",
+ "s"
+ ],
+ [
+ "▁",
+ "parents"
+ ],
+ [
+ "▁pop",
+ "ular"
+ ],
+ [
+ "▁popula",
+ "r"
+ ],
+ [
+ "▁popul",
+ "ar"
+ ],
+ [
+ "▁m",
+ "ode"
+ ],
+ [
+ "▁mod",
+ "e"
+ ],
+ [
+ "▁mo",
+ "de"
+ ],
+ [
+ "▁",
+ "mode"
+ ],
+ [
+ "▁m",
+ "ention"
+ ],
+ [
+ "▁ment",
+ "ion"
+ ],
+ [
+ "N",
+ "E"
+ ],
+ [
+ "Lo",
+ "ad"
+ ],
+ [
+ "L",
+ "oad"
+ ],
+ [
+ "▁reg",
+ "ular"
+ ],
+ [
+ "▁regul",
+ "ar"
+ ],
+ [
+ "▁",
+ "regular"
+ ],
+ [
+ "ave",
+ "d"
+ ],
+ [
+ "av",
+ "ed"
+ ],
+ [
+ "a",
+ "ved"
+ ],
+ [
+ "?",
+ ":"
+ ],
+ [
+ "ye",
+ "ar"
+ ],
+ [
+ "y",
+ "ear"
+ ],
+ [
+ "fun",
+ "c"
+ ],
+ [
+ "fu",
+ "nc"
+ ],
+ [
+ "f",
+ "unc"
+ ],
+ [
+ "▁per",
+ "formance"
+ ],
+ [
+ "▁perform",
+ "ance"
+ ],
+ [
+ "▁J",
+ "uly"
+ ],
+ [
+ "▁Jul",
+ "y"
+ ],
+ [
+ "▁Ju",
+ "ly"
+ ],
+ [
+ "th",
+ "ern"
+ ],
+ [
+ "ther",
+ "n"
+ ],
+ [
+ "the",
+ "rn"
+ ],
+ [
+ "▁we",
+ "bsite"
+ ],
+ [
+ "▁webs",
+ "ite"
+ ],
+ [
+ "▁web",
+ "site"
+ ],
+ [
+ "fo",
+ "rd"
+ ],
+ [
+ "for",
+ "d"
+ ],
+ [
+ "f",
+ "ord"
+ ],
+ [
+ "P",
+ "R"
+ ],
+ [
+ "el",
+ "a"
+ ],
+ [
+ "e",
+ "la"
+ ],
+ [
+ "le",
+ "vel"
+ ],
+ [
+ "lev",
+ "el"
+ ],
+ [
+ "l",
+ "evel"
+ ],
+ [
+ "ui",
+ "t"
+ ],
+ [
+ "u",
+ "it"
+ ],
+ [
+ "fl",
+ "ags"
+ ],
+ [
+ "flag",
+ "s"
+ ],
+ [
+ "▁w",
+ "orth"
+ ],
+ [
+ "▁wor",
+ "th"
+ ],
+ [
+ "▁",
+ "worth"
+ ],
+ [
+ "▁cor",
+ "respon"
+ ],
+ [
+ "▁Brit",
+ "ish"
+ ],
+ [
+ "si",
+ "m"
+ ],
+ [
+ "s",
+ "im"
+ ],
+ [
+ "▁al",
+ "one"
+ ],
+ [
+ "▁",
+ "alone"
+ ],
+ [
+ "▁h",
+ "ar"
+ ],
+ [
+ "▁ha",
+ "r"
+ ],
+ [
+ "▁",
+ "har"
+ ],
+ [
+ "▁o",
+ "nes"
+ ],
+ [
+ "▁on",
+ "es"
+ ],
+ [
+ "▁one",
+ "s"
+ ],
+ [
+ "▁",
+ "ones"
+ ],
+ [
+ "ob",
+ "ile"
+ ],
+ [
+ "obi",
+ "le"
+ ],
+ [
+ "obil",
+ "e"
+ ],
+ [
+ "▁d",
+ "ru"
+ ],
+ [
+ "▁dr",
+ "u"
+ ],
+ [
+ "▁",
+ "dru"
+ ],
+ [
+ "ch",
+ "i"
+ ],
+ [
+ "c",
+ "hi"
+ ],
+ [
+ "▁D",
+ "avid"
+ ],
+ [
+ "▁Dav",
+ "id"
+ ],
+ [
+ "▁Da",
+ "vid"
+ ],
+ [
+ "▁",
+ "David"
+ ],
+ [
+ "▁proble",
+ "ms"
+ ],
+ [
+ "▁problem",
+ "s"
+ ],
+ [
+ "▁col",
+ "umn"
+ ],
+ [
+ "▁",
+ "column"
+ ],
+ [
+ "()",
+ ";\r"
+ ],
+ [
+ "();",
+ "\r"
+ ],
+ [
+ "(",
+ ");\r"
+ ],
+ [
+ "Z",
+ "E"
+ ],
+ [
+ "▁re",
+ "lig"
+ ],
+ [
+ "▁rel",
+ "ig"
+ ],
+ [
+ "▁reli",
+ "g"
+ ],
+ [
+ "olog",
+ "ical"
+ ],
+ [
+ "▁reg",
+ "ion"
+ ],
+ [
+ "▁",
+ "region"
+ ],
+ [
+ "ad",
+ "y"
+ ],
+ [
+ "a",
+ "dy"
+ ],
+ [
+ "I",
+ "O"
+ ],
+ [
+ "an",
+ "der"
+ ],
+ [
+ "and",
+ "er"
+ ],
+ [
+ "ande",
+ "r"
+ ],
+ [
+ "a",
+ "nder"
+ ],
+ [
+ "Ne",
+ "t"
+ ],
+ [
+ "N",
+ "et"
+ ],
+ [
+ "▁bu",
+ "ilt"
+ ],
+ [
+ "▁",
+ "built"
+ ],
+ [
+ "▁inst",
+ "all"
+ ],
+ [
+ "▁",
+ "install"
+ ],
+ [
+ "▁appro",
+ "ach"
+ ],
+ [
+ "C",
+ "ur"
+ ],
+ [
+ "▁f",
+ "ine"
+ ],
+ [
+ "▁fin",
+ "e"
+ ],
+ [
+ "▁fi",
+ "ne"
+ ],
+ [
+ "▁talk",
+ "ing"
+ ],
+ [
+ "▁tal",
+ "king"
+ ],
+ [
+ "▁ch",
+ "anges"
+ ],
+ [
+ "▁chang",
+ "es"
+ ],
+ [
+ "▁change",
+ "s"
+ ],
+ [
+ "▁",
+ "changes"
+ ],
+ [
+ "St",
+ "yle"
+ ],
+ [
+ "▁M",
+ "art"
+ ],
+ [
+ "▁Mar",
+ "t"
+ ],
+ [
+ "▁Ma",
+ "rt"
+ ],
+ [
+ "▁",
+ "Mart"
+ ],
+ [
+ "л",
+ "ю"
+ ],
+ [
+ "res",
+ "ponse"
+ ],
+ [
+ "respon",
+ "se"
+ ],
+ [
+ "respons",
+ "e"
+ ],
+ [
+ "te",
+ "ger"
+ ],
+ [
+ "{",
+ "\r"
+ ],
+ [
+ "ir",
+ "it"
+ ],
+ [
+ "iri",
+ "t"
+ ],
+ [
+ "i",
+ "rit"
+ ],
+ [
+ "▁prote",
+ "cted"
+ ],
+ [
+ "▁protect",
+ "ed"
+ ],
+ [
+ "▁",
+ "protected"
+ ],
+ [
+ "▁re",
+ "le"
+ ],
+ [
+ "▁r",
+ "ele"
+ ],
+ [
+ "▁rel",
+ "e"
+ ],
+ [
+ "er",
+ "ship"
+ ],
+ [
+ "ers",
+ "hip"
+ ],
+ [
+ "те",
+ "ль"
+ ],
+ [
+ "тел",
+ "ь"
+ ],
+ [
+ "un",
+ "signed"
+ ],
+ [
+ "uns",
+ "igned"
+ ],
+ [
+ "ial",
+ "ize"
+ ],
+ [
+ "▁htt",
+ "ps"
+ ],
+ [
+ "▁http",
+ "s"
+ ],
+ [
+ "▁",
+ "https"
+ ],
+ [
+ "T",
+ "ag"
+ ],
+ [
+ "▁$",
+ "("
+ ],
+ [
+ "▁",
+ "$("
+ ],
+ [
+ "mo",
+ "re"
+ ],
+ [
+ "mor",
+ "e"
+ ],
+ [
+ "m",
+ "ore"
+ ],
+ [
+ "ype",
+ "s"
+ ],
+ [
+ "yp",
+ "es"
+ ],
+ [
+ "y",
+ "pes"
+ ],
+ [
+ "▁st",
+ "ream"
+ ],
+ [
+ "▁stre",
+ "am"
+ ],
+ [
+ "▁",
+ "stream"
+ ],
+ [
+ "et",
+ "ch"
+ ],
+ [
+ "etc",
+ "h"
+ ],
+ [
+ "▁eng",
+ "ine"
+ ],
+ [
+ "▁",
+ "engine"
+ ],
+ [
+ "K",
+ "E"
+ ],
+ [
+ "cm",
+ "d"
+ ],
+ [
+ "c",
+ "md"
+ ],
+ [
+ "sc",
+ "ript"
+ ],
+ [
+ "scri",
+ "pt"
+ ],
+ [
+ "scr",
+ "ipt"
+ ],
+ [
+ "s",
+ "cript"
+ ],
+ [
+ "tt",
+ "p"
+ ],
+ [
+ "t",
+ "tp"
+ ],
+ [
+ "▁a",
+ "void"
+ ],
+ [
+ "▁av",
+ "oid"
+ ],
+ [
+ "▁t",
+ "err"
+ ],
+ [
+ "▁te",
+ "rr"
+ ],
+ [
+ "▁ter",
+ "r"
+ ],
+ [
+ "▁r",
+ "ock"
+ ],
+ [
+ "▁ro",
+ "ck"
+ ],
+ [
+ "▁",
+ "rock"
+ ],
+ [
+ "▁f",
+ "ul"
+ ],
+ [
+ "▁fu",
+ "l"
+ ],
+ [
+ "▁",
+ "ful"
+ ],
+ [
+ "Up",
+ "date"
+ ],
+ [
+ "▁env",
+ "ironment"
+ ],
+ [
+ "▁environ",
+ "ment"
+ ],
+ [
+ "▁",
+ "environment"
+ ],
+ [
+ "▁p",
+ "rec"
+ ],
+ [
+ "▁pre",
+ "c"
+ ],
+ [
+ "▁pr",
+ "ec"
+ ],
+ [
+ "▁",
+ "prec"
+ ],
+ [
+ "▁с",
+ "а"
+ ],
+ [
+ "▁",
+ "са"
+ ],
+ [
+ "▁c",
+ "ases"
+ ],
+ [
+ "▁case",
+ "s"
+ ],
+ [
+ "▁cas",
+ "es"
+ ],
+ [
+ "▁ca",
+ "ses"
+ ],
+ [
+ "▁",
+ "cases"
+ ],
+ [
+ "▁off",
+ "set"
+ ],
+ [
+ "▁",
+ "offset"
+ ],
+ [
+ "▁r",
+ "ais"
+ ],
+ [
+ "▁ra",
+ "is"
+ ],
+ [
+ "▁",
+ "rais"
+ ],
+ [
+ "li",
+ "b"
+ ],
+ [
+ "l",
+ "ib"
+ ],
+ [
+ "ée",
+ "s"
+ ],
+ [
+ "é",
+ "es"
+ ],
+ [
+ "a",
+ "a"
+ ],
+ [
+ "y",
+ "t"
+ ],
+ [
+ "▁a",
+ "rr"
+ ],
+ [
+ "▁ar",
+ "r"
+ ],
+ [
+ "▁",
+ "arr"
+ ],
+ [
+ "opy",
+ "right"
+ ],
+ [
+ "f",
+ "irst"
+ ],
+ [
+ "▁u",
+ "til"
+ ],
+ [
+ "▁ut",
+ "il"
+ ],
+ [
+ "▁",
+ "util"
+ ],
+ [
+ "▁fe",
+ "ature"
+ ],
+ [
+ "▁feat",
+ "ure"
+ ],
+ [
+ "▁",
+ "feature"
+ ],
+ [
+ "pos",
+ "ed"
+ ],
+ [
+ "po",
+ "sed"
+ ],
+ [
+ "pose",
+ "d"
+ ],
+ [
+ "p",
+ "osed"
+ ],
+ [
+ "ff",
+ "ect"
+ ],
+ [
+ "f",
+ "fect"
+ ],
+ [
+ "ж",
+ "а"
+ ],
+ [
+ "it",
+ "ude"
+ ],
+ [
+ "itu",
+ "de"
+ ],
+ [
+ "itud",
+ "e"
+ ],
+ [
+ "em",
+ "ents"
+ ],
+ [
+ "ement",
+ "s"
+ ],
+ [
+ "emen",
+ "ts"
+ ],
+ [
+ "e",
+ "ments"
+ ],
+ [
+ "as",
+ "c"
+ ],
+ [
+ "a",
+ "sc"
+ ],
+ [
+ "ad",
+ "or"
+ ],
+ [
+ "ado",
+ "r"
+ ],
+ [
+ "le",
+ "ctions"
+ ],
+ [
+ "lect",
+ "ions"
+ ],
+ [
+ "lection",
+ "s"
+ ],
+ [
+ "▁cl",
+ "ub"
+ ],
+ [
+ "▁",
+ "club"
+ ],
+ [
+ "]",
+ "{"
+ ],
+ [
+ "▁*",
+ ")"
+ ],
+ [
+ "▁",
+ "*)"
+ ],
+ [
+ "ст",
+ "во"
+ ],
+ [
+ "ств",
+ "о"
+ ],
+ [
+ "с",
+ "тво"
+ ],
+ [
+ "▁im",
+ "m"
+ ],
+ [
+ "▁i",
+ "mm"
+ ],
+ [
+ "▁",
+ "imm"
+ ],
+ [
+ "▁for",
+ "mer"
+ ],
+ [
+ "▁form",
+ "er"
+ ],
+ [
+ "▁forme",
+ "r"
+ ],
+ [
+ "▁",
+ "former"
+ ],
+ [
+ "▁r",
+ "ights"
+ ],
+ [
+ "▁right",
+ "s"
+ ],
+ [
+ "▁dec",
+ "ided"
+ ],
+ [
+ "▁decide",
+ "d"
+ ],
+ [
+ "▁decid",
+ "ed"
+ ],
+ [
+ "▁re",
+ "v"
+ ],
+ [
+ "▁r",
+ "ev"
+ ],
+ [
+ "▁",
+ "rev"
+ ],
+ [
+ "▁m",
+ "ent"
+ ],
+ [
+ "▁me",
+ "nt"
+ ],
+ [
+ "▁men",
+ "t"
+ ],
+ [
+ "▁",
+ "ment"
+ ],
+ [
+ "an",
+ "i"
+ ],
+ [
+ "a",
+ "ni"
+ ],
+ [
+ "▁st",
+ "ru"
+ ],
+ [
+ "▁str",
+ "u"
+ ],
+ [
+ "▁",
+ "stru"
+ ],
+ [
+ "▁att",
+ "ention"
+ ],
+ [
+ "art",
+ "ment"
+ ],
+ [
+ "▁I",
+ "tal"
+ ],
+ [
+ "▁It",
+ "al"
+ ],
+ [
+ "al",
+ "le"
+ ],
+ [
+ "all",
+ "e"
+ ],
+ [
+ "a",
+ "lle"
+ ],
+ [
+ "▁b",
+ "is"
+ ],
+ [
+ "▁bi",
+ "s"
+ ],
+ [
+ "▁",
+ "bis"
+ ],
+ [
+ "ge",
+ "ner"
+ ],
+ [
+ "gen",
+ "er"
+ ],
+ [
+ "g",
+ "ener"
+ ],
+ [
+ "▁in",
+ "tegr"
+ ],
+ [
+ "▁int",
+ "egr"
+ ],
+ [
+ "▁inte",
+ "gr"
+ ],
+ [
+ "▁",
+ "integr"
+ ],
+ [
+ "el",
+ "lo"
+ ],
+ [
+ "ell",
+ "o"
+ ],
+ [
+ "ry",
+ "pt"
+ ],
+ [
+ "▁a",
+ "chie"
+ ],
+ [
+ "ne",
+ "s"
+ ],
+ [
+ "n",
+ "es"
+ ],
+ [
+ "▁s",
+ "tra"
+ ],
+ [
+ "▁st",
+ "ra"
+ ],
+ [
+ "▁str",
+ "a"
+ ],
+ [
+ "▁",
+ "stra"
+ ],
+ [
+ "s",
+ "b"
+ ],
+ [
+ "▁t",
+ "ypes"
+ ],
+ [
+ "▁type",
+ "s"
+ ],
+ [
+ "▁typ",
+ "es"
+ ],
+ [
+ "▁ty",
+ "pes"
+ ],
+ [
+ "▁",
+ "types"
+ ],
+ [
+ "▁R",
+ "E"
+ ],
+ [
+ "▁",
+ "RE"
+ ],
+ [
+ "In",
+ "it"
+ ],
+ [
+ "I",
+ "nit"
+ ],
+ [
+ "▁com",
+ "ment"
+ ],
+ [
+ "▁comm",
+ "ent"
+ ],
+ [
+ "▁comme",
+ "nt"
+ ],
+ [
+ "▁",
+ "comment"
+ ],
+ [
+ "▁add",
+ "ition"
+ ],
+ [
+ "▁I",
+ "D"
+ ],
+ [
+ "▁",
+ "ID"
+ ],
+ [
+ "AR",
+ "T"
+ ],
+ [
+ "A",
+ "RT"
+ ],
+ [
+ "F",
+ "O"
+ ],
+ [
+ "щ",
+ "и"
+ ],
+ [
+ "Con",
+ "ne"
+ ],
+ [
+ "Conn",
+ "e"
+ ],
+ [
+ "C",
+ "onne"
+ ],
+ [
+ "▁s",
+ "qu"
+ ],
+ [
+ "▁sq",
+ "u"
+ ],
+ [
+ "▁consider",
+ "ed"
+ ],
+ [
+ "▁consid",
+ "ered"
+ ],
+ [
+ "id",
+ "ad"
+ ],
+ [
+ "ida",
+ "d"
+ ],
+ [
+ "▁Oct",
+ "ober"
+ ],
+ [
+ "ci",
+ "al"
+ ],
+ [
+ "cia",
+ "l"
+ ],
+ [
+ "c",
+ "ial"
+ ],
+ [
+ "▁O",
+ "f"
+ ],
+ [
+ "▁",
+ "Of"
+ ],
+ [
+ "▁tr",
+ "avel"
+ ],
+ [
+ "▁tra",
+ "vel"
+ ],
+ [
+ "▁trav",
+ "el"
+ ],
+ [
+ "▁b",
+ "oy"
+ ],
+ [
+ "▁bo",
+ "y"
+ ],
+ [
+ "▁",
+ "boy"
+ ],
+ [
+ "')",
+ "."
+ ],
+ [
+ "'",
+ ")."
+ ],
+ [
+ "u",
+ "y"
+ ],
+ [
+ "il",
+ "la"
+ ],
+ [
+ "ill",
+ "a"
+ ],
+ [
+ "i",
+ "lla"
+ ],
+ [
+ "is",
+ "try"
+ ],
+ [
+ "ist",
+ "ry"
+ ],
+ [
+ "istr",
+ "y"
+ ],
+ [
+ "▁v",
+ "a"
+ ],
+ [
+ "▁",
+ "va"
+ ],
+ [
+ "▁C",
+ "he"
+ ],
+ [
+ "▁Ch",
+ "e"
+ ],
+ [
+ "▁",
+ "Che"
+ ],
+ [
+ "ER",
+ "T"
+ ],
+ [
+ "E",
+ "RT"
+ ],
+ [
+ "en",
+ "de"
+ ],
+ [
+ "end",
+ "e"
+ ],
+ [
+ "e",
+ "nde"
+ ],
+ [
+ "un",
+ "gen"
+ ],
+ [
+ "ung",
+ "en"
+ ],
+ [
+ "unge",
+ "n"
+ ],
+ [
+ "ab",
+ "y"
+ ],
+ [
+ "a",
+ "by"
+ ],
+ [
+ "▁R",
+ "ober"
+ ],
+ [
+ "▁Ro",
+ "ber"
+ ],
+ [
+ "▁Rob",
+ "er"
+ ],
+ [
+ "▁play",
+ "ing"
+ ],
+ [
+ "il",
+ "s"
+ ],
+ [
+ "i",
+ "ls"
+ ],
+ [
+ "▁s",
+ "am"
+ ],
+ [
+ "▁sa",
+ "m"
+ ],
+ [
+ "▁",
+ "sam"
+ ],
+ [
+ "▁ex",
+ "ecut"
+ ],
+ [
+ "▁exec",
+ "ut"
+ ],
+ [
+ "▁",
+ "execut"
+ ],
+ [
+ "▁U",
+ "s"
+ ],
+ [
+ "▁",
+ "Us"
+ ],
+ [
+ "▁m",
+ "ut"
+ ],
+ [
+ "▁mu",
+ "t"
+ ],
+ [
+ "▁",
+ "mut"
+ ],
+ [
+ "▁b",
+ "al"
+ ],
+ [
+ "▁ba",
+ "l"
+ ],
+ [
+ "▁",
+ "bal"
+ ],
+ [
+ "as",
+ "se"
+ ],
+ [
+ "ass",
+ "e"
+ ],
+ [
+ "▁k",
+ "ids"
+ ],
+ [
+ "▁kid",
+ "s"
+ ],
+ [
+ "▁ki",
+ "ds"
+ ],
+ [
+ "▁fin",
+ "anc"
+ ],
+ [
+ "go",
+ "r"
+ ],
+ [
+ "g",
+ "or"
+ ],
+ [
+ "▁S",
+ "ec"
+ ],
+ [
+ "▁Se",
+ "c"
+ ],
+ [
+ "▁",
+ "Sec"
+ ],
+ [
+ "ber",
+ "t"
+ ],
+ [
+ "be",
+ "rt"
+ ],
+ [
+ "b",
+ "ert"
+ ],
+ [
+ "▁H",
+ "igh"
+ ],
+ [
+ "▁Hig",
+ "h"
+ ],
+ [
+ "▁Hi",
+ "gh"
+ ],
+ [
+ "▁",
+ "High"
+ ],
+ [
+ "▁",
+ "је"
+ ],
+ [
+ "▁ke",
+ "pt"
+ ],
+ [
+ "but",
+ "ton"
+ ],
+ [
+ "b",
+ "utton"
+ ],
+ [
+ "it",
+ "ory"
+ ],
+ [
+ "itor",
+ "y"
+ ],
+ [
+ "ito",
+ "ry"
+ ],
+ [
+ "▁R",
+ "em"
+ ],
+ [
+ "▁Re",
+ "m"
+ ],
+ [
+ "▁",
+ "Rem"
+ ],
+ [
+ "▁D",
+ "E"
+ ],
+ [
+ "▁",
+ "DE"
+ ],
+ [
+ "▁re",
+ "ach"
+ ],
+ [
+ "▁r",
+ "each"
+ ],
+ [
+ "▁",
+ "reach"
+ ],
+ [
+ "▁b",
+ "ur"
+ ],
+ [
+ "▁bu",
+ "r"
+ ],
+ [
+ "▁",
+ "bur"
+ ],
+ [
+ "La",
+ "bel"
+ ],
+ [
+ "L",
+ "abel"
+ ],
+ [
+ "á",
+ "t"
+ ],
+ [
+ "ag",
+ "o"
+ ],
+ [
+ "a",
+ "go"
+ ],
+ [
+ "▁pass",
+ "ed"
+ ],
+ [
+ "▁pas",
+ "sed"
+ ],
+ [
+ "▁be",
+ "hav"
+ ],
+ [
+ "▁beh",
+ "av"
+ ],
+ [
+ "xF",
+ "F"
+ ],
+ [
+ "x",
+ "FF"
+ ],
+ [
+ "▁R",
+ "eturn"
+ ],
+ [
+ "▁Re",
+ "turn"
+ ],
+ [
+ "▁Ret",
+ "urn"
+ ],
+ [
+ "▁",
+ "Return"
+ ],
+ [
+ "ST",
+ "R"
+ ],
+ [
+ "S",
+ "TR"
+ ],
+ [
+ "▁L",
+ "es"
+ ],
+ [
+ "▁Le",
+ "s"
+ ],
+ [
+ "▁",
+ "Les"
+ ],
+ [
+ "▁o",
+ "rd"
+ ],
+ [
+ "▁or",
+ "d"
+ ],
+ [
+ "▁",
+ "ord"
+ ],
+ [
+ "al",
+ "a"
+ ],
+ [
+ "a",
+ "la"
+ ],
+ [
+ "in",
+ "ger"
+ ],
+ [
+ "ing",
+ "er"
+ ],
+ [
+ "inge",
+ "r"
+ ],
+ [
+ "▁S",
+ "ince"
+ ],
+ [
+ "▁Sin",
+ "ce"
+ ],
+ [
+ "▁",
+ "Since"
+ ],
+ [
+ "▁exper",
+ "i"
+ ],
+ [
+ "▁exp",
+ "eri"
+ ],
+ [
+ "▁s",
+ "hall"
+ ],
+ [
+ "▁sh",
+ "all"
+ ],
+ [
+ "▁sha",
+ "ll"
+ ],
+ [
+ "▁",
+ "shall"
+ ],
+ [
+ "▁s",
+ "tar"
+ ],
+ [
+ "▁st",
+ "ar"
+ ],
+ [
+ "▁sta",
+ "r"
+ ],
+ [
+ "▁",
+ "star"
+ ],
+ [
+ "no",
+ "n"
+ ],
+ [
+ "n",
+ "on"
+ ],
+ [
+ "▁g",
+ "un"
+ ],
+ [
+ "▁gu",
+ "n"
+ ],
+ [
+ "▁",
+ "gun"
+ ],
+ [
+ "▁B",
+ "el"
+ ],
+ [
+ "▁Be",
+ "l"
+ ],
+ [
+ "▁",
+ "Bel"
+ ],
+ [
+ "▁ob",
+ "j"
+ ],
+ [
+ "▁",
+ "obj"
+ ],
+ [
+ "ar",
+ "es"
+ ],
+ [
+ "are",
+ "s"
+ ],
+ [
+ "a",
+ "res"
+ ],
+ [
+ "r",
+ "s"
+ ],
+ [
+ "▁we",
+ "eks"
+ ],
+ [
+ "▁week",
+ "s"
+ ],
+ [
+ "ne",
+ "n"
+ ],
+ [
+ "n",
+ "en"
+ ],
+ [
+ "▁S",
+ "tre"
+ ],
+ [
+ "▁St",
+ "re"
+ ],
+ [
+ "▁Str",
+ "e"
+ ],
+ [
+ "or",
+ "ing"
+ ],
+ [
+ "ori",
+ "ng"
+ ],
+ [
+ "o",
+ "ring"
+ ],
+ [
+ "▁",
+ "î"
+ ],
+ [
+ "▁ser",
+ "ious"
+ ],
+ [
+ "time",
+ "s"
+ ],
+ [
+ "ti",
+ "mes"
+ ],
+ [
+ "tim",
+ "es"
+ ],
+ [
+ "t",
+ "imes"
+ ],
+ [
+ "▁H",
+ "ouse"
+ ],
+ [
+ "▁Ho",
+ "use"
+ ],
+ [
+ "▁Hou",
+ "se"
+ ],
+ [
+ "▁r",
+ "oll"
+ ],
+ [
+ "▁ro",
+ "ll"
+ ],
+ [
+ "▁",
+ "roll"
+ ],
+ [
+ "▁reg",
+ "ister"
+ ],
+ [
+ "▁",
+ "register"
+ ],
+ [
+ "▁mod",
+ "ule"
+ ],
+ [
+ "▁mo",
+ "dule"
+ ],
+ [
+ "▁",
+ "module"
+ ],
+ [
+ "▁app",
+ "lic"
+ ],
+ [
+ "▁ap",
+ "plic"
+ ],
+ [
+ "▁appl",
+ "ic"
+ ],
+ [
+ "I",
+ "R"
+ ],
+ [
+ "▁c",
+ "ook"
+ ],
+ [
+ "▁co",
+ "ok"
+ ],
+ [
+ "▁",
+ "cook"
+ ],
+ [
+ "au",
+ "x"
+ ],
+ [
+ "a",
+ "ux"
+ ],
+ [
+ "▁s",
+ "ave"
+ ],
+ [
+ "▁sa",
+ "ve"
+ ],
+ [
+ "▁sav",
+ "e"
+ ],
+ [
+ "▁",
+ "save"
+ ],
+ [
+ "▁C",
+ "r"
+ ],
+ [
+ "▁",
+ "Cr"
+ ],
+ [
+ ",",
+ "\r"
+ ],
+ [
+ "▁st",
+ "ates"
+ ],
+ [
+ "▁stat",
+ "es"
+ ],
+ [
+ "▁state",
+ "s"
+ ],
+ [
+ "▁sta",
+ "tes"
+ ],
+ [
+ "▁",
+ "states"
+ ],
+ [
+ "▁em",
+ "pty"
+ ],
+ [
+ "▁emp",
+ "ty"
+ ],
+ [
+ "▁empt",
+ "y"
+ ],
+ [
+ "▁",
+ "empty"
+ ],
+ [
+ "▁aut",
+ "om"
+ ],
+ [
+ "▁au",
+ "tom"
+ ],
+ [
+ "▁auto",
+ "m"
+ ],
+ [
+ "▁",
+ "autom"
+ ],
+ [
+ "fig",
+ "ure"
+ ],
+ [
+ "ian",
+ "ce"
+ ],
+ [
+ "i",
+ "ance"
+ ],
+ [
+ "▁h",
+ "appy"
+ ],
+ [
+ "▁happ",
+ "y"
+ ],
+ [
+ "▁f",
+ "n"
+ ],
+ [
+ "▁",
+ "fn"
+ ],
+ [
+ "▁j",
+ "ud"
+ ],
+ [
+ "▁ju",
+ "d"
+ ],
+ [
+ "▁",
+ "jud"
+ ],
+ [
+ "▁h",
+ "at"
+ ],
+ [
+ "▁ha",
+ "t"
+ ],
+ [
+ "▁",
+ "hat"
+ ],
+ [
+ "AC",
+ "K"
+ ],
+ [
+ "A",
+ "CK"
+ ],
+ [
+ "▁F",
+ "e"
+ ],
+ [
+ "▁",
+ "Fe"
+ ],
+ [
+ "$",
+ "-"
+ ],
+ [
+ "iv",
+ "il"
+ ],
+ [
+ "ivi",
+ "l"
+ ],
+ [
+ "i",
+ "vil"
+ ],
+ [
+ "ot",
+ "ed"
+ ],
+ [
+ "ote",
+ "d"
+ ],
+ [
+ "o",
+ "ted"
+ ],
+ [
+ "▁size",
+ "of"
+ ],
+ [
+ "▁",
+ "sizeof"
+ ],
+ [
+ "▁sit",
+ "uation"
+ ],
+ [
+ "▁situ",
+ "ation"
+ ],
+ [
+ "▁l",
+ "ives"
+ ],
+ [
+ "▁li",
+ "ves"
+ ],
+ [
+ "▁live",
+ "s"
+ ],
+ [
+ "▁liv",
+ "es"
+ ],
+ [
+ "▁fe",
+ "eling"
+ ],
+ [
+ "▁feel",
+ "ing"
+ ],
+ [
+ "▁fee",
+ "ling"
+ ],
+ [
+ "▁r",
+ "isk"
+ ],
+ [
+ "▁ri",
+ "sk"
+ ],
+ [
+ "▁ris",
+ "k"
+ ],
+ [
+ "▁Jan",
+ "uary"
+ ],
+ [
+ "▁Januar",
+ "y"
+ ],
+ [
+ "▁Ob",
+ "ject"
+ ],
+ [
+ "▁",
+ "Object"
+ ],
+ [
+ "▁re",
+ "comm"
+ ],
+ [
+ "▁rec",
+ "omm"
+ ],
+ [
+ "▁в",
+ "ы"
+ ],
+ [
+ "▁",
+ "вы"
+ ],
+ [
+ "▁pot",
+ "ential"
+ ],
+ [
+ "ea",
+ "h"
+ ],
+ [
+ "e",
+ "ah"
+ ],
+ [
+ "▁com",
+ "plex"
+ ],
+ [
+ "▁comp",
+ "lex"
+ ],
+ [
+ "▁compl",
+ "ex"
+ ],
+ [
+ "▁",
+ "complex"
+ ],
+ [
+ "print",
+ "f"
+ ],
+ [
+ "ist",
+ "ance"
+ ],
+ [
+ "istan",
+ "ce"
+ ],
+ [
+ "i",
+ "stance"
+ ],
+ [
+ "ir",
+ "th"
+ ],
+ [
+ "irt",
+ "h"
+ ],
+ [
+ "li",
+ "k"
+ ],
+ [
+ "l",
+ "ik"
+ ],
+ [
+ "as",
+ "te"
+ ],
+ [
+ "ast",
+ "e"
+ ],
+ [
+ "a",
+ "ste"
+ ],
+ [
+ "▁wh",
+ "ose"
+ ],
+ [
+ "▁who",
+ "se"
+ ],
+ [
+ "Ar",
+ "g"
+ ],
+ [
+ "A",
+ "rg"
+ ],
+ [
+ "▁mod",
+ "ern"
+ ],
+ [
+ "▁mo",
+ "dern"
+ ],
+ [
+ "▁mode",
+ "rn"
+ ],
+ [
+ "▁moder",
+ "n"
+ ],
+ [
+ "ion",
+ "es"
+ ],
+ [
+ "io",
+ "nes"
+ ],
+ [
+ "ione",
+ "s"
+ ],
+ [
+ "i",
+ "ones"
+ ],
+ [
+ "▁ч",
+ "е"
+ ],
+ [
+ "▁",
+ "че"
+ ],
+ [
+ "▁s",
+ "ett"
+ ],
+ [
+ "▁se",
+ "tt"
+ ],
+ [
+ "▁set",
+ "t"
+ ],
+ [
+ "▁M",
+ "ag"
+ ],
+ [
+ "▁Ma",
+ "g"
+ ],
+ [
+ "▁",
+ "Mag"
+ ],
+ [
+ "a",
+ "e"
+ ],
+ [
+ "▁cond",
+ "ition"
+ ],
+ [
+ "▁",
+ "condition"
+ ],
+ [
+ "Le",
+ "ngth"
+ ],
+ [
+ "L",
+ "ength"
+ ],
+ [
+ "▁f",
+ "it"
+ ],
+ [
+ "▁fi",
+ "t"
+ ],
+ [
+ "▁",
+ "fit"
+ ],
+ [
+ "ound",
+ "s"
+ ],
+ [
+ "oun",
+ "ds"
+ ],
+ [
+ "▁ch",
+ "anged"
+ ],
+ [
+ "▁chang",
+ "ed"
+ ],
+ [
+ "▁change",
+ "d"
+ ],
+ [
+ "▁",
+ "changed"
+ ],
+ [
+ "▁g",
+ "uy"
+ ],
+ [
+ "▁gu",
+ "y"
+ ],
+ [
+ "fil",
+ "ter"
+ ],
+ [
+ "at",
+ "ever"
+ ],
+ [
+ "ate",
+ "ver"
+ ],
+ [
+ "é",
+ "d"
+ ],
+ [
+ "re",
+ "move"
+ ],
+ [
+ "rem",
+ "ove"
+ ],
+ [
+ "▁h",
+ "op"
+ ],
+ [
+ "▁ho",
+ "p"
+ ],
+ [
+ "▁",
+ "hop"
+ ],
+ [
+ "▁O",
+ "ut"
+ ],
+ [
+ "▁",
+ "Out"
+ ],
+ [
+ "▁R",
+ "ich"
+ ],
+ [
+ "▁Ric",
+ "h"
+ ],
+ [
+ "▁",
+ "Rich"
+ ],
+ [
+ "ch",
+ "ild"
+ ],
+ [
+ "chi",
+ "ld"
+ ],
+ [
+ "▁in",
+ "cluded"
+ ],
+ [
+ "▁incl",
+ "uded"
+ ],
+ [
+ "▁includ",
+ "ed"
+ ],
+ [
+ "▁include",
+ "d"
+ ],
+ [
+ "▁inclu",
+ "ded"
+ ],
+ [
+ "$",
+ "\\"
+ ],
+ [
+ "▁T",
+ "om"
+ ],
+ [
+ "▁To",
+ "m"
+ ],
+ [
+ "▁",
+ "Tom"
+ ],
+ [
+ "el",
+ "ine"
+ ],
+ [
+ "eli",
+ "ne"
+ ],
+ [
+ "elin",
+ "e"
+ ],
+ [
+ "e",
+ "line"
+ ],
+ [
+ "▁s",
+ "ometimes"
+ ],
+ [
+ "▁some",
+ "times"
+ ],
+ [
+ "▁somet",
+ "imes"
+ ],
+ [
+ "▁sometime",
+ "s"
+ ],
+ [
+ "▁dr",
+ "ink"
+ ],
+ [
+ "▁qu",
+ "ant"
+ ],
+ [
+ "▁",
+ "quant"
+ ],
+ [
+ "▁p",
+ "lease"
+ ],
+ [
+ "▁ple",
+ "ase"
+ ],
+ [
+ "▁I",
+ "nt"
+ ],
+ [
+ "▁In",
+ "t"
+ ],
+ [
+ "▁",
+ "Int"
+ ],
+ [
+ "ri",
+ "ef"
+ ],
+ [
+ "rie",
+ "f"
+ ],
+ [
+ "r",
+ "ief"
+ ],
+ [
+ "▁ex",
+ "actly"
+ ],
+ [
+ "▁exact",
+ "ly"
+ ],
+ [
+ "ci",
+ "ng"
+ ],
+ [
+ "cin",
+ "g"
+ ],
+ [
+ "c",
+ "ing"
+ ],
+ [
+ "▁all",
+ "owed"
+ ],
+ [
+ "▁allow",
+ "ed"
+ ],
+ [
+ "▁",
+ "allowed"
+ ],
+ [
+ "bu",
+ "ild"
+ ],
+ [
+ "b",
+ "uild"
+ ],
+ [
+ "▁beaut",
+ "iful"
+ ],
+ [
+ "▁W",
+ "ell"
+ ],
+ [
+ "▁We",
+ "ll"
+ ],
+ [
+ "▁Wel",
+ "l"
+ ],
+ [
+ "▁",
+ "Well"
+ ],
+ [
+ "▁look",
+ "s"
+ ],
+ [
+ "▁lo",
+ "oks"
+ ],
+ [
+ "▁",
+ "ü"
+ ],
+ [
+ "▁ch",
+ "ance"
+ ],
+ [
+ "▁w",
+ "rote"
+ ],
+ [
+ "▁wr",
+ "ote"
+ ],
+ [
+ "▁n",
+ "or"
+ ],
+ [
+ "▁no",
+ "r"
+ ],
+ [
+ "▁",
+ "nor"
+ ],
+ [
+ "▁f",
+ "ailed"
+ ],
+ [
+ "▁fa",
+ "iled"
+ ],
+ [
+ "▁fail",
+ "ed"
+ ],
+ [
+ "▁",
+ "failed"
+ ],
+ [
+ "Me",
+ "t"
+ ],
+ [
+ "M",
+ "et"
+ ],
+ [
+ "▁p",
+ "rior"
+ ],
+ [
+ "▁pr",
+ "ior"
+ ],
+ [
+ "▁pri",
+ "or"
+ ],
+ [
+ "▁h",
+ "undred"
+ ],
+ [
+ "ско",
+ "й"
+ ],
+ [
+ "с",
+ "кой"
+ ],
+ [
+ "or",
+ "ia"
+ ],
+ [
+ "ori",
+ "a"
+ ],
+ [
+ "o",
+ "ria"
+ ],
+ [
+ "▁c",
+ "y"
+ ],
+ [
+ "▁",
+ "cy"
+ ],
+ [
+ "▁w",
+ "eb"
+ ],
+ [
+ "▁we",
+ "b"
+ ],
+ [
+ "▁",
+ "web"
+ ],
+ [
+ "▁m",
+ "ess"
+ ],
+ [
+ "▁me",
+ "ss"
+ ],
+ [
+ "▁mes",
+ "s"
+ ],
+ [
+ "le",
+ "q"
+ ],
+ [
+ "l",
+ "eq"
+ ],
+ [
+ "d",
+ "y"
+ ],
+ [
+ "te",
+ "x"
+ ],
+ [
+ "t",
+ "ex"
+ ],
+ [
+ "▁a",
+ "nim"
+ ],
+ [
+ "▁an",
+ "im"
+ ],
+ [
+ "▁",
+ "anim"
+ ],
+ [
+ "at",
+ "ur"
+ ],
+ [
+ "atu",
+ "r"
+ ],
+ [
+ "▁str",
+ "ucture"
+ ],
+ [
+ "▁struct",
+ "ure"
+ ],
+ [
+ "▁",
+ "structure"
+ ],
+ [
+ "opt",
+ "ion"
+ ],
+ [
+ "o",
+ "ption"
+ ],
+ [
+ "▁act",
+ "ual"
+ ],
+ [
+ "▁",
+ "actual"
+ ],
+ [
+ "▁Fr",
+ "anc"
+ ],
+ [
+ "▁Fra",
+ "nc"
+ ],
+ [
+ "▁Fran",
+ "c"
+ ],
+ [
+ "en",
+ "ced"
+ ],
+ [
+ "ence",
+ "d"
+ ],
+ [
+ "enc",
+ "ed"
+ ],
+ [
+ ".<",
+ "/"
+ ],
+ [
+ ".",
+ ""
+ ],
+ [
+ "▁f",
+ "low"
+ ],
+ [
+ "▁fl",
+ "ow"
+ ],
+ [
+ "▁flo",
+ "w"
+ ],
+ [
+ "▁",
+ "flow"
+ ],
+ [
+ "▁A",
+ "fr"
+ ],
+ [
+ "▁Af",
+ "r"
+ ],
+ [
+ "de",
+ "t"
+ ],
+ [
+ "d",
+ "et"
+ ],
+ [
+ "▁K",
+ "e"
+ ],
+ [
+ "▁",
+ "Ke"
+ ],
+ [
+ "et",
+ "y"
+ ],
+ [
+ "e",
+ "ty"
+ ],
+ [
+ "ски",
+ "й"
+ ],
+ [
+ "с",
+ "кий"
+ ],
+ [
+ "▁st",
+ "uff"
+ ],
+ [
+ "it",
+ "ter"
+ ],
+ [
+ "itt",
+ "er"
+ ],
+ [
+ "itte",
+ "r"
+ ],
+ [
+ "▁ar",
+ "gs"
+ ],
+ [
+ "▁arg",
+ "s"
+ ],
+ [
+ "▁",
+ "args"
+ ],
+ [
+ "▁al",
+ "bum"
+ ],
+ [
+ "▁",
+ "album"
+ ],
+ [
+ "▁",
+ "]"
+ ],
+ [
+ "ug",
+ "in"
+ ],
+ [
+ "u",
+ "gin"
+ ],
+ [
+ "S",
+ "U"
+ ],
+ [
+ "Pe",
+ "r"
+ ],
+ [
+ "P",
+ "er"
+ ],
+ [
+ "▁cir",
+ "c"
+ ],
+ [
+ "▁ci",
+ "rc"
+ ],
+ [
+ "▁",
+ "circ"
+ ],
+ [
+ "▁cor",
+ "rect"
+ ],
+ [
+ "▁corre",
+ "ct"
+ ],
+ [
+ "▁",
+ "correct"
+ ],
+ [
+ "▁l",
+ "ines"
+ ],
+ [
+ "▁li",
+ "nes"
+ ],
+ [
+ "▁line",
+ "s"
+ ],
+ [
+ "▁lin",
+ "es"
+ ],
+ [
+ "▁",
+ "lines"
+ ],
+ [
+ "▁complet",
+ "ely"
+ ],
+ [
+ "▁complete",
+ "ly"
+ ],
+ [
+ "kn",
+ "own"
+ ],
+ [
+ "know",
+ "n"
+ ],
+ [
+ "k",
+ "nown"
+ ],
+ [
+ "▁t",
+ "ree"
+ ],
+ [
+ "▁tr",
+ "ee"
+ ],
+ [
+ "▁tre",
+ "e"
+ ],
+ [
+ "▁",
+ "tree"
+ ],
+ [
+ "ro",
+ "ot"
+ ],
+ [
+ "r",
+ "oot"
+ ],
+ [
+ "▁J",
+ "apan"
+ ],
+ [
+ "▁Ja",
+ "pan"
+ ],
+ [
+ "▁Jap",
+ "an"
+ ],
+ [
+ "ol",
+ "es"
+ ],
+ [
+ "ole",
+ "s"
+ ],
+ [
+ "o",
+ "les"
+ ],
+ [
+ "en",
+ "do"
+ ],
+ [
+ "end",
+ "o"
+ ],
+ [
+ "▁l",
+ "ocation"
+ ],
+ [
+ "▁loc",
+ "ation"
+ ],
+ [
+ "▁",
+ "location"
+ ],
+ [
+ "▁",
+ "Х"
+ ],
+ [
+ "▁m",
+ "id"
+ ],
+ [
+ "▁mi",
+ "d"
+ ],
+ [
+ "▁",
+ "mid"
+ ],
+ [
+ "al",
+ "ing"
+ ],
+ [
+ "ali",
+ "ng"
+ ],
+ [
+ "alin",
+ "g"
+ ],
+ [
+ "a",
+ "ling"
+ ],
+ [
+ "G",
+ "L"
+ ],
+ [
+ "ia",
+ "no"
+ ],
+ [
+ "ian",
+ "o"
+ ],
+ [
+ "i",
+ "ano"
+ ],
+ [
+ "▁{",
+ "}"
+ ],
+ [
+ "▁",
+ "{}"
+ ],
+ [
+ "la",
+ "ng"
+ ],
+ [
+ "lan",
+ "g"
+ ],
+ [
+ "l",
+ "ang"
+ ],
+ [
+ "▁equ",
+ "ip"
+ ],
+ [
+ "ERR",
+ "OR"
+ ],
+ [
+ "▁mem",
+ "ory"
+ ],
+ [
+ "▁memor",
+ "y"
+ ],
+ [
+ "▁memo",
+ "ry"
+ ],
+ [
+ "▁",
+ "memory"
+ ],
+ [
+ "▁(",
+ "\""
+ ],
+ [
+ "▁",
+ "(\""
+ ],
+ [
+ "▁n",
+ "ature"
+ ],
+ [
+ "▁nat",
+ "ure"
+ ],
+ [
+ "▁natur",
+ "e"
+ ],
+ [
+ "go",
+ "ogle"
+ ],
+ [
+ "ab",
+ "s"
+ ],
+ [
+ "a",
+ "bs"
+ ],
+ [
+ "B",
+ "C"
+ ],
+ [
+ "▁g",
+ "ets"
+ ],
+ [
+ "▁get",
+ "s"
+ ],
+ [
+ "▁ge",
+ "ts"
+ ],
+ [
+ "▁",
+ "gets"
+ ],
+ [
+ "Com",
+ "mand"
+ ],
+ [
+ "Comm",
+ "and"
+ ],
+ [
+ "TE",
+ "R"
+ ],
+ [
+ "T",
+ "ER"
+ ],
+ [
+ "al",
+ "ed"
+ ],
+ [
+ "ale",
+ "d"
+ ],
+ [
+ "a",
+ "led"
+ ],
+ [
+ "c",
+ "p"
+ ],
+ [
+ "▁p",
+ "urch"
+ ],
+ [
+ "▁pur",
+ "ch"
+ ],
+ [
+ "▁D",
+ "en"
+ ],
+ [
+ "▁De",
+ "n"
+ ],
+ [
+ "▁",
+ "Den"
+ ],
+ [
+ "▁her",
+ "self"
+ ],
+ [
+ "▁hers",
+ "elf"
+ ],
+ [
+ "▁I",
+ "r"
+ ],
+ [
+ "▁",
+ "Ir"
+ ],
+ [
+ "▁s",
+ "ie"
+ ],
+ [
+ "▁si",
+ "e"
+ ],
+ [
+ "ga",
+ "r"
+ ],
+ [
+ "g",
+ "ar"
+ ],
+ [
+ "A",
+ "p"
+ ],
+ [
+ "▁n",
+ "el"
+ ],
+ [
+ "▁ne",
+ "l"
+ ],
+ [
+ "▁",
+ "nel"
+ ],
+ [
+ "ot",
+ "a"
+ ],
+ [
+ "o",
+ "ta"
+ ],
+ [
+ ")",
+ "]"
+ ],
+ [
+ "co",
+ "r"
+ ],
+ [
+ "c",
+ "or"
+ ],
+ [
+ "ac",
+ "ht"
+ ],
+ [
+ "ach",
+ "t"
+ ],
+ [
+ "a",
+ "cht"
+ ],
+ [
+ "(",
+ "*"
+ ],
+ [
+ "irt",
+ "ual"
+ ],
+ [
+ "▁pol",
+ "ice"
+ ],
+ [
+ "▁polic",
+ "e"
+ ],
+ [
+ "▁s",
+ "kin"
+ ],
+ [
+ "▁sk",
+ "in"
+ ],
+ [
+ "▁ski",
+ "n"
+ ],
+ [
+ "▁",
+ "skin"
+ ],
+ [
+ "sh",
+ "ip"
+ ],
+ [
+ "s",
+ "hip"
+ ],
+ [
+ "ef",
+ "ined"
+ ],
+ [
+ "augh",
+ "ter"
+ ],
+ [
+ "aught",
+ "er"
+ ],
+ [
+ "in",
+ "ding"
+ ],
+ [
+ "ind",
+ "ing"
+ ],
+ [
+ "indi",
+ "ng"
+ ],
+ [
+ "▁S",
+ "l"
+ ],
+ [
+ "▁",
+ "Sl"
+ ],
+ [
+ "▁in",
+ "flu"
+ ],
+ [
+ "▁infl",
+ "u"
+ ],
+ [
+ "▁inf",
+ "lu"
+ ],
+ [
+ "▁m",
+ "ount"
+ ],
+ [
+ "▁mo",
+ "unt"
+ ],
+ [
+ "▁mou",
+ "nt"
+ ],
+ [
+ "▁",
+ "mount"
+ ],
+ [
+ "▁a",
+ "z"
+ ],
+ [
+ "▁",
+ "az"
+ ],
+ [
+ "▁w",
+ "ood"
+ ],
+ [
+ "▁wo",
+ "od"
+ ],
+ [
+ "▁",
+ "wood"
+ ],
+ [
+ "ot",
+ "es"
+ ],
+ [
+ "ote",
+ "s"
+ ],
+ [
+ "o",
+ "tes"
+ ],
+ [
+ "eg",
+ "a"
+ ],
+ [
+ "e",
+ "ga"
+ ],
+ [
+ "▁acc",
+ "ording"
+ ],
+ [
+ "▁accord",
+ "ing"
+ ],
+ [
+ "▁name",
+ "space"
+ ],
+ [
+ "▁names",
+ "pace"
+ ],
+ [
+ "▁",
+ "namespace"
+ ],
+ [
+ "Del",
+ "ta"
+ ],
+ [
+ "D",
+ "elta"
+ ],
+ [
+ "st",
+ "ant"
+ ],
+ [
+ "sta",
+ "nt"
+ ],
+ [
+ "stan",
+ "t"
+ ],
+ [
+ "▁pub",
+ "lished"
+ ],
+ [
+ "▁publish",
+ "ed"
+ ],
+ [
+ "▁",
+ "published"
+ ],
+ [
+ "ak",
+ "er"
+ ],
+ [
+ "ake",
+ "r"
+ ],
+ [
+ "a",
+ "ker"
+ ],
+ [
+ "▁Bl",
+ "ack"
+ ],
+ [
+ "▁",
+ "Black"
+ ],
+ [
+ "l",
+ "n"
+ ],
+ [
+ "▁indust",
+ "ry"
+ ],
+ [
+ "SO",
+ "N"
+ ],
+ [
+ "S",
+ "ON"
+ ],
+ [
+ "Re",
+ "p"
+ ],
+ [
+ "R",
+ "ep"
+ ],
+ [
+ "▁ch",
+ "oice"
+ ],
+ [
+ "▁cho",
+ "ice"
+ ],
+ [
+ "▁",
+ "choice"
+ ],
+ [
+ "▁in",
+ "n"
+ ],
+ [
+ "▁i",
+ "nn"
+ ],
+ [
+ "▁",
+ "inn"
+ ],
+ [
+ "k",
+ "l"
+ ],
+ [
+ "▁p",
+ "al"
+ ],
+ [
+ "▁pa",
+ "l"
+ ],
+ [
+ "▁",
+ "pal"
+ ],
+ [
+ "▁a",
+ "ud"
+ ],
+ [
+ "▁au",
+ "d"
+ ],
+ [
+ "▁",
+ "aud"
+ ],
+ [
+ "▁stand",
+ "ard"
+ ],
+ [
+ "▁",
+ "standard"
+ ],
+ [
+ "▁know",
+ "ledge"
+ ],
+ [
+ "**",
+ ","
+ ],
+ [
+ "*",
+ "*,"
+ ],
+ [
+ "▁F",
+ "rank"
+ ],
+ [
+ "▁Fr",
+ "ank"
+ ],
+ [
+ "▁Fran",
+ "k"
+ ],
+ [
+ "s",
+ "q"
+ ],
+ [
+ "Out",
+ "put"
+ ],
+ [
+ "▁f",
+ "ör"
+ ],
+ [
+ "▁fö",
+ "r"
+ ],
+ [
+ "▁",
+ "för"
+ ],
+ [
+ "Val",
+ "id"
+ ],
+ [
+ "ug",
+ "h"
+ ],
+ [
+ "u",
+ "gh"
+ ],
+ [
+ "▁bo",
+ "oks"
+ ],
+ [
+ "▁book",
+ "s"
+ ],
+ [
+ "▁",
+ "books"
+ ],
+ [
+ "▁J",
+ "ames"
+ ],
+ [
+ "▁Jam",
+ "es"
+ ],
+ [
+ "▁Ja",
+ "mes"
+ ],
+ [
+ "k",
+ "o"
+ ],
+ [
+ "▁compan",
+ "ies"
+ ],
+ [
+ "an",
+ "ning"
+ ],
+ [
+ "ann",
+ "ing"
+ ],
+ [
+ "anni",
+ "ng"
+ ],
+ [
+ "▁v",
+ "ict"
+ ],
+ [
+ "▁vi",
+ "ct"
+ ],
+ [
+ "▁vic",
+ "t"
+ ],
+ [
+ "▁re",
+ "pl"
+ ],
+ [
+ "▁rep",
+ "l"
+ ],
+ [
+ "▁s",
+ "che"
+ ],
+ [
+ "▁sc",
+ "he"
+ ],
+ [
+ "▁sch",
+ "e"
+ ],
+ [
+ "▁",
+ "sche"
+ ],
+ [
+ "▁h",
+ "appen"
+ ],
+ [
+ "▁happ",
+ "en"
+ ],
+ [
+ "▁ha",
+ "ppen"
+ ],
+ [
+ "ft",
+ "y"
+ ],
+ [
+ "f",
+ "ty"
+ ],
+ [
+ "ac",
+ "ity"
+ ],
+ [
+ "aci",
+ "ty"
+ ],
+ [
+ "a",
+ "city"
+ ],
+ [
+ "ir",
+ "a"
+ ],
+ [
+ "i",
+ "ra"
+ ],
+ [
+ "▁im",
+ "plement"
+ ],
+ [
+ "▁imp",
+ "lement"
+ ],
+ [
+ "▁impl",
+ "ement"
+ ],
+ [
+ "▁",
+ "implement"
+ ],
+ [
+ "ско",
+ "го"
+ ],
+ [
+ "ск",
+ "ого"
+ ],
+ [
+ "с",
+ "кого"
+ ],
+ [
+ "num",
+ "ber"
+ ],
+ [
+ "nu",
+ "mber"
+ ],
+ [
+ "n",
+ "umber"
+ ],
+ [
+ "S",
+ "H"
+ ],
+ [
+ "ir",
+ "o"
+ ],
+ [
+ "i",
+ "ro"
+ ],
+ [
+ "▁f",
+ "ear"
+ ],
+ [
+ "▁fe",
+ "ar"
+ ],
+ [
+ "▁t",
+ "ouch"
+ ],
+ [
+ "▁to",
+ "uch"
+ ],
+ [
+ "▁tou",
+ "ch"
+ ],
+ [
+ "▁",
+ "touch"
+ ],
+ [
+ "▁c",
+ "ast"
+ ],
+ [
+ "▁cas",
+ "t"
+ ],
+ [
+ "▁ca",
+ "st"
+ ],
+ [
+ "▁",
+ "cast"
+ ],
+ [
+ "AS",
+ "S"
+ ],
+ [
+ "A",
+ "SS"
+ ],
+ [
+ "▁cons",
+ "ist"
+ ],
+ [
+ "T",
+ "ask"
+ ],
+ [
+ "▁s",
+ "ig"
+ ],
+ [
+ "▁si",
+ "g"
+ ],
+ [
+ "▁",
+ "sig"
+ ],
+ [
+ "б",
+ "а"
+ ],
+ [
+ "ig",
+ "ation"
+ ],
+ [
+ "▁M",
+ "ost"
+ ],
+ [
+ "▁Mo",
+ "st"
+ ],
+ [
+ "▁Mos",
+ "t"
+ ],
+ [
+ "▁",
+ "Most"
+ ],
+ [
+ "▁D",
+ "er"
+ ],
+ [
+ "▁De",
+ "r"
+ ],
+ [
+ "▁",
+ "Der"
+ ],
+ [
+ "}(",
+ "\\"
+ ],
+ [
+ "}",
+ "(\\"
+ ],
+ [
+ ":",
+ "\""
+ ],
+ [
+ "▁F",
+ "ig"
+ ],
+ [
+ "▁Fi",
+ "g"
+ ],
+ [
+ "▁",
+ "Fig"
+ ],
+ [
+ "al",
+ "i"
+ ],
+ [
+ "a",
+ "li"
+ ],
+ [
+ "in",
+ "er"
+ ],
+ [
+ "ine",
+ "r"
+ ],
+ [
+ "i",
+ "ner"
+ ],
+ [
+ "')",
+ ","
+ ],
+ [
+ "'",
+ "),"
+ ],
+ [
+ "▁C",
+ "oun"
+ ],
+ [
+ "▁Co",
+ "un"
+ ],
+ [
+ "▁Cou",
+ "n"
+ ],
+ [
+ "(",
+ "_"
+ ],
+ [
+ "▁d",
+ "istributed"
+ ],
+ [
+ "▁distribut",
+ "ed"
+ ],
+ [
+ "▁distribute",
+ "d"
+ ],
+ [
+ "NA",
+ "ME"
+ ],
+ [
+ "N",
+ "AME"
+ ],
+ [
+ "▁m",
+ "ur"
+ ],
+ [
+ "▁mu",
+ "r"
+ ],
+ [
+ "▁care",
+ "er"
+ ],
+ [
+ "~",
+ "~"
+ ],
+ [
+ "pe",
+ "rs"
+ ],
+ [
+ "per",
+ "s"
+ ],
+ [
+ "p",
+ "ers"
+ ],
+ [
+ "ar",
+ "ies"
+ ],
+ [
+ "ari",
+ "es"
+ ],
+ [
+ "a",
+ "ries"
+ ],
+ [
+ "en",
+ "ses"
+ ],
+ [
+ "ens",
+ "es"
+ ],
+ [
+ "ense",
+ "s"
+ ],
+ [
+ "▁Al",
+ "so"
+ ],
+ [
+ "▁Als",
+ "o"
+ ],
+ [
+ "Vers",
+ "ion"
+ ],
+ [
+ "V",
+ "ersion"
+ ],
+ [
+ "▁un",
+ "ique"
+ ],
+ [
+ "▁uniqu",
+ "e"
+ ],
+ [
+ "▁",
+ "unique"
+ ],
+ [
+ "▁Fr",
+ "ance"
+ ],
+ [
+ "▁Franc",
+ "e"
+ ],
+ [
+ "▁Fran",
+ "ce"
+ ],
+ [
+ "B",
+ "A"
+ ],
+ [
+ "k",
+ "y"
+ ],
+ [
+ "▁F",
+ "ebru"
+ ],
+ [
+ "▁Fe",
+ "bru"
+ ],
+ [
+ "▁Feb",
+ "ru"
+ ],
+ [
+ "▁d",
+ "ied"
+ ],
+ [
+ "▁di",
+ "ed"
+ ],
+ [
+ "▁die",
+ "d"
+ ],
+ [
+ "om",
+ "ega"
+ ],
+ [
+ "ome",
+ "ga"
+ ],
+ [
+ "▁F",
+ "orm"
+ ],
+ [
+ "▁For",
+ "m"
+ ],
+ [
+ "▁Fo",
+ "rm"
+ ],
+ [
+ "▁",
+ "Form"
+ ],
+ [
+ "▁w",
+ "idth"
+ ],
+ [
+ "▁wid",
+ "th"
+ ],
+ [
+ "▁",
+ "width"
+ ],
+ [
+ "to",
+ "col"
+ ],
+ [
+ "t",
+ "ocol"
+ ],
+ [
+ "▁l",
+ "ie"
+ ],
+ [
+ "▁li",
+ "e"
+ ],
+ [
+ "▁",
+ "lie"
+ ],
+ [
+ "Sh",
+ "e"
+ ],
+ [
+ "S",
+ "he"
+ ],
+ [
+ "é",
+ "m"
+ ],
+ [
+ "▁stra",
+ "ight"
+ ],
+ [
+ "▁n",
+ "ach"
+ ],
+ [
+ "▁na",
+ "ch"
+ ],
+ [
+ "▁st",
+ "ood"
+ ],
+ [
+ "▁sto",
+ "od"
+ ],
+ [
+ "▁",
+ "stood"
+ ],
+ [
+ "ol",
+ "ds"
+ ],
+ [
+ "old",
+ "s"
+ ],
+ [
+ "▁g",
+ "oes"
+ ],
+ [
+ "▁go",
+ "es"
+ ],
+ [
+ "ce",
+ "ll"
+ ],
+ [
+ "cel",
+ "l"
+ ],
+ [
+ "c",
+ "ell"
+ ],
+ [
+ "▁t",
+ "ill"
+ ],
+ [
+ "▁til",
+ "l"
+ ],
+ [
+ "▁ti",
+ "ll"
+ ],
+ [
+ "L",
+ "I"
+ ],
+ [
+ "dr",
+ "aw"
+ ],
+ [
+ "d",
+ "raw"
+ ],
+ [
+ "▁s",
+ "atisf"
+ ],
+ [
+ "▁sat",
+ "isf"
+ ],
+ [
+ "▁re",
+ "ading"
+ ],
+ [
+ "▁read",
+ "ing"
+ ],
+ [
+ "AT",
+ "ION"
+ ],
+ [
+ "A",
+ "TION"
+ ],
+ [
+ "▁A",
+ "re"
+ ],
+ [
+ "▁Ar",
+ "e"
+ ],
+ [
+ "▁",
+ "Are"
+ ],
+ [
+ "▁A",
+ "c"
+ ],
+ [
+ "▁",
+ "Ac"
+ ],
+ [
+ ")",
+ "*"
+ ],
+ [
+ "▁add",
+ "itional"
+ ],
+ [
+ "▁addition",
+ "al"
+ ],
+ [
+ "wo",
+ "od"
+ ],
+ [
+ "w",
+ "ood"
+ ],
+ [
+ "ci",
+ "l"
+ ],
+ [
+ "c",
+ "il"
+ ],
+ [
+ "п",
+ "у"
+ ],
+ [
+ "UL",
+ "T"
+ ],
+ [
+ "U",
+ "LT"
+ ],
+ [
+ "▁b",
+ "ill"
+ ],
+ [
+ "▁bi",
+ "ll"
+ ],
+ [
+ "▁bil",
+ "l"
+ ],
+ [
+ "ma",
+ "s"
+ ],
+ [
+ "m",
+ "as"
+ ],
+ [
+ "an",
+ "ia"
+ ],
+ [
+ "ani",
+ "a"
+ ],
+ [
+ "a",
+ "nia"
+ ],
+ [
+ "с",
+ "у"
+ ],
+ [
+ "an",
+ "z"
+ ],
+ [
+ "he",
+ "ight"
+ ],
+ [
+ "h",
+ "eight"
+ ],
+ [
+ "j",
+ "o"
+ ],
+ [
+ "▁d",
+ "os"
+ ],
+ [
+ "▁do",
+ "s"
+ ],
+ [
+ "\\",
+ "\""
+ ],
+ [
+ "▁/",
+ ">"
+ ],
+ [
+ "▁",
+ "/>"
+ ],
+ [
+ "▁p",
+ "roduction"
+ ],
+ [
+ "▁produ",
+ "ction"
+ ],
+ [
+ "▁product",
+ "ion"
+ ],
+ [
+ "▁prod",
+ "uction"
+ ],
+ [
+ "▁",
+ "production"
+ ],
+ [
+ "ig",
+ "er"
+ ],
+ [
+ "ige",
+ "r"
+ ],
+ [
+ "i",
+ "ger"
+ ],
+ [
+ "▁с",
+ "т"
+ ],
+ [
+ "▁",
+ "ст"
+ ],
+ [
+ "sh",
+ "ow"
+ ],
+ [
+ "s",
+ "how"
+ ],
+ [
+ "▁pop",
+ "ulation"
+ ],
+ [
+ "▁popul",
+ "ation"
+ ],
+ [
+ "▁p",
+ "ark"
+ ],
+ [
+ "▁par",
+ "k"
+ ],
+ [
+ "▁",
+ "park"
+ ],
+ [
+ "▁Z",
+ "e"
+ ],
+ [
+ "▁necess",
+ "ary"
+ ],
+ [
+ "▁",
+ "necessary"
+ ],
+ [
+ "▁t",
+ "rust"
+ ],
+ [
+ "▁tr",
+ "ust"
+ ],
+ [
+ "▁sh",
+ "own"
+ ],
+ [
+ "▁show",
+ "n"
+ ],
+ [
+ "mod",
+ "ule"
+ ],
+ [
+ "mo",
+ "dule"
+ ],
+ [
+ "G",
+ "E"
+ ],
+ [
+ "▁l",
+ "ay"
+ ],
+ [
+ "▁la",
+ "y"
+ ],
+ [
+ "▁",
+ "lay"
+ ],
+ [
+ "▁ann",
+ "oun"
+ ],
+ [
+ "▁class",
+ "Name"
+ ],
+ [
+ "▁",
+ "className"
+ ],
+ [
+ "▁cal",
+ "cul"
+ ],
+ [
+ "▁calc",
+ "ul"
+ ],
+ [
+ "Fun",
+ "ction"
+ ],
+ [
+ "F",
+ "unction"
+ ],
+ [
+ "▁S",
+ "al"
+ ],
+ [
+ "▁Sa",
+ "l"
+ ],
+ [
+ "▁",
+ "Sal"
+ ],
+ [
+ "O",
+ "K"
+ ],
+ [
+ "T",
+ "P"
+ ],
+ [
+ "▁en",
+ "try"
+ ],
+ [
+ "▁ent",
+ "ry"
+ ],
+ [
+ "▁entr",
+ "y"
+ ],
+ [
+ "▁",
+ "entry"
+ ],
+ [
+ "▁St",
+ "ud"
+ ],
+ [
+ "▁",
+ "Stud"
+ ],
+ [
+ "▁it",
+ "ems"
+ ],
+ [
+ "▁item",
+ "s"
+ ],
+ [
+ "▁",
+ "items"
+ ],
+ [
+ "▁se",
+ "curity"
+ ],
+ [
+ "▁sec",
+ "urity"
+ ],
+ [
+ "▁secur",
+ "ity"
+ ],
+ [
+ "▁",
+ "security"
+ ],
+ [
+ "En",
+ "try"
+ ],
+ [
+ "Ent",
+ "ry"
+ ],
+ [
+ "f",
+ "loat"
+ ],
+ [
+ "l",
+ "s"
+ ],
+ [
+ "ib",
+ "ly"
+ ],
+ [
+ "▁cont",
+ "ribut"
+ ],
+ [
+ "▁C",
+ "heck"
+ ],
+ [
+ "▁Che",
+ "ck"
+ ],
+ [
+ "▁",
+ "Check"
+ ],
+ [
+ "M",
+ "D"
+ ],
+ [
+ "▁impro",
+ "ve"
+ ],
+ [
+ "Par",
+ "t"
+ ],
+ [
+ "P",
+ "art"
+ ],
+ [
+ "▁system",
+ "s"
+ ],
+ [
+ "▁syst",
+ "ems"
+ ],
+ [
+ "B",
+ "l"
+ ],
+ [
+ "▁pol",
+ "icy"
+ ],
+ [
+ "▁polic",
+ "y"
+ ],
+ [
+ "▁",
+ "policy"
+ ],
+ [
+ "▁s",
+ "creen"
+ ],
+ [
+ "▁sc",
+ "reen"
+ ],
+ [
+ "▁scr",
+ "een"
+ ],
+ [
+ "▁",
+ "screen"
+ ],
+ [
+ "▁A",
+ "ny"
+ ],
+ [
+ "▁An",
+ "y"
+ ],
+ [
+ "▁",
+ "Any"
+ ],
+ [
+ "▁op",
+ "ened"
+ ],
+ [
+ "▁open",
+ "ed"
+ ],
+ [
+ "al",
+ "loc"
+ ],
+ [
+ "all",
+ "oc"
+ ],
+ [
+ "allo",
+ "c"
+ ],
+ [
+ "▁De",
+ "cember"
+ ],
+ [
+ "▁Dec",
+ "ember"
+ ],
+ [
+ "▁",
+ "É"
+ ],
+ [
+ "▁e",
+ "mail"
+ ],
+ [
+ "▁em",
+ "ail"
+ ],
+ [
+ "▁",
+ "email"
+ ],
+ [
+ "ad",
+ "er"
+ ],
+ [
+ "ade",
+ "r"
+ ],
+ [
+ "a",
+ "der"
+ ],
+ [
+ "=",
+ ">"
+ ],
+ [
+ "▁H",
+ "en"
+ ],
+ [
+ "▁He",
+ "n"
+ ],
+ [
+ "▁",
+ "Hen"
+ ],
+ [
+ "▁in",
+ "fo"
+ ],
+ [
+ "▁inf",
+ "o"
+ ],
+ [
+ "▁",
+ "info"
+ ],
+ [
+ "▁f",
+ "loat"
+ ],
+ [
+ "▁flo",
+ "at"
+ ],
+ [
+ "▁",
+ "float"
+ ],
+ [
+ "▁sw",
+ "itch"
+ ],
+ [
+ "▁",
+ "switch"
+ ],
+ [
+ "ра",
+ "н"
+ ],
+ [
+ "р",
+ "ан"
+ ],
+ [
+ "ur",
+ "ance"
+ ],
+ [
+ "▁as",
+ "sum"
+ ],
+ [
+ "▁ass",
+ "um"
+ ],
+ [
+ "us",
+ "tr"
+ ],
+ [
+ "ust",
+ "r"
+ ],
+ [
+ "u",
+ "str"
+ ],
+ [
+ "▁g",
+ "roups"
+ ],
+ [
+ "▁group",
+ "s"
+ ],
+ [
+ "▁gro",
+ "ups"
+ ],
+ [
+ "▁",
+ "groups"
+ ],
+ [
+ "▁R",
+ "ead"
+ ],
+ [
+ "▁Re",
+ "ad"
+ ],
+ [
+ "▁",
+ "Read"
+ ],
+ [
+ "▁w",
+ "at"
+ ],
+ [
+ "▁wa",
+ "t"
+ ],
+ [
+ "S",
+ "p"
+ ],
+ [
+ "ве",
+ "р"
+ ],
+ [
+ "в",
+ "ер"
+ ],
+ [
+ "RA",
+ "N"
+ ],
+ [
+ "R",
+ "AN"
+ ],
+ [
+ "hi",
+ "b"
+ ],
+ [
+ "h",
+ "ib"
+ ],
+ [
+ "AL",
+ "L"
+ ],
+ [
+ "A",
+ "LL"
+ ],
+ [
+ "▁h",
+ "us"
+ ],
+ [
+ "▁",
+ "hus"
+ ],
+ [
+ "Sp",
+ "ec"
+ ],
+ [
+ "Spe",
+ "c"
+ ],
+ [
+ "S",
+ "pec"
+ ],
+ [
+ "\")",
+ ")"
+ ],
+ [
+ "\"",
+ "))"
+ ],
+ [
+ "▁F",
+ "rench"
+ ],
+ [
+ "▁C",
+ "lass"
+ ],
+ [
+ "▁Cl",
+ "ass"
+ ],
+ [
+ "▁",
+ "Class"
+ ],
+ [
+ "▁pres",
+ "ident"
+ ],
+ [
+ "▁presid",
+ "ent"
+ ],
+ [
+ "▁def",
+ "init"
+ ],
+ [
+ "▁defin",
+ "it"
+ ],
+ [
+ "▁N",
+ "or"
+ ],
+ [
+ "▁No",
+ "r"
+ ],
+ [
+ "▁T",
+ "hom"
+ ],
+ [
+ "▁Th",
+ "om"
+ ],
+ [
+ "ai",
+ "gn"
+ ],
+ [
+ "a",
+ "ign"
+ ],
+ [
+ "W",
+ "idth"
+ ],
+ [
+ "D",
+ "o"
+ ],
+ [
+ "▁{",
+ "@"
+ ],
+ [
+ "ag",
+ "on"
+ ],
+ [
+ "ago",
+ "n"
+ ],
+ [
+ "a",
+ "gon"
+ ],
+ [
+ "▁L",
+ "u"
+ ],
+ [
+ "▁",
+ "Lu"
+ ],
+ [
+ "▁follow",
+ "ed"
+ ],
+ [
+ "M",
+ "M"
+ ],
+ [
+ "as",
+ "ons"
+ ],
+ [
+ "ason",
+ "s"
+ ],
+ [
+ "tm",
+ "p"
+ ],
+ [
+ "t",
+ "mp"
+ ],
+ [
+ "▁th",
+ "rows"
+ ],
+ [
+ "▁throw",
+ "s"
+ ],
+ [
+ "▁thr",
+ "ows"
+ ],
+ [
+ "▁thro",
+ "ws"
+ ],
+ [
+ "▁",
+ "throws"
+ ],
+ [
+ "IT",
+ "Y"
+ ],
+ [
+ "I",
+ "TY"
+ ],
+ [
+ "но",
+ "м"
+ ],
+ [
+ "▁f",
+ "air"
+ ],
+ [
+ "▁fa",
+ "ir"
+ ],
+ [
+ "▁p",
+ "en"
+ ],
+ [
+ "▁pe",
+ "n"
+ ],
+ [
+ "▁",
+ "pen"
+ ],
+ [
+ "é",
+ "g"
+ ],
+ [
+ "▁inter",
+ "face"
+ ],
+ [
+ "▁",
+ "interface"
+ ],
+ [
+ "▁s",
+ "af"
+ ],
+ [
+ "▁sa",
+ "f"
+ ],
+ [
+ "oo",
+ "n"
+ ],
+ [
+ "o",
+ "on"
+ ],
+ [
+ "B",
+ "ack"
+ ],
+ [
+ "▁s",
+ "peed"
+ ],
+ [
+ "▁sp",
+ "eed"
+ ],
+ [
+ "▁spe",
+ "ed"
+ ],
+ [
+ "▁",
+ "speed"
+ ],
+ [
+ "▁ext",
+ "ends"
+ ],
+ [
+ "▁extend",
+ "s"
+ ],
+ [
+ "em",
+ "pty"
+ ],
+ [
+ "empt",
+ "y"
+ ],
+ [
+ "emp",
+ "ty"
+ ],
+ [
+ "▁п",
+ "ере"
+ ],
+ [
+ "▁пер",
+ "е"
+ ],
+ [
+ "▁пе",
+ "ре"
+ ],
+ [
+ "▁pro",
+ "per"
+ ],
+ [
+ "▁pr",
+ "oper"
+ ],
+ [
+ "▁prop",
+ "er"
+ ],
+ [
+ "▁d",
+ "riv"
+ ],
+ [
+ "▁dr",
+ "iv"
+ ],
+ [
+ "▁dri",
+ "v"
+ ],
+ [
+ "ф",
+ "и"
+ ],
+ [
+ "▁c",
+ "enter"
+ ],
+ [
+ "▁cent",
+ "er"
+ ],
+ [
+ "▁",
+ "center"
+ ],
+ [
+ "he",
+ "ader"
+ ],
+ [
+ "head",
+ "er"
+ ],
+ [
+ "▁}",
+ ")"
+ ],
+ [
+ "▁",
+ "})"
+ ],
+ [
+ "w",
+ "a"
+ ],
+ [
+ "▁m",
+ "iddle"
+ ],
+ [
+ "▁",
+ "middle"
+ ],
+ [
+ "▁ch",
+ "oose"
+ ],
+ [
+ "▁cho",
+ "ose"
+ ],
+ [
+ "▁St",
+ "ad"
+ ],
+ [
+ "▁Sta",
+ "d"
+ ],
+ [
+ "S",
+ "O"
+ ],
+ [
+ "Fact",
+ "ory"
+ ],
+ [
+ "Factor",
+ "y"
+ ],
+ [
+ "F",
+ "actory"
+ ],
+ [
+ "De",
+ "v"
+ ],
+ [
+ "D",
+ "ev"
+ ],
+ [
+ "ic",
+ "les"
+ ],
+ [
+ "icle",
+ "s"
+ ],
+ [
+ "icl",
+ "es"
+ ],
+ [
+ "i",
+ "cles"
+ ],
+ [
+ "▁ap",
+ "plication"
+ ],
+ [
+ "▁applic",
+ "ation"
+ ],
+ [
+ "▁appl",
+ "ication"
+ ],
+ [
+ "▁",
+ "application"
+ ],
+ [
+ "▁mod",
+ "els"
+ ],
+ [
+ "▁model",
+ "s"
+ ],
+ [
+ "▁mode",
+ "ls"
+ ],
+ [
+ "▁",
+ "models"
+ ],
+ [
+ "pi",
+ "te"
+ ],
+ [
+ "pit",
+ "e"
+ ],
+ [
+ "p",
+ "ite"
+ ],
+ [
+ "ca",
+ "p"
+ ],
+ [
+ "c",
+ "ap"
+ ],
+ [
+ "x",
+ "i"
+ ],
+ [
+ "osp",
+ "ital"
+ ],
+ [
+ "▁d",
+ "ream"
+ ],
+ [
+ "▁dre",
+ "am"
+ ],
+ [
+ "EN",
+ "D"
+ ],
+ [
+ "E",
+ "ND"
+ ],
+ [
+ "▁con",
+ "tract"
+ ],
+ [
+ "▁cont",
+ "ract"
+ ],
+ [
+ "▁contr",
+ "act"
+ ],
+ [
+ "▁contra",
+ "ct"
+ ],
+ [
+ "▁",
+ "contract"
+ ],
+ [
+ "icro",
+ "soft"
+ ],
+ [
+ "▁th",
+ "ous"
+ ],
+ [
+ "▁thou",
+ "s"
+ ],
+ [
+ "iz",
+ "es"
+ ],
+ [
+ "ize",
+ "s"
+ ],
+ [
+ "i",
+ "zes"
+ ],
+ [
+ "▁д",
+ "а"
+ ],
+ [
+ "▁",
+ "да"
+ ],
+ [
+ "▁C",
+ "O"
+ ],
+ [
+ "▁",
+ "CO"
+ ],
+ [
+ "▁d",
+ "irection"
+ ],
+ [
+ "▁di",
+ "rection"
+ ],
+ [
+ "▁direct",
+ "ion"
+ ],
+ [
+ "▁dire",
+ "ction"
+ ],
+ [
+ "▁dir",
+ "ection"
+ ],
+ [
+ "▁",
+ "direction"
+ ],
+ [
+ "▁`",
+ "`"
+ ],
+ [
+ "▁",
+ "``"
+ ],
+ [
+ "▁d",
+ "rive"
+ ],
+ [
+ "▁dr",
+ "ive"
+ ],
+ [
+ "▁dri",
+ "ve"
+ ],
+ [
+ "▁driv",
+ "e"
+ ],
+ [
+ "▁",
+ "drive"
+ ],
+ [
+ "Ma",
+ "x"
+ ],
+ [
+ "M",
+ "ax"
+ ],
+ [
+ "ci",
+ "a"
+ ],
+ [
+ "c",
+ "ia"
+ ],
+ [
+ "▁contin",
+ "u"
+ ],
+ [
+ "▁A",
+ "lex"
+ ],
+ [
+ "▁Al",
+ "ex"
+ ],
+ [
+ "▁Ale",
+ "x"
+ ],
+ [
+ "▁",
+ "Alex"
+ ],
+ [
+ "▁g",
+ "old"
+ ],
+ [
+ "▁go",
+ "ld"
+ ],
+ [
+ "▁gol",
+ "d"
+ ],
+ [
+ "▁",
+ "gold"
+ ],
+ [
+ "▁p",
+ "rep"
+ ],
+ [
+ "▁pre",
+ "p"
+ ],
+ [
+ "▁pr",
+ "ep"
+ ],
+ [
+ "▁or",
+ "igin"
+ ],
+ [
+ "▁orig",
+ "in"
+ ],
+ [
+ "▁",
+ "origin"
+ ],
+ [
+ "▁r",
+ "ap"
+ ],
+ [
+ "▁ra",
+ "p"
+ ],
+ [
+ "▁",
+ "rap"
+ ],
+ [
+ "O",
+ "p"
+ ],
+ [
+ "ous",
+ "ly"
+ ],
+ [
+ "▁are",
+ "as"
+ ],
+ [
+ "▁area",
+ "s"
+ ],
+ [
+ "PO",
+ "RT"
+ ],
+ [
+ "P",
+ "ORT"
+ ],
+ [
+ "он",
+ "а"
+ ],
+ [
+ "о",
+ "на"
+ ],
+ [
+ "▁sa",
+ "fe"
+ ],
+ [
+ "▁saf",
+ "e"
+ ],
+ [
+ "▁",
+ "safe"
+ ],
+ [
+ "▁profess",
+ "ional"
+ ],
+ [
+ "▁profession",
+ "al"
+ ],
+ [
+ "ap",
+ "ache"
+ ],
+ [
+ "apa",
+ "che"
+ ],
+ [
+ "▁t",
+ "emper"
+ ],
+ [
+ "▁tem",
+ "per"
+ ],
+ [
+ "▁temp",
+ "er"
+ ],
+ [
+ "s",
+ "z"
+ ],
+ [
+ "▁u",
+ "nit"
+ ],
+ [
+ "▁un",
+ "it"
+ ],
+ [
+ "▁",
+ "unit"
+ ],
+ [
+ "▁c",
+ "op"
+ ],
+ [
+ "▁co",
+ "p"
+ ],
+ [
+ "▁",
+ "cop"
+ ],
+ [
+ "eq",
+ "n"
+ ],
+ [
+ "List",
+ "ener"
+ ],
+ [
+ "Listen",
+ "er"
+ ],
+ [
+ "▁for",
+ "mat"
+ ],
+ [
+ "▁form",
+ "at"
+ ],
+ [
+ "▁forma",
+ "t"
+ ],
+ [
+ "▁",
+ "format"
+ ],
+ [
+ "se",
+ "lect"
+ ],
+ [
+ "sel",
+ "ect"
+ ],
+ [
+ "s",
+ "elect"
+ ],
+ [
+ "▁com",
+ "fort"
+ ],
+ [
+ "▁",
+ "comfort"
+ ],
+ [
+ "▁me",
+ "ant"
+ ],
+ [
+ "▁mean",
+ "t"
+ ],
+ [
+ "id",
+ "ay"
+ ],
+ [
+ "ida",
+ "y"
+ ],
+ [
+ "i",
+ "day"
+ ],
+ [
+ "em",
+ "e"
+ ],
+ [
+ "e",
+ "me"
+ ],
+ [
+ "▁act",
+ "ive"
+ ],
+ [
+ "▁activ",
+ "e"
+ ],
+ [
+ "▁",
+ "active"
+ ],
+ [
+ "▁n",
+ "ote"
+ ],
+ [
+ "▁not",
+ "e"
+ ],
+ [
+ "▁no",
+ "te"
+ ],
+ [
+ "▁",
+ "note"
+ ],
+ [
+ "▁M",
+ "il"
+ ],
+ [
+ "▁Mi",
+ "l"
+ ],
+ [
+ "▁",
+ "Mil"
+ ],
+ [
+ "on",
+ "ly"
+ ],
+ [
+ "▁<",
+ "="
+ ],
+ [
+ "▁",
+ "<="
+ ],
+ [
+ "▁ne",
+ "igh"
+ ],
+ [
+ "▁nei",
+ "gh"
+ ],
+ [
+ "a",
+ "o"
+ ],
+ [
+ "▁bl",
+ "ue"
+ ],
+ [
+ "▁",
+ "blue"
+ ],
+ [
+ "▁T",
+ "V"
+ ],
+ [
+ "▁",
+ "TV"
+ ],
+ [
+ "Ch",
+ "ild"
+ ],
+ [
+ "▁re",
+ "ached"
+ ],
+ [
+ "▁reach",
+ "ed"
+ ],
+ [
+ "Add",
+ "ress"
+ ],
+ [
+ "Addr",
+ "ess"
+ ],
+ [
+ "ст",
+ "в"
+ ],
+ [
+ "▁cl",
+ "osed"
+ ],
+ [
+ "▁close",
+ "d"
+ ],
+ [
+ "▁clos",
+ "ed"
+ ],
+ [
+ "▁clo",
+ "sed"
+ ],
+ [
+ "▁",
+ "closed"
+ ],
+ [
+ "in",
+ "der"
+ ],
+ [
+ "ind",
+ "er"
+ ],
+ [
+ "inde",
+ "r"
+ ],
+ [
+ "i",
+ "nder"
+ ],
+ [
+ "ol",
+ "o"
+ ],
+ [
+ "o",
+ "lo"
+ ],
+ [
+ "▁a",
+ "lt"
+ ],
+ [
+ "▁al",
+ "t"
+ ],
+ [
+ "▁",
+ "alt"
+ ],
+ [
+ "▁a",
+ "dm"
+ ],
+ [
+ "▁ad",
+ "m"
+ ],
+ [
+ "Form",
+ "at"
+ ],
+ [
+ "For",
+ "mat"
+ ],
+ [
+ "U",
+ "I"
+ ],
+ [
+ "▁H",
+ "am"
+ ],
+ [
+ "▁Ha",
+ "m"
+ ],
+ [
+ "▁f",
+ "requ"
+ ],
+ [
+ "▁fr",
+ "equ"
+ ],
+ [
+ "▁fre",
+ "qu"
+ ],
+ [
+ "▁in",
+ "depend"
+ ],
+ [
+ "▁inde",
+ "pend"
+ ],
+ [
+ "▁",
+ "independ"
+ ],
+ [
+ "▁eas",
+ "ily"
+ ],
+ [
+ "▁L",
+ "and"
+ ],
+ [
+ "▁La",
+ "nd"
+ ],
+ [
+ "▁Lan",
+ "d"
+ ],
+ [
+ "▁",
+ "Land"
+ ],
+ [
+ "▁t",
+ "or"
+ ],
+ [
+ "▁to",
+ "r"
+ ],
+ [
+ "▁",
+ "tor"
+ ],
+ [
+ "ograph",
+ "y"
+ ],
+ [
+ "ograp",
+ "hy"
+ ],
+ [
+ "in",
+ "fty"
+ ],
+ [
+ "inf",
+ "ty"
+ ],
+ [
+ "▁W",
+ "ork"
+ ],
+ [
+ "▁Wor",
+ "k"
+ ],
+ [
+ "▁",
+ "Work"
+ ],
+ [
+ "iv",
+ "en"
+ ],
+ [
+ "ive",
+ "n"
+ ],
+ [
+ "i",
+ "ven"
+ ],
+ [
+ "▁Count",
+ "y"
+ ],
+ [
+ "▁Coun",
+ "ty"
+ ],
+ [
+ "▁s",
+ "rc"
+ ],
+ [
+ "▁",
+ "src"
+ ],
+ [
+ "}$",
+ ","
+ ],
+ [
+ "}",
+ "$,"
+ ],
+ [
+ "par",
+ "se"
+ ],
+ [
+ "pars",
+ "e"
+ ],
+ [
+ "p",
+ "arse"
+ ],
+ [
+ "C",
+ "D"
+ ],
+ [
+ "▁C",
+ "our"
+ ],
+ [
+ "▁Co",
+ "ur"
+ ],
+ [
+ "▁Cou",
+ "r"
+ ],
+ [
+ "▁f",
+ "ol"
+ ],
+ [
+ "▁fo",
+ "l"
+ ],
+ [
+ "▁",
+ "fol"
+ ],
+ [
+ "Ent",
+ "ity"
+ ],
+ [
+ "pg",
+ "f"
+ ],
+ [
+ "▁Ch",
+ "ina"
+ ],
+ [
+ "▁Chi",
+ "na"
+ ],
+ [
+ "▁S",
+ "ub"
+ ],
+ [
+ "▁Su",
+ "b"
+ ],
+ [
+ "▁",
+ "Sub"
+ ],
+ [
+ "ho",
+ "od"
+ ],
+ [
+ "h",
+ "ood"
+ ],
+ [
+ "▁field",
+ "s"
+ ],
+ [
+ "▁",
+ "fields"
+ ],
+ [
+ "▁y",
+ "es"
+ ],
+ [
+ "▁ye",
+ "s"
+ ],
+ [
+ "▁",
+ "yes"
+ ],
+ [
+ "re",
+ "nd"
+ ],
+ [
+ "ren",
+ "d"
+ ],
+ [
+ "r",
+ "end"
+ ],
+ [
+ "▁to",
+ "wards"
+ ],
+ [
+ "▁toward",
+ "s"
+ ],
+ [
+ "▁tow",
+ "ards"
+ ],
+ [
+ "▁st",
+ "aff"
+ ],
+ [
+ "▁sta",
+ "ff"
+ ],
+ [
+ "▁",
+ "staff"
+ ],
+ [
+ "▁A",
+ "ir"
+ ],
+ [
+ "▁",
+ "Air"
+ ],
+ [
+ "▁st",
+ "ation"
+ ],
+ [
+ "▁stat",
+ "ion"
+ ],
+ [
+ "▁",
+ "station"
+ ],
+ [
+ "at",
+ "ives"
+ ],
+ [
+ "ative",
+ "s"
+ ],
+ [
+ "ati",
+ "ves"
+ ],
+ [
+ "ativ",
+ "es"
+ ],
+ [
+ "▁imp",
+ "act"
+ ],
+ [
+ "в",
+ "ы"
+ ],
+ [
+ "▁direct",
+ "ly"
+ ],
+ [
+ "iss",
+ "ions"
+ ],
+ [
+ "ission",
+ "s"
+ ],
+ [
+ "iv",
+ "a"
+ ],
+ [
+ "i",
+ "va"
+ ],
+ [
+ "|",
+ "\\"
+ ],
+ [
+ "Pt",
+ "r"
+ ],
+ [
+ "P",
+ "tr"
+ ],
+ [
+ "▁S",
+ "ant"
+ ],
+ [
+ "▁San",
+ "t"
+ ],
+ [
+ "▁Sa",
+ "nt"
+ ],
+ [
+ "Po",
+ "l"
+ ],
+ [
+ "P",
+ "ol"
+ ],
+ [
+ "▁pro",
+ "gress"
+ ],
+ [
+ "▁",
+ "progress"
+ ],
+ [
+ "it",
+ "ar"
+ ],
+ [
+ "ita",
+ "r"
+ ],
+ [
+ "i",
+ "tar"
+ ],
+ [
+ "▁p",
+ "arts"
+ ],
+ [
+ "▁part",
+ "s"
+ ],
+ [
+ "▁par",
+ "ts"
+ ],
+ [
+ "▁",
+ "parts"
+ ],
+ [
+ "▁pl",
+ "ant"
+ ],
+ [
+ "▁plan",
+ "t"
+ ],
+ [
+ "▁",
+ "plant"
+ ],
+ [
+ "▁abs",
+ "olut"
+ ],
+ [
+ "▁gu",
+ "ess"
+ ],
+ [
+ "eq",
+ "ref"
+ ],
+ [
+ "▁t",
+ "im"
+ ],
+ [
+ "▁ti",
+ "m"
+ ],
+ [
+ "▁",
+ "tim"
+ ],
+ [
+ "▁L",
+ "ou"
+ ],
+ [
+ "▁Lo",
+ "u"
+ ],
+ [
+ "▁",
+ "Lou"
+ ],
+ [
+ "▁c",
+ "ool"
+ ],
+ [
+ "▁co",
+ "ol"
+ ],
+ [
+ "al",
+ "u"
+ ],
+ [
+ "a",
+ "lu"
+ ],
+ [
+ "▁m",
+ "outh"
+ ],
+ [
+ "▁mo",
+ "uth"
+ ],
+ [
+ "▁mou",
+ "th"
+ ],
+ [
+ "▁",
+ "mouth"
+ ],
+ [
+ "ни",
+ "х"
+ ],
+ [
+ "▁h",
+ "eight"
+ ],
+ [
+ "▁he",
+ "ight"
+ ],
+ [
+ "▁",
+ "height"
+ ],
+ [
+ "ge",
+ "st"
+ ],
+ [
+ "ges",
+ "t"
+ ],
+ [
+ "g",
+ "est"
+ ],
+ [
+ "▁P",
+ "ost"
+ ],
+ [
+ "▁Po",
+ "st"
+ ],
+ [
+ "▁Pos",
+ "t"
+ ],
+ [
+ "▁",
+ "Post"
+ ],
+ [
+ "▁b",
+ "oard"
+ ],
+ [
+ "▁bo",
+ "ard"
+ ],
+ [
+ "▁",
+ "board"
+ ],
+ [
+ "▁t",
+ "it"
+ ],
+ [
+ "▁ti",
+ "t"
+ ],
+ [
+ "▁",
+ "tit"
+ ],
+ [
+ "▁h",
+ "our"
+ ],
+ [
+ "▁ho",
+ "ur"
+ ],
+ [
+ "▁",
+ "hour"
+ ],
+ [
+ "▁ser",
+ "ver"
+ ],
+ [
+ "▁serv",
+ "er"
+ ],
+ [
+ "▁serve",
+ "r"
+ ],
+ [
+ "▁",
+ "server"
+ ],
+ [
+ "▁p",
+ "layers"
+ ],
+ [
+ "▁play",
+ "ers"
+ ],
+ [
+ "▁player",
+ "s"
+ ],
+ [
+ "ri",
+ "er"
+ ],
+ [
+ "rie",
+ "r"
+ ],
+ [
+ "r",
+ "ier"
+ ],
+ [
+ "Lin",
+ "k"
+ ],
+ [
+ "L",
+ "ink"
+ ],
+ [
+ "▁Pres",
+ "ident"
+ ],
+ [
+ "]",
+ "("
+ ],
+ [
+ "▁con",
+ "struct"
+ ],
+ [
+ "▁const",
+ "ruct"
+ ],
+ [
+ "▁constr",
+ "uct"
+ ],
+ [
+ "▁constru",
+ "ct"
+ ],
+ [
+ "▁",
+ "construct"
+ ],
+ [
+ "hand",
+ "le"
+ ],
+ [
+ "}$",
+ "."
+ ],
+ [
+ "}",
+ "$."
+ ],
+ [
+ "ry",
+ "ing"
+ ],
+ [
+ "r",
+ "ying"
+ ],
+ [
+ "▁s",
+ "hop"
+ ],
+ [
+ "▁sh",
+ "op"
+ ],
+ [
+ "▁",
+ "shop"
+ ],
+ [
+ "ia",
+ "na"
+ ],
+ [
+ "ian",
+ "a"
+ ],
+ [
+ "i",
+ "ana"
+ ],
+ [
+ "ex",
+ "p"
+ ],
+ [
+ "e",
+ "xp"
+ ],
+ [
+ "Hel",
+ "per"
+ ],
+ [
+ "Help",
+ "er"
+ ],
+ [
+ "Off",
+ "set"
+ ],
+ [
+ "ac",
+ "hes"
+ ],
+ [
+ "ach",
+ "es"
+ ],
+ [
+ "ache",
+ "s"
+ ],
+ [
+ "a",
+ "ches"
+ ],
+ [
+ "▁conne",
+ "ction"
+ ],
+ [
+ "▁connect",
+ "ion"
+ ],
+ [
+ "▁conn",
+ "ection"
+ ],
+ [
+ "▁",
+ "connection"
+ ],
+ [
+ "▁d",
+ "ifference"
+ ],
+ [
+ "▁dif",
+ "ference"
+ ],
+ [
+ "▁differ",
+ "ence"
+ ],
+ [
+ "serv",
+ "ice"
+ ],
+ [
+ "s",
+ "ervice"
+ ],
+ [
+ "▁g",
+ "as"
+ ],
+ [
+ "▁ga",
+ "s"
+ ],
+ [
+ "▁",
+ "gas"
+ ],
+ [
+ "▁p",
+ "riv"
+ ],
+ [
+ "▁pr",
+ "iv"
+ ],
+ [
+ "▁pri",
+ "v"
+ ],
+ [
+ "▁",
+ "priv"
+ ],
+ [
+ "▁un",
+ "ivers"
+ ],
+ [
+ "▁",
+ "univers"
+ ],
+ [
+ "▁w",
+ "ish"
+ ],
+ [
+ "▁wis",
+ "h"
+ ],
+ [
+ "Re",
+ "m"
+ ],
+ [
+ "R",
+ "em"
+ ],
+ [
+ "U",
+ "rl"
+ ],
+ [
+ "ge",
+ "b"
+ ],
+ [
+ "g",
+ "eb"
+ ],
+ [
+ "S",
+ "o"
+ ],
+ [
+ "ens",
+ "ions"
+ ],
+ [
+ "ension",
+ "s"
+ ],
+ [
+ "Mod",
+ "ule"
+ ],
+ [
+ "Mo",
+ "dule"
+ ],
+ [
+ "SI",
+ "ZE"
+ ],
+ [
+ "▁p",
+ "rem"
+ ],
+ [
+ "▁pre",
+ "m"
+ ],
+ [
+ "▁pr",
+ "em"
+ ],
+ [
+ "wind",
+ "ow"
+ ],
+ [
+ "w",
+ "indow"
+ ],
+ [
+ "▁d",
+ "ies"
+ ],
+ [
+ "▁di",
+ "es"
+ ],
+ [
+ "▁die",
+ "s"
+ ],
+ [
+ "de",
+ "l"
+ ],
+ [
+ "d",
+ "el"
+ ],
+ [
+ "▁r",
+ "ow"
+ ],
+ [
+ "▁ro",
+ "w"
+ ],
+ [
+ "▁",
+ "row"
+ ],
+ [
+ "▁a",
+ "verage"
+ ],
+ [
+ "▁aver",
+ "age"
+ ],
+ [
+ "▁ave",
+ "rage"
+ ],
+ [
+ "xi",
+ "m"
+ ],
+ [
+ "x",
+ "im"
+ ],
+ [
+ "▁p",
+ "u"
+ ],
+ [
+ "▁",
+ "pu"
+ ],
+ [
+ "an",
+ "ç"
+ ],
+ [
+ "De",
+ "t"
+ ],
+ [
+ "D",
+ "et"
+ ],
+ [
+ "ke",
+ "r"
+ ],
+ [
+ "k",
+ "er"
+ ],
+ [
+ "y",
+ "a"
+ ],
+ [
+ "▁D",
+ "et"
+ ],
+ [
+ "▁De",
+ "t"
+ ],
+ [
+ "▁",
+ "Det"
+ ],
+ [
+ "▁p",
+ "å"
+ ],
+ [
+ "▁n",
+ "amed"
+ ],
+ [
+ "▁name",
+ "d"
+ ],
+ [
+ "▁na",
+ "med"
+ ],
+ [
+ "▁nam",
+ "ed"
+ ],
+ [
+ "▁",
+ "named"
+ ],
+ [
+ "▁dec",
+ "ision"
+ ],
+ [
+ "▁decis",
+ "ion"
+ ],
+ [
+ "wi",
+ "n"
+ ],
+ [
+ "w",
+ "in"
+ ],
+ [
+ "▁Ge",
+ "orge"
+ ],
+ [
+ "▁Georg",
+ "e"
+ ],
+ [
+ "ar",
+ "ily"
+ ],
+ [
+ "ari",
+ "ly"
+ ],
+ [
+ "▁s",
+ "olution"
+ ],
+ [
+ "▁sol",
+ "ution"
+ ],
+ [
+ "▁mult",
+ "iple"
+ ],
+ [
+ "▁multi",
+ "ple"
+ ],
+ [
+ "▁multip",
+ "le"
+ ],
+ [
+ "▁",
+ "multiple"
+ ],
+ [
+ "at",
+ "egy"
+ ],
+ [
+ "ate",
+ "gy"
+ ],
+ [
+ "ateg",
+ "y"
+ ],
+ [
+ "▁le",
+ "arning"
+ ],
+ [
+ "▁learn",
+ "ing"
+ ],
+ [
+ "▁lear",
+ "ning"
+ ],
+ [
+ "▁",
+ "learning"
+ ],
+ [
+ "▁se",
+ "cret"
+ ],
+ [
+ "▁sec",
+ "ret"
+ ],
+ [
+ "▁secre",
+ "t"
+ ],
+ [
+ "▁",
+ "secret"
+ ],
+ [
+ "D",
+ "O"
+ ],
+ [
+ "▁n",
+ "ice"
+ ],
+ [
+ "▁ni",
+ "ce"
+ ],
+ [
+ "▁nic",
+ "e"
+ ],
+ [
+ "▁",
+ "nice"
+ ],
+ [
+ "////////",
+ "////////"
+ ],
+ [
+ "S",
+ "u"
+ ],
+ [
+ "it",
+ "ation"
+ ],
+ [
+ "itat",
+ "ion"
+ ],
+ [
+ "▁j",
+ "oin"
+ ],
+ [
+ "▁jo",
+ "in"
+ ],
+ [
+ "▁",
+ "join"
+ ],
+ [
+ "▁el",
+ "ements"
+ ],
+ [
+ "▁element",
+ "s"
+ ],
+ [
+ "▁ele",
+ "ments"
+ ],
+ [
+ "▁elem",
+ "ents"
+ ],
+ [
+ "▁",
+ "elements"
+ ],
+ [
+ "▁e",
+ "mer"
+ ],
+ [
+ "▁em",
+ "er"
+ ],
+ [
+ "til",
+ "de"
+ ],
+ [
+ "t",
+ "ilde"
+ ],
+ [
+ "▁d",
+ "ep"
+ ],
+ [
+ "▁de",
+ "p"
+ ],
+ [
+ "▁",
+ "dep"
+ ],
+ [
+ "▁s",
+ "hot"
+ ],
+ [
+ "▁sh",
+ "ot"
+ ],
+ [
+ "▁",
+ "shot"
+ ],
+ [
+ "▁pl",
+ "atform"
+ ],
+ [
+ "▁plat",
+ "form"
+ ],
+ [
+ "▁",
+ "platform"
+ ],
+ [
+ "ot",
+ "hing"
+ ],
+ [
+ "oth",
+ "ing"
+ ],
+ [
+ "o",
+ "thing"
+ ],
+ [
+ "M",
+ "y"
+ ],
+ [
+ "ed",
+ "ia"
+ ],
+ [
+ "edi",
+ "a"
+ ],
+ [
+ "om",
+ "s"
+ ],
+ [
+ "o",
+ "ms"
+ ],
+ [
+ "ail",
+ "y"
+ ],
+ [
+ "ai",
+ "ly"
+ ],
+ [
+ "a",
+ "ily"
+ ],
+ [
+ "(",
+ "["
+ ],
+ [
+ "▁d",
+ "ress"
+ ],
+ [
+ "▁dr",
+ "ess"
+ ],
+ [
+ "▁dre",
+ "ss"
+ ],
+ [
+ "▁off",
+ "icial"
+ ],
+ [
+ "▁offic",
+ "ial"
+ ],
+ [
+ "es",
+ "tern"
+ ],
+ [
+ "est",
+ "ern"
+ ],
+ [
+ "ester",
+ "n"
+ ],
+ [
+ "este",
+ "rn"
+ ],
+ [
+ "▁dis",
+ "cover"
+ ],
+ [
+ "▁disc",
+ "over"
+ ],
+ [
+ "▁disco",
+ "ver"
+ ],
+ [
+ "▁m",
+ "i"
+ ],
+ [
+ "▁",
+ "mi"
+ ],
+ [
+ "ны",
+ "е"
+ ],
+ [
+ "C",
+ "A"
+ ],
+ [
+ "od",
+ "ing"
+ ],
+ [
+ "odi",
+ "ng"
+ ],
+ [
+ "o",
+ "ding"
+ ],
+ [
+ "▁F",
+ "ound"
+ ],
+ [
+ "▁Fou",
+ "nd"
+ ],
+ [
+ "▁Fo",
+ "und"
+ ],
+ [
+ "▁",
+ "Found"
+ ],
+ [
+ "▁a",
+ "ffect"
+ ],
+ [
+ "▁aff",
+ "ect"
+ ],
+ [
+ "▁af",
+ "fect"
+ ],
+ [
+ "Vi",
+ "s"
+ ],
+ [
+ "V",
+ "is"
+ ],
+ [
+ "st",
+ "ract"
+ ],
+ [
+ "str",
+ "act"
+ ],
+ [
+ "stra",
+ "ct"
+ ],
+ [
+ "s",
+ "tract"
+ ],
+ [
+ "ic",
+ "ed"
+ ],
+ [
+ "ice",
+ "d"
+ ],
+ [
+ "i",
+ "ced"
+ ],
+ [
+ "de",
+ "bug"
+ ],
+ [
+ "d",
+ "ebug"
+ ],
+ [
+ "▁rel",
+ "ated"
+ ],
+ [
+ "▁relate",
+ "d"
+ ],
+ [
+ "▁",
+ "related"
+ ],
+ [
+ "▁s",
+ "pect"
+ ],
+ [
+ "▁sp",
+ "ect"
+ ],
+ [
+ "▁spec",
+ "t"
+ ],
+ [
+ "▁spe",
+ "ct"
+ ],
+ [
+ "▁",
+ "spect"
+ ],
+ [
+ "us",
+ "hed"
+ ],
+ [
+ "ush",
+ "ed"
+ ],
+ [
+ "сь",
+ "ко"
+ ],
+ [
+ "▁b",
+ "ank"
+ ],
+ [
+ "▁ban",
+ "k"
+ ],
+ [
+ "▁",
+ "bank"
+ ],
+ [
+ "▁c",
+ "ele"
+ ],
+ [
+ "▁ce",
+ "le"
+ ],
+ [
+ "▁cel",
+ "e"
+ ],
+ [
+ "AN",
+ "D"
+ ],
+ [
+ "A",
+ "ND"
+ ],
+ [
+ "ol",
+ "f"
+ ],
+ [
+ "е",
+ "м"
+ ],
+ [
+ "▁f",
+ "ill"
+ ],
+ [
+ "▁fil",
+ "l"
+ ],
+ [
+ "▁fi",
+ "ll"
+ ],
+ [
+ "▁",
+ "fill"
+ ],
+ [
+ "▁g",
+ "ives"
+ ],
+ [
+ "▁giv",
+ "es"
+ ],
+ [
+ "▁give",
+ "s"
+ ],
+ [
+ "▁gi",
+ "ves"
+ ],
+ [
+ "▁б",
+ "у"
+ ],
+ [
+ "▁",
+ "бу"
+ ],
+ [
+ "ar",
+ "on"
+ ],
+ [
+ "aro",
+ "n"
+ ],
+ [
+ "a",
+ "ron"
+ ],
+ [
+ "▁J",
+ "es"
+ ],
+ [
+ "▁Je",
+ "s"
+ ],
+ [
+ "RE",
+ "G"
+ ],
+ [
+ "▁s",
+ "udd"
+ ],
+ [
+ "▁su",
+ "dd"
+ ],
+ [
+ "▁sud",
+ "d"
+ ],
+ [
+ "date",
+ "d"
+ ],
+ [
+ "da",
+ "ted"
+ ],
+ [
+ "dat",
+ "ed"
+ ],
+ [
+ "d",
+ "ated"
+ ],
+ [
+ "v",
+ "i"
+ ],
+ [
+ "▁g",
+ "i"
+ ],
+ [
+ "▁",
+ "gi"
+ ],
+ [
+ "se",
+ "nd"
+ ],
+ [
+ "sen",
+ "d"
+ ],
+ [
+ "s",
+ "end"
+ ],
+ [
+ "cp",
+ "p"
+ ],
+ [
+ "c",
+ "pp"
+ ],
+ [
+ "▁s",
+ "pent"
+ ],
+ [
+ "▁sp",
+ "ent"
+ ],
+ [
+ "▁spe",
+ "nt"
+ ],
+ [
+ "an",
+ "de"
+ ],
+ [
+ "and",
+ "e"
+ ],
+ [
+ "a",
+ "nde"
+ ],
+ [
+ "▁oper",
+ "ation"
+ ],
+ [
+ "▁",
+ "operation"
+ ],
+ [
+ "pro",
+ "cess"
+ ],
+ [
+ "proc",
+ "ess"
+ ],
+ [
+ "▁in",
+ "form"
+ ],
+ [
+ "▁inf",
+ "orm"
+ ],
+ [
+ "▁info",
+ "rm"
+ ],
+ [
+ "▁F",
+ "ree"
+ ],
+ [
+ "▁Fr",
+ "ee"
+ ],
+ [
+ "▁Fre",
+ "e"
+ ],
+ [
+ "▁",
+ "Free"
+ ],
+ [
+ "yo",
+ "nd"
+ ],
+ [
+ "y",
+ "ond"
+ ],
+ [
+ "▁per",
+ "haps"
+ ],
+ [
+ "▁su",
+ "rv"
+ ],
+ [
+ "▁sur",
+ "v"
+ ],
+ [
+ "▁L",
+ "oc"
+ ],
+ [
+ "▁Lo",
+ "c"
+ ],
+ [
+ "▁",
+ "Loc"
+ ],
+ [
+ "▁con",
+ "cl"
+ ],
+ [
+ "▁conc",
+ "l"
+ ],
+ [
+ "▁ра",
+ "з"
+ ],
+ [
+ "▁",
+ "раз"
+ ],
+ [
+ "▁O",
+ "ver"
+ ],
+ [
+ "▁",
+ "Over"
+ ],
+ [
+ "ho",
+ "l"
+ ],
+ [
+ "h",
+ "ol"
+ ],
+ [
+ "ra",
+ "z"
+ ],
+ [
+ "r",
+ "az"
+ ],
+ [
+ "Wr",
+ "ite"
+ ],
+ [
+ "Writ",
+ "e"
+ ],
+ [
+ "W",
+ "rite"
+ ],
+ [
+ "▁g",
+ "iving"
+ ],
+ [
+ "▁giv",
+ "ing"
+ ],
+ [
+ "▁gi",
+ "ving"
+ ],
+ [
+ "r",
+ "d"
+ ],
+ [
+ "in",
+ "stance"
+ ],
+ [
+ "inst",
+ "ance"
+ ],
+ [
+ "▁re",
+ "leased"
+ ],
+ [
+ "▁rele",
+ "ased"
+ ],
+ [
+ "▁release",
+ "d"
+ ],
+ [
+ "▁R",
+ "o"
+ ],
+ [
+ "▁",
+ "Ro"
+ ],
+ [
+ "R",
+ "A"
+ ],
+ [
+ "▁pract",
+ "ice"
+ ],
+ [
+ "▁g",
+ "raph"
+ ],
+ [
+ "▁gr",
+ "aph"
+ ],
+ [
+ "▁gra",
+ "ph"
+ ],
+ [
+ "▁grap",
+ "h"
+ ],
+ [
+ "▁",
+ "graph"
+ ],
+ [
+ "▁incre",
+ "ase"
+ ],
+ [
+ "▁fig",
+ "ure"
+ ],
+ [
+ "▁",
+ "figure"
+ ],
+ [
+ "Fil",
+ "ter"
+ ],
+ [
+ "HE",
+ "CK"
+ ],
+ [
+ "id",
+ "x"
+ ],
+ [
+ "i",
+ "dx"
+ ],
+ [
+ "▁g",
+ "lass"
+ ],
+ [
+ "▁gl",
+ "ass"
+ ],
+ [
+ "▁",
+ "glass"
+ ],
+ [
+ "sk",
+ "i"
+ ],
+ [
+ "s",
+ "ki"
+ ],
+ [
+ "com",
+ "es"
+ ],
+ [
+ "co",
+ "mes"
+ ],
+ [
+ "come",
+ "s"
+ ],
+ [
+ "c",
+ "omes"
+ ],
+ [
+ "▁c",
+ "at"
+ ],
+ [
+ "▁ca",
+ "t"
+ ],
+ [
+ "▁",
+ "cat"
+ ],
+ [
+ "▁c",
+ "old"
+ ],
+ [
+ "▁col",
+ "d"
+ ],
+ [
+ "▁co",
+ "ld"
+ ],
+ [
+ "go",
+ "to"
+ ],
+ [
+ "got",
+ "o"
+ ],
+ [
+ "g",
+ "oto"
+ ],
+ [
+ "uf",
+ "act"
+ ],
+ [
+ "u",
+ "fact"
+ ],
+ [
+ "▁C",
+ "opyright"
+ ],
+ [
+ "▁Copy",
+ "right"
+ ],
+ [
+ "▁",
+ "Copyright"
+ ],
+ [
+ "}}",
+ "\\"
+ ],
+ [
+ "}",
+ "}\\"
+ ],
+ [
+ "▁str",
+ "eng"
+ ],
+ [
+ "▁stre",
+ "ng"
+ ],
+ [
+ "▁d",
+ "ir"
+ ],
+ [
+ "▁di",
+ "r"
+ ],
+ [
+ "▁",
+ "dir"
+ ],
+ [
+ "to",
+ "ken"
+ ],
+ [
+ "tok",
+ "en"
+ ],
+ [
+ "t",
+ "oken"
+ ],
+ [
+ "▁occ",
+ "ur"
+ ],
+ [
+ "▁oc",
+ "cur"
+ ],
+ [
+ "arl",
+ "ier"
+ ],
+ [
+ "▁me",
+ "asure"
+ ],
+ [
+ "▁meas",
+ "ure"
+ ],
+ [
+ "▁",
+ "measure"
+ ],
+ [
+ "▁s",
+ "ec"
+ ],
+ [
+ "▁se",
+ "c"
+ ],
+ [
+ "▁",
+ "sec"
+ ],
+ [
+ "▁m",
+ "ás"
+ ],
+ [
+ "▁má",
+ "s"
+ ],
+ [
+ "▁N",
+ "et"
+ ],
+ [
+ "▁Ne",
+ "t"
+ ],
+ [
+ "▁",
+ "Net"
+ ],
+ [
+ "▁arg",
+ "ument"
+ ],
+ [
+ "▁",
+ "argument"
+ ],
+ [
+ "▁s",
+ "ou"
+ ],
+ [
+ "▁so",
+ "u"
+ ],
+ [
+ "▁m",
+ "oving"
+ ],
+ [
+ "▁mov",
+ "ing"
+ ],
+ [
+ "▁mo",
+ "ving"
+ ],
+ [
+ "▁p",
+ "refer"
+ ],
+ [
+ "▁pre",
+ "fer"
+ ],
+ [
+ "▁pref",
+ "er"
+ ],
+ [
+ "ma",
+ "sk"
+ ],
+ [
+ "mas",
+ "k"
+ ],
+ [
+ "m",
+ "ask"
+ ],
+ [
+ "<",
+ "<"
+ ],
+ [
+ "▁bre",
+ "ath"
+ ],
+ [
+ "▁breat",
+ "h"
+ ],
+ [
+ "▁phys",
+ "ical"
+ ],
+ [
+ "▁pos",
+ "itive"
+ ],
+ [
+ "▁posit",
+ "ive"
+ ],
+ [
+ "▁s",
+ "or"
+ ],
+ [
+ "▁so",
+ "r"
+ ],
+ [
+ "▁",
+ "sor"
+ ],
+ [
+ "▁de",
+ "part"
+ ],
+ [
+ "▁dep",
+ "art"
+ ],
+ [
+ "▁re",
+ "move"
+ ],
+ [
+ "▁rem",
+ "ove"
+ ],
+ [
+ "▁",
+ "remove"
+ ],
+ [
+ "▁k",
+ "it"
+ ],
+ [
+ "▁ki",
+ "t"
+ ],
+ [
+ "▁",
+ "kit"
+ ],
+ [
+ "▁me",
+ "eting"
+ ],
+ [
+ "▁meet",
+ "ing"
+ ],
+ [
+ "▁D",
+ "ata"
+ ],
+ [
+ "▁Da",
+ "ta"
+ ],
+ [
+ "▁Dat",
+ "a"
+ ],
+ [
+ "▁",
+ "Data"
+ ],
+ [
+ "og",
+ "raf"
+ ],
+ [
+ "act",
+ "ions"
+ ],
+ [
+ "action",
+ "s"
+ ],
+ [
+ "a",
+ "ctions"
+ ],
+ [
+ "▁param",
+ "eters"
+ ],
+ [
+ "▁parameter",
+ "s"
+ ],
+ [
+ "▁",
+ "parameters"
+ ],
+ [
+ "▁A",
+ "tt"
+ ],
+ [
+ "▁At",
+ "t"
+ ],
+ [
+ "▁",
+ "Att"
+ ],
+ [
+ "es",
+ "ch"
+ ],
+ [
+ "esc",
+ "h"
+ ],
+ [
+ "e",
+ "sch"
+ ],
+ [
+ "▁inv",
+ "olved"
+ ],
+ [
+ "▁invol",
+ "ved"
+ ],
+ [
+ "▁involve",
+ "d"
+ ],
+ [
+ "ä",
+ "t"
+ ],
+ [
+ "L",
+ "L"
+ ],
+ [
+ "B",
+ "ar"
+ ],
+ [
+ "▁с",
+ "и"
+ ],
+ [
+ "▁",
+ "си"
+ ],
+ [
+ "ec",
+ "h"
+ ],
+ [
+ "e",
+ "ch"
+ ],
+ [
+ "GE",
+ "T"
+ ],
+ [
+ "G",
+ "ET"
+ ],
+ [
+ "▁pre",
+ "vent"
+ ],
+ [
+ "▁pr",
+ "event"
+ ],
+ [
+ "▁prev",
+ "ent"
+ ],
+ [
+ "▁",
+ "prevent"
+ ],
+ [
+ "▁be",
+ "yond"
+ ],
+ [
+ "▁O",
+ "ther"
+ ],
+ [
+ "▁Ot",
+ "her"
+ ],
+ [
+ "▁",
+ "Other"
+ ],
+ [
+ "ä",
+ "n"
+ ],
+ [
+ "by",
+ "te"
+ ],
+ [
+ "▁sudd",
+ "en"
+ ],
+ [
+ "▁sud",
+ "den"
+ ],
+ [
+ "ol",
+ "ve"
+ ],
+ [
+ "olv",
+ "e"
+ ],
+ [
+ "▁н",
+ "о"
+ ],
+ [
+ "▁",
+ "но"
+ ],
+ [
+ "LO",
+ "G"
+ ],
+ [
+ "L",
+ "OG"
+ ],
+ [
+ "un",
+ "it"
+ ],
+ [
+ "uni",
+ "t"
+ ],
+ [
+ "u",
+ "nit"
+ ],
+ [
+ "▁tr",
+ "uth"
+ ],
+ [
+ "ra",
+ "t"
+ ],
+ [
+ "r",
+ "at"
+ ],
+ [
+ "S",
+ "D"
+ ],
+ [
+ "▁e",
+ "at"
+ ],
+ [
+ "▁M",
+ "ad"
+ ],
+ [
+ "▁Ma",
+ "d"
+ ],
+ [
+ "▁",
+ "Mad"
+ ],
+ [
+ "▁prov",
+ "ides"
+ ],
+ [
+ "▁provide",
+ "s"
+ ],
+ [
+ "▁s",
+ "ession"
+ ],
+ [
+ "▁",
+ "session"
+ ],
+ [
+ "De",
+ "le"
+ ],
+ [
+ "Del",
+ "e"
+ ],
+ [
+ "D",
+ "ele"
+ ],
+ [
+ "▁con",
+ "vers"
+ ],
+ [
+ "▁conv",
+ "ers"
+ ],
+ [
+ "▁conver",
+ "s"
+ ],
+ [
+ "▁conve",
+ "rs"
+ ],
+ [
+ "cent",
+ "er"
+ ],
+ [
+ "cen",
+ "ter"
+ ],
+ [
+ "c",
+ "enter"
+ ],
+ [
+ "▁contin",
+ "ued"
+ ],
+ [
+ "▁continue",
+ "d"
+ ],
+ [
+ "▁continu",
+ "ed"
+ ],
+ [
+ "ot",
+ "ion"
+ ],
+ [
+ "oti",
+ "on"
+ ],
+ [
+ "ca",
+ "che"
+ ],
+ [
+ "c",
+ "ache"
+ ],
+ [
+ "dis",
+ "play"
+ ],
+ [
+ "disp",
+ "lay"
+ ],
+ [
+ "▁prote",
+ "ct"
+ ],
+ [
+ "▁prot",
+ "ect"
+ ],
+ [
+ "am",
+ "s"
+ ],
+ [
+ "a",
+ "ms"
+ ],
+ [
+ "▁p",
+ "ow"
+ ],
+ [
+ "▁po",
+ "w"
+ ],
+ [
+ "▁",
+ "pow"
+ ],
+ [
+ "CT",
+ "ION"
+ ],
+ [
+ "C",
+ "TION"
+ ],
+ [
+ "▁M",
+ "ac"
+ ],
+ [
+ "▁Ma",
+ "c"
+ ],
+ [
+ "▁",
+ "Mac"
+ ],
+ [
+ "m",
+ "o"
+ ],
+ [
+ "х",
+ "а"
+ ],
+ [
+ "▁d",
+ "istance"
+ ],
+ [
+ "▁di",
+ "stance"
+ ],
+ [
+ "▁dist",
+ "ance"
+ ],
+ [
+ "▁",
+ "distance"
+ ],
+ [
+ "▁T",
+ "ime"
+ ],
+ [
+ "▁Tim",
+ "e"
+ ],
+ [
+ "▁Ti",
+ "me"
+ ],
+ [
+ "▁",
+ "Time"
+ ],
+ [
+ "g",
+ "i"
+ ],
+ [
+ "▁s",
+ "equ"
+ ],
+ [
+ "▁se",
+ "qu"
+ ],
+ [
+ "▁seq",
+ "u"
+ ],
+ [
+ "▁",
+ "sequ"
+ ],
+ [
+ "T",
+ "arget"
+ ],
+ [
+ "с",
+ "ле"
+ ],
+ [
+ "Ser",
+ "ver"
+ ],
+ [
+ "Serv",
+ "er"
+ ],
+ [
+ "▁w",
+ "ide"
+ ],
+ [
+ "▁wid",
+ "e"
+ ],
+ [
+ "▁",
+ "wide"
+ ],
+ [
+ "cl",
+ "ose"
+ ],
+ [
+ "clos",
+ "e"
+ ],
+ [
+ "▁c",
+ "ru"
+ ],
+ [
+ "▁cr",
+ "u"
+ ],
+ [
+ "Ex",
+ "t"
+ ],
+ [
+ "E",
+ "xt"
+ ],
+ [
+ "▁s",
+ "elect"
+ ],
+ [
+ "▁se",
+ "lect"
+ ],
+ [
+ "▁sel",
+ "ect"
+ ],
+ [
+ "▁sele",
+ "ct"
+ ],
+ [
+ "▁",
+ "select"
+ ],
+ [
+ "▁pat",
+ "tern"
+ ],
+ [
+ "▁",
+ "pattern"
+ ],
+ [
+ "\")",
+ ");"
+ ],
+ [
+ "\"))",
+ ";"
+ ],
+ [
+ "\"",
+ "));"
+ ],
+ [
+ "Pro",
+ "vider"
+ ],
+ [
+ "Prov",
+ "ider"
+ ],
+ [
+ "UR",
+ "L"
+ ],
+ [
+ "U",
+ "RL"
+ ],
+ [
+ "▁g",
+ "reen"
+ ],
+ [
+ "▁gr",
+ "een"
+ ],
+ [
+ "▁gre",
+ "en"
+ ],
+ [
+ "▁",
+ "green"
+ ],
+ [
+ "▁wait",
+ "ing"
+ ],
+ [
+ "▁wa",
+ "iting"
+ ],
+ [
+ "pro",
+ "to"
+ ],
+ [
+ "pr",
+ "oto"
+ ],
+ [
+ "prot",
+ "o"
+ ],
+ [
+ "▁immedi",
+ "ately"
+ ],
+ [
+ "▁immediate",
+ "ly"
+ ],
+ [
+ "com",
+ "mon"
+ ],
+ [
+ "comm",
+ "on"
+ ],
+ [
+ "az",
+ "ione"
+ ],
+ [
+ "azi",
+ "one"
+ ],
+ [
+ "a",
+ "zione"
+ ],
+ [
+ "ri",
+ "ver"
+ ],
+ [
+ "riv",
+ "er"
+ ],
+ [
+ "rive",
+ "r"
+ ],
+ [
+ "r",
+ "iver"
+ ],
+ [
+ "▁s",
+ "en"
+ ],
+ [
+ "▁se",
+ "n"
+ ],
+ [
+ "▁",
+ "sen"
+ ],
+ [
+ "▁!",
+ "=="
+ ],
+ [
+ "▁!=",
+ "="
+ ],
+ [
+ "▁Febru",
+ "ary"
+ ],
+ [
+ "▁Februar",
+ "y"
+ ],
+ [
+ "ur",
+ "b"
+ ],
+ [
+ "u",
+ "rb"
+ ],
+ [
+ "▁S",
+ "en"
+ ],
+ [
+ "▁Se",
+ "n"
+ ],
+ [
+ "de",
+ "st"
+ ],
+ [
+ "des",
+ "t"
+ ],
+ [
+ "d",
+ "est"
+ ],
+ [
+ "<",
+ "?"
+ ],
+ [
+ "▁ed",
+ "ge"
+ ],
+ [
+ "▁",
+ "edge"
+ ],
+ [
+ "▁m",
+ "ais"
+ ],
+ [
+ "▁ma",
+ "is"
+ ],
+ [
+ "▁mai",
+ "s"
+ ],
+ [
+ "gor",
+ "ith"
+ ],
+ [
+ "cp",
+ "u"
+ ],
+ [
+ "c",
+ "pu"
+ ],
+ [
+ "▁educ",
+ "ation"
+ ],
+ [
+ "▁associ",
+ "ated"
+ ],
+ [
+ "▁associate",
+ "d"
+ ],
+ [
+ "No",
+ "ne"
+ ],
+ [
+ "Non",
+ "e"
+ ],
+ [
+ "N",
+ "one"
+ ],
+ [
+ "h",
+ "i"
+ ],
+ [
+ "▁p",
+ "oor"
+ ],
+ [
+ "▁po",
+ "or"
+ ],
+ [
+ "se",
+ "m"
+ ],
+ [
+ "s",
+ "em"
+ ],
+ [
+ "▁W",
+ "il"
+ ],
+ [
+ "▁Wi",
+ "l"
+ ],
+ [
+ "▁b",
+ "ud"
+ ],
+ [
+ "▁bu",
+ "d"
+ ],
+ [
+ "▁",
+ "bud"
+ ],
+ [
+ "▁a",
+ "uch"
+ ],
+ [
+ "▁au",
+ "ch"
+ ],
+ [
+ "▁",
+ "auch"
+ ],
+ [
+ "el",
+ "ler"
+ ],
+ [
+ "ell",
+ "er"
+ ],
+ [
+ "elle",
+ "r"
+ ],
+ [
+ "▁L",
+ "ife"
+ ],
+ [
+ "▁Li",
+ "fe"
+ ],
+ [
+ "▁",
+ "Life"
+ ],
+ [
+ "▁f",
+ "iles"
+ ],
+ [
+ "▁fil",
+ "es"
+ ],
+ [
+ "▁file",
+ "s"
+ ],
+ [
+ "▁fi",
+ "les"
+ ],
+ [
+ "▁",
+ "files"
+ ],
+ [
+ "▁le",
+ "ading"
+ ],
+ [
+ "▁lead",
+ "ing"
+ ],
+ [
+ "▁",
+ "leading"
+ ],
+ [
+ "▁ob",
+ "tain"
+ ],
+ [
+ "▁obt",
+ "ain"
+ ],
+ [
+ "▁J",
+ "ul"
+ ],
+ [
+ "▁Ju",
+ "l"
+ ],
+ [
+ "at",
+ "ory"
+ ],
+ [
+ "ator",
+ "y"
+ ],
+ [
+ "ato",
+ "ry"
+ ],
+ [
+ "г",
+ "у"
+ ],
+ [
+ "it",
+ "able"
+ ],
+ [
+ "ita",
+ "ble"
+ ],
+ [
+ "i",
+ "table"
+ ],
+ [
+ "▁on",
+ "to"
+ ],
+ [
+ "▁ont",
+ "o"
+ ],
+ [
+ "▁",
+ "onto"
+ ],
+ [
+ "▁b",
+ "orn"
+ ],
+ [
+ "▁bo",
+ "rn"
+ ],
+ [
+ "▁bor",
+ "n"
+ ],
+ [
+ "▁",
+ "born"
+ ],
+ [
+ "or",
+ "em"
+ ],
+ [
+ "ore",
+ "m"
+ ],
+ [
+ "o",
+ "rem"
+ ],
+ [
+ "▁Stre",
+ "et"
+ ],
+ [
+ "▁m",
+ "aint"
+ ],
+ [
+ "▁main",
+ "t"
+ ],
+ [
+ "▁ma",
+ "int"
+ ],
+ [
+ "▁mai",
+ "nt"
+ ],
+ [
+ "Param",
+ "s"
+ ],
+ [
+ "Par",
+ "ams"
+ ],
+ [
+ "ri",
+ "p"
+ ],
+ [
+ "r",
+ "ip"
+ ],
+ [
+ "▁S",
+ "T"
+ ],
+ [
+ "▁",
+ "ST"
+ ],
+ [
+ "u",
+ "v"
+ ],
+ [
+ "ma",
+ "in"
+ ],
+ [
+ "m",
+ "ain"
+ ],
+ [
+ "▁re",
+ "cent"
+ ],
+ [
+ "▁rec",
+ "ent"
+ ],
+ [
+ "▁rece",
+ "nt"
+ ],
+ [
+ "We",
+ "b"
+ ],
+ [
+ "W",
+ "eb"
+ ],
+ [
+ "ov",
+ "a"
+ ],
+ [
+ "o",
+ "va"
+ ],
+ [
+ "ц",
+ "а"
+ ],
+ [
+ "ais",
+ "e"
+ ],
+ [
+ "ai",
+ "se"
+ ],
+ [
+ "a",
+ "ise"
+ ],
+ [
+ "yle",
+ "s"
+ ],
+ [
+ "yl",
+ "es"
+ ],
+ [
+ "y",
+ "les"
+ ],
+ [
+ "▁de",
+ "scribed"
+ ],
+ [
+ "▁desc",
+ "ribed"
+ ],
+ [
+ "▁describ",
+ "ed"
+ ],
+ [
+ "▁describe",
+ "d"
+ ],
+ [
+ "▁begin",
+ "ning"
+ ],
+ [
+ "▁D",
+ "ay"
+ ],
+ [
+ "▁Da",
+ "y"
+ ],
+ [
+ "▁",
+ "Day"
+ ],
+ [
+ "▁V",
+ "ol"
+ ],
+ [
+ "▁Vo",
+ "l"
+ ],
+ [
+ "▁",
+ "Vol"
+ ],
+ [
+ "▁h",
+ "uge"
+ ],
+ [
+ "▁hug",
+ "e"
+ ],
+ [
+ "Ha",
+ "s"
+ ],
+ [
+ "H",
+ "as"
+ ],
+ [
+ "an",
+ "cy"
+ ],
+ [
+ "anc",
+ "y"
+ ],
+ [
+ "He",
+ "ader"
+ ],
+ [
+ "Head",
+ "er"
+ ],
+ [
+ "▁a",
+ "ren"
+ ],
+ [
+ "▁are",
+ "n"
+ ],
+ [
+ "▁ar",
+ "en"
+ ],
+ [
+ "▁",
+ "aren"
+ ],
+ [
+ "ва",
+ "н"
+ ],
+ [
+ "в",
+ "ан"
+ ],
+ [
+ "▁en",
+ "sure"
+ ],
+ [
+ "▁ens",
+ "ure"
+ ],
+ [
+ "▁",
+ "ensure"
+ ],
+ [
+ "▁p",
+ "et"
+ ],
+ [
+ "▁pe",
+ "t"
+ ],
+ [
+ "▁",
+ "pet"
+ ],
+ [
+ "mu",
+ "lt"
+ ],
+ [
+ "mul",
+ "t"
+ ],
+ [
+ "m",
+ "ult"
+ ],
+ [
+ "▁L",
+ "ike"
+ ],
+ [
+ "▁Li",
+ "ke"
+ ],
+ [
+ "▁",
+ "Like"
+ ],
+ [
+ "▁man",
+ "agement"
+ ],
+ [
+ "▁manage",
+ "ment"
+ ],
+ [
+ "▁",
+ "management"
+ ],
+ [
+ "P",
+ "S"
+ ],
+ [
+ "wh",
+ "ile"
+ ],
+ [
+ "▁back",
+ "ground"
+ ],
+ [
+ "▁",
+ "background"
+ ],
+ [
+ "ount",
+ "er"
+ ],
+ [
+ "oun",
+ "ter"
+ ],
+ [
+ "o",
+ "unter"
+ ],
+ [
+ "bo",
+ "ol"
+ ],
+ [
+ "b",
+ "ool"
+ ],
+ [
+ "F",
+ "C"
+ ],
+ [
+ "N",
+ "um"
+ ],
+ [
+ "R",
+ "L"
+ ],
+ [
+ "▁ex",
+ "cl"
+ ],
+ [
+ "▁exc",
+ "l"
+ ],
+ [
+ "▁e",
+ "ye"
+ ],
+ [
+ "▁ey",
+ "e"
+ ],
+ [
+ "im",
+ "g"
+ ],
+ [
+ "i",
+ "mg"
+ ],
+ [
+ "▁r",
+ "om"
+ ],
+ [
+ "▁ro",
+ "m"
+ ],
+ [
+ "▁",
+ "rom"
+ ],
+ [
+ "▁H",
+ "el"
+ ],
+ [
+ "▁He",
+ "l"
+ ],
+ [
+ "▁",
+ "Hel"
+ ],
+ [
+ "Opt",
+ "ion"
+ ],
+ [
+ "O",
+ "ption"
+ ],
+ [
+ "▁stop",
+ "ped"
+ ],
+ [
+ "▁sto",
+ "pped"
+ ],
+ [
+ "▁th",
+ "read"
+ ],
+ [
+ "▁thr",
+ "ead"
+ ],
+ [
+ "▁",
+ "thread"
+ ],
+ [
+ "to",
+ "type"
+ ],
+ [
+ "tot",
+ "ype"
+ ],
+ [
+ "t",
+ "otype"
+ ],
+ [
+ "))",
+ ")"
+ ],
+ [
+ ")",
+ "))"
+ ],
+ [
+ "▁st",
+ "age"
+ ],
+ [
+ "▁stag",
+ "e"
+ ],
+ [
+ "▁sta",
+ "ge"
+ ],
+ [
+ "▁",
+ "stage"
+ ],
+ [
+ "▁ü",
+ "ber"
+ ],
+ [
+ "▁",
+ "über"
+ ],
+ [
+ "▁al",
+ "though"
+ ],
+ [
+ "▁",
+ "although"
+ ],
+ [
+ "Type",
+ "s"
+ ],
+ [
+ "Ty",
+ "pes"
+ ],
+ [
+ "Typ",
+ "es"
+ ],
+ [
+ "T",
+ "ypes"
+ ],
+ [
+ "▁O",
+ "h"
+ ],
+ [
+ "▁",
+ "Oh"
+ ],
+ [
+ "▁e",
+ "ight"
+ ],
+ [
+ "▁",
+ "eight"
+ ],
+ [
+ "▁de",
+ "scription"
+ ],
+ [
+ "▁des",
+ "cription"
+ ],
+ [
+ "▁",
+ "description"
+ ],
+ [
+ "'",
+ "'"
+ ],
+ [
+ "ö",
+ "n"
+ ],
+ [
+ "▁sur",
+ "face"
+ ],
+ [
+ "▁surf",
+ "ace"
+ ],
+ [
+ "▁",
+ "surface"
+ ],
+ [
+ "▁Intern",
+ "ational"
+ ],
+ [
+ "▁ch",
+ "arg"
+ ],
+ [
+ "▁char",
+ "g"
+ ],
+ [
+ "▁cha",
+ "rg"
+ ],
+ [
+ "▁",
+ "charg"
+ ],
+ [
+ "▁col",
+ "lection"
+ ],
+ [
+ "▁coll",
+ "ection"
+ ],
+ [
+ "▁collect",
+ "ion"
+ ],
+ [
+ "▁colle",
+ "ction"
+ ],
+ [
+ "▁",
+ "collection"
+ ],
+ [
+ "▁us",
+ "ers"
+ ],
+ [
+ "▁use",
+ "rs"
+ ],
+ [
+ "▁user",
+ "s"
+ ],
+ [
+ "▁",
+ "users"
+ ],
+ [
+ "▁ob",
+ "vious"
+ ],
+ [
+ "▁cent",
+ "ury"
+ ],
+ [
+ "▁",
+ "century"
+ ],
+ [
+ "ic",
+ "ks"
+ ],
+ [
+ "ick",
+ "s"
+ ],
+ [
+ "i",
+ "cks"
+ ],
+ [
+ "▁art",
+ "icle"
+ ],
+ [
+ "▁artic",
+ "le"
+ ],
+ [
+ "▁",
+ "article"
+ ],
+ [
+ "▁\"",
+ "\\"
+ ],
+ [
+ "▁",
+ "\"\\"
+ ],
+ [
+ "di",
+ "m"
+ ],
+ [
+ "d",
+ "im"
+ ],
+ [
+ "▁s",
+ "in"
+ ],
+ [
+ "▁si",
+ "n"
+ ],
+ [
+ "▁",
+ "sin"
+ ],
+ [
+ "en",
+ "ge"
+ ],
+ [
+ "eng",
+ "e"
+ ],
+ [
+ "Cont",
+ "rol"
+ ],
+ [
+ "▁com",
+ "mit"
+ ],
+ [
+ "▁comm",
+ "it"
+ ],
+ [
+ "▁",
+ "commit"
+ ],
+ [
+ "ens",
+ "ity"
+ ],
+ [
+ "▁t",
+ "ra"
+ ],
+ [
+ "▁tr",
+ "a"
+ ],
+ [
+ "▁",
+ "tra"
+ ],
+ [
+ "cript",
+ "or"
+ ],
+ [
+ "▁N",
+ "OT"
+ ],
+ [
+ "▁NO",
+ "T"
+ ],
+ [
+ "▁",
+ "NOT"
+ ],
+ [
+ "we",
+ "ll"
+ ],
+ [
+ "w",
+ "ell"
+ ],
+ [
+ "▁M",
+ "ichael"
+ ],
+ [
+ "▁Mich",
+ "ael"
+ ],
+ [
+ "▁n",
+ "od"
+ ],
+ [
+ "▁no",
+ "d"
+ ],
+ [
+ "▁",
+ "nod"
+ ],
+ [
+ "▁m",
+ "ort"
+ ],
+ [
+ "▁mor",
+ "t"
+ ],
+ [
+ "▁mo",
+ "rt"
+ ],
+ [
+ "iv",
+ "o"
+ ],
+ [
+ "i",
+ "vo"
+ ],
+ [
+ "is",
+ "ation"
+ ],
+ [
+ "▁P",
+ "o"
+ ],
+ [
+ "▁",
+ "Po"
+ ],
+ [
+ "▁P",
+ "aris"
+ ],
+ [
+ "▁Par",
+ "is"
+ ],
+ [
+ "▁Pa",
+ "ris"
+ ],
+ [
+ "▁ad",
+ "ministr"
+ ],
+ [
+ "▁admin",
+ "istr"
+ ],
+ [
+ "▁",
+ "administr"
+ ],
+ [
+ "bu",
+ "rg"
+ ],
+ [
+ "bur",
+ "g"
+ ],
+ [
+ "b",
+ "urg"
+ ],
+ [
+ "cd",
+ "ot"
+ ],
+ [
+ "c",
+ "dot"
+ ],
+ [
+ "▁mil",
+ "itary"
+ ],
+ [
+ "▁milit",
+ "ary"
+ ],
+ [
+ "▁militar",
+ "y"
+ ],
+ [
+ "▁B",
+ "est"
+ ],
+ [
+ "▁Be",
+ "st"
+ ],
+ [
+ "▁Bes",
+ "t"
+ ],
+ [
+ "▁",
+ "Best"
+ ],
+ [
+ "▁К",
+ "а"
+ ],
+ [
+ "▁",
+ "Ка"
+ ],
+ [
+ "IN",
+ "E"
+ ],
+ [
+ "I",
+ "NE"
+ ],
+ [
+ "▁through",
+ "out"
+ ],
+ [
+ "S",
+ "l"
+ ],
+ [
+ "▁im",
+ "pl"
+ ],
+ [
+ "▁imp",
+ "l"
+ ],
+ [
+ "▁",
+ "impl"
+ ],
+ [
+ "cont",
+ "rol"
+ ],
+ [
+ "contr",
+ "ol"
+ ],
+ [
+ "▁",
+ "Ч"
+ ],
+ [
+ "▁u",
+ "it"
+ ],
+ [
+ "▁ui",
+ "t"
+ ],
+ [
+ "▁",
+ "uit"
+ ],
+ [
+ "▁un",
+ "signed"
+ ],
+ [
+ "▁uns",
+ "igned"
+ ],
+ [
+ "▁",
+ "unsigned"
+ ],
+ [
+ "▁M",
+ "ary"
+ ],
+ [
+ "▁Mar",
+ "y"
+ ],
+ [
+ "▁Ma",
+ "ry"
+ ],
+ [
+ "Ch",
+ "ar"
+ ],
+ [
+ "C",
+ "har"
+ ],
+ [
+ "м",
+ "і"
+ ],
+ [
+ "▁th",
+ "reat"
+ ],
+ [
+ "▁c",
+ "ourt"
+ ],
+ [
+ "▁co",
+ "urt"
+ ],
+ [
+ "▁cour",
+ "t"
+ ],
+ [
+ "▁cou",
+ "rt"
+ ],
+ [
+ "▁",
+ "court"
+ ],
+ [
+ "vi",
+ "lle"
+ ],
+ [
+ "vil",
+ "le"
+ ],
+ [
+ "v",
+ "ille"
+ ],
+ [
+ "▁",
+ "ш"
+ ],
+ [
+ "▁C",
+ "am"
+ ],
+ [
+ "▁Ca",
+ "m"
+ ],
+ [
+ "▁",
+ "Cam"
+ ],
+ [
+ ".",
+ "\r"
+ ],
+ [
+ "▁current",
+ "ly"
+ ],
+ [
+ "▁curr",
+ "ently"
+ ],
+ [
+ "ro",
+ "t"
+ ],
+ [
+ "r",
+ "ot"
+ ],
+ [
+ "▁D",
+ "ate"
+ ],
+ [
+ "▁Da",
+ "te"
+ ],
+ [
+ "▁Dat",
+ "e"
+ ],
+ [
+ "▁",
+ "Date"
+ ],
+ [
+ "▁s",
+ "hit"
+ ],
+ [
+ "▁sh",
+ "it"
+ ],
+ [
+ "▁",
+ "shit"
+ ],
+ [
+ "▁$",
+ "{\\"
+ ],
+ [
+ "▁${",
+ "\\"
+ ],
+ [
+ "un",
+ "n"
+ ],
+ [
+ "u",
+ "nn"
+ ],
+ [
+ "U",
+ "s"
+ ],
+ [
+ "▁b",
+ "uffer"
+ ],
+ [
+ "▁buff",
+ "er"
+ ],
+ [
+ "▁buf",
+ "fer"
+ ],
+ [
+ "▁",
+ "buffer"
+ ],
+ [
+ "▁s",
+ "ont"
+ ],
+ [
+ "▁so",
+ "nt"
+ ],
+ [
+ "▁son",
+ "t"
+ ],
+ [
+ "▁let",
+ "ter"
+ ],
+ [
+ "▁lett",
+ "er"
+ ],
+ [
+ "▁",
+ "letter"
+ ],
+ [
+ "in",
+ "ated"
+ ],
+ [
+ "ina",
+ "ted"
+ ],
+ [
+ "inate",
+ "d"
+ ],
+ [
+ "Ch",
+ "ange"
+ ],
+ [
+ "▁h",
+ "ref"
+ ],
+ [
+ "▁hr",
+ "ef"
+ ],
+ [
+ "▁",
+ "href"
+ ],
+ [
+ "▁l",
+ "ack"
+ ],
+ [
+ "▁la",
+ "ck"
+ ],
+ [
+ "▁lac",
+ "k"
+ ],
+ [
+ "▁o",
+ "il"
+ ],
+ [
+ "▁C",
+ "ons"
+ ],
+ [
+ "▁Con",
+ "s"
+ ],
+ [
+ "▁Co",
+ "ns"
+ ],
+ [
+ "▁",
+ "Cons"
+ ],
+ [
+ "▁J",
+ "er"
+ ],
+ [
+ "▁Je",
+ "r"
+ ],
+ [
+ "BU",
+ "G"
+ ],
+ [
+ "B",
+ "UG"
+ ],
+ [
+ "if",
+ "orn"
+ ],
+ [
+ "▁pro",
+ "perties"
+ ],
+ [
+ "▁proper",
+ "ties"
+ ],
+ [
+ "▁",
+ "properties"
+ ],
+ [
+ "▁r",
+ "andom"
+ ],
+ [
+ "▁ran",
+ "dom"
+ ],
+ [
+ "▁rand",
+ "om"
+ ],
+ [
+ "▁",
+ "random"
+ ],
+ [
+ "▁br",
+ "other"
+ ],
+ [
+ "▁bro",
+ "ther"
+ ],
+ [
+ "▁p",
+ "iece"
+ ],
+ [
+ "▁pie",
+ "ce"
+ ],
+ [
+ "▁",
+ "piece"
+ ],
+ [
+ "б",
+ "у"
+ ],
+ [
+ "ist",
+ "ics"
+ ],
+ [
+ "istic",
+ "s"
+ ],
+ [
+ "isti",
+ "cs"
+ ],
+ [
+ "▁techn",
+ "ology"
+ ],
+ [
+ "gl",
+ "obal"
+ ],
+ [
+ "glob",
+ "al"
+ ],
+ [
+ "▁trans",
+ "form"
+ ],
+ [
+ "▁",
+ "transform"
+ ],
+ [
+ "er",
+ "d"
+ ],
+ [
+ "e",
+ "rd"
+ ],
+ [
+ "▁B",
+ "ecause"
+ ],
+ [
+ "▁",
+ "Because"
+ ],
+ [
+ "PE",
+ "CT"
+ ],
+ [
+ "P",
+ "ECT"
+ ],
+ [
+ "pr",
+ "et"
+ ],
+ [
+ "pre",
+ "t"
+ ],
+ [
+ "p",
+ "ret"
+ ],
+ [
+ "▁го",
+ "ду"
+ ],
+ [
+ "▁год",
+ "у"
+ ],
+ [
+ "▁M",
+ "et"
+ ],
+ [
+ "▁Me",
+ "t"
+ ],
+ [
+ "▁",
+ "Met"
+ ],
+ [
+ "▁p",
+ "sy"
+ ],
+ [
+ "▁ps",
+ "y"
+ ],
+ [
+ "▁",
+ "psy"
+ ],
+ [
+ "▁о",
+ "д"
+ ],
+ [
+ "▁g",
+ "od"
+ ],
+ [
+ "▁go",
+ "d"
+ ],
+ [
+ "▁",
+ "god"
+ ],
+ [
+ "▁D",
+ "el"
+ ],
+ [
+ "▁De",
+ "l"
+ ],
+ [
+ "▁",
+ "Del"
+ ],
+ [
+ "base",
+ "d"
+ ],
+ [
+ "ba",
+ "sed"
+ ],
+ [
+ "bas",
+ "ed"
+ ],
+ [
+ "b",
+ "ased"
+ ],
+ [
+ "▁v",
+ "oor"
+ ],
+ [
+ "▁vo",
+ "or"
+ ],
+ [
+ "▁C",
+ "all"
+ ],
+ [
+ "▁Cal",
+ "l"
+ ],
+ [
+ "▁Ca",
+ "ll"
+ ],
+ [
+ "▁",
+ "Call"
+ ],
+ [
+ "S",
+ "A"
+ ],
+ [
+ "▁fil",
+ "ter"
+ ],
+ [
+ "▁",
+ "filter"
+ ],
+ [
+ "▁incl",
+ "udes"
+ ],
+ [
+ "▁includ",
+ "es"
+ ],
+ [
+ "▁include",
+ "s"
+ ],
+ [
+ "▁inclu",
+ "des"
+ ],
+ [
+ "▁",
+ "includes"
+ ],
+ [
+ "olut",
+ "ions"
+ ],
+ [
+ "olution",
+ "s"
+ ],
+ [
+ "f",
+ "d"
+ ],
+ [
+ "▁w",
+ "ind"
+ ],
+ [
+ "▁win",
+ "d"
+ ],
+ [
+ "▁",
+ "wind"
+ ],
+ [
+ "▁б",
+ "о"
+ ],
+ [
+ "▁",
+ "бо"
+ ],
+ [
+ "▁ab",
+ "ility"
+ ],
+ [
+ "▁",
+ "ability"
+ ],
+ [
+ "ca",
+ "rd"
+ ],
+ [
+ "car",
+ "d"
+ ],
+ [
+ "c",
+ "ard"
+ ],
+ [
+ "▁n",
+ "umer"
+ ],
+ [
+ "▁num",
+ "er"
+ ],
+ [
+ "▁nu",
+ "mer"
+ ],
+ [
+ "▁",
+ "numer"
+ ],
+ [
+ "add",
+ "ress"
+ ],
+ [
+ "addr",
+ "ess"
+ ],
+ [
+ "▁go",
+ "al"
+ ],
+ [
+ "ash",
+ "ington"
+ ],
+ [
+ "ashing",
+ "ton"
+ ],
+ [
+ "▁s",
+ "light"
+ ],
+ [
+ "▁sl",
+ "ight"
+ ],
+ [
+ "ab",
+ "a"
+ ],
+ [
+ "a",
+ "ba"
+ ],
+ [
+ "▁L",
+ "og"
+ ],
+ [
+ "▁Lo",
+ "g"
+ ],
+ [
+ "▁",
+ "Log"
+ ],
+ [
+ "Set",
+ "tings"
+ ],
+ [
+ "Setting",
+ "s"
+ ],
+ [
+ "ad",
+ "ow"
+ ],
+ [
+ "ado",
+ "w"
+ ],
+ [
+ "▁p",
+ "i"
+ ],
+ [
+ "▁",
+ "pi"
+ ],
+ [
+ "ir",
+ "ing"
+ ],
+ [
+ "iri",
+ "ng"
+ ],
+ [
+ "i",
+ "ring"
+ ],
+ [
+ "F",
+ "T"
+ ],
+ [
+ "▁number",
+ "s"
+ ],
+ [
+ "▁num",
+ "bers"
+ ],
+ [
+ "con",
+ "f"
+ ],
+ [
+ "co",
+ "nf"
+ ],
+ [
+ "ta",
+ "sk"
+ ],
+ [
+ "t",
+ "ask"
+ ],
+ [
+ "▁î",
+ "n"
+ ],
+ [
+ "т",
+ "ы"
+ ],
+ [
+ "▁re",
+ "ceive"
+ ],
+ [
+ "▁rece",
+ "ive"
+ ],
+ [
+ "▁r",
+ "oot"
+ ],
+ [
+ "▁ro",
+ "ot"
+ ],
+ [
+ "▁",
+ "root"
+ ],
+ [
+ "▁Ind",
+ "ia"
+ ],
+ [
+ "pat",
+ "ch"
+ ],
+ [
+ "p",
+ "atch"
+ ],
+ [
+ "é",
+ "l"
+ ],
+ [
+ "▁sum",
+ "mer"
+ ],
+ [
+ "▁method",
+ "s"
+ ],
+ [
+ "▁",
+ "methods"
+ ],
+ [
+ "▁pl",
+ "aces"
+ ],
+ [
+ "▁place",
+ "s"
+ ],
+ [
+ "▁plac",
+ "es"
+ ],
+ [
+ "▁М",
+ "а"
+ ],
+ [
+ "▁",
+ "Ма"
+ ],
+ [
+ "▁cap",
+ "ital"
+ ],
+ [
+ "▁capit",
+ "al"
+ ],
+ [
+ "▁ev",
+ "idence"
+ ],
+ [
+ "▁G",
+ "erman"
+ ],
+ [
+ "▁Germ",
+ "an"
+ ],
+ [
+ "▁Ger",
+ "man"
+ ],
+ [
+ "\\",
+ ","
+ ],
+ [
+ "D",
+ "A"
+ ],
+ [
+ "ec",
+ "ute"
+ ],
+ [
+ "ecut",
+ "e"
+ ],
+ [
+ "col",
+ "umn"
+ ],
+ [
+ "▁fun",
+ "ctions"
+ ],
+ [
+ "▁function",
+ "s"
+ ],
+ [
+ "▁",
+ "functions"
+ ],
+ [
+ "▁c",
+ "ounter"
+ ],
+ [
+ "▁co",
+ "unter"
+ ],
+ [
+ "▁coun",
+ "ter"
+ ],
+ [
+ "▁count",
+ "er"
+ ],
+ [
+ "▁",
+ "counter"
+ ],
+ [
+ "▁ar",
+ "ms"
+ ],
+ [
+ "▁arm",
+ "s"
+ ],
+ [
+ "▁",
+ "arms"
+ ],
+ [
+ "▁f",
+ "eed"
+ ],
+ [
+ "▁fe",
+ "ed"
+ ],
+ [
+ "▁fee",
+ "d"
+ ],
+ [
+ "▁",
+ "feed"
+ ],
+ [
+ "ve",
+ "y"
+ ],
+ [
+ "v",
+ "ey"
+ ],
+ [
+ "he",
+ "nt"
+ ],
+ [
+ "hen",
+ "t"
+ ],
+ [
+ "h",
+ "ent"
+ ],
+ [
+ "MA",
+ "X"
+ ],
+ [
+ "M",
+ "AX"
+ ],
+ [
+ "▁ac",
+ "qu"
+ ],
+ [
+ "▁app",
+ "ly"
+ ],
+ [
+ "▁ap",
+ "ply"
+ ],
+ [
+ "▁appl",
+ "y"
+ ],
+ [
+ "▁",
+ "apply"
+ ],
+ [
+ "▁hus",
+ "band"
+ ],
+ [
+ "▁k",
+ "illed"
+ ],
+ [
+ "▁kill",
+ "ed"
+ ],
+ [
+ "▁kil",
+ "led"
+ ],
+ [
+ "▁S",
+ "pec"
+ ],
+ [
+ "▁Sp",
+ "ec"
+ ],
+ [
+ "▁Spe",
+ "c"
+ ],
+ [
+ "▁",
+ "Spec"
+ ],
+ [
+ "ent",
+ "ity"
+ ],
+ [
+ "enti",
+ "ty"
+ ],
+ [
+ "▁e",
+ "arlier"
+ ],
+ [
+ "▁M",
+ "iss"
+ ],
+ [
+ "▁Mi",
+ "ss"
+ ],
+ [
+ "▁Mis",
+ "s"
+ ],
+ [
+ "▁",
+ "Miss"
+ ],
+ [
+ "▁set",
+ "ting"
+ ],
+ [
+ "▁sett",
+ "ing"
+ ],
+ [
+ "▁",
+ "setting"
+ ],
+ [
+ "it",
+ "ect"
+ ],
+ [
+ "ite",
+ "ct"
+ ],
+ [
+ "▁d",
+ "ed"
+ ],
+ [
+ "▁de",
+ "d"
+ ],
+ [
+ "▁",
+ "ded"
+ ],
+ [
+ "Ro",
+ "w"
+ ],
+ [
+ "R",
+ "ow"
+ ],
+ [
+ "▁r",
+ "an"
+ ],
+ [
+ "▁ra",
+ "n"
+ ],
+ [
+ "▁",
+ "ran"
+ ],
+ [
+ "▁Y",
+ "es"
+ ],
+ [
+ "▁Ye",
+ "s"
+ ],
+ [
+ "▁",
+ "Yes"
+ ],
+ [
+ "▁fin",
+ "ancial"
+ ],
+ [
+ "▁financ",
+ "ial"
+ ],
+ [
+ "s",
+ "ession"
+ ],
+ [
+ "le",
+ "ar"
+ ],
+ [
+ "l",
+ "ear"
+ ],
+ [
+ "is",
+ "hing"
+ ],
+ [
+ "ish",
+ "ing"
+ ],
+ [
+ "ishi",
+ "ng"
+ ],
+ [
+ "▁ne",
+ "arly"
+ ],
+ [
+ "▁near",
+ "ly"
+ ],
+ [
+ "▁d",
+ "ur"
+ ],
+ [
+ "▁du",
+ "r"
+ ],
+ [
+ "▁m",
+ "achine"
+ ],
+ [
+ "▁mach",
+ "ine"
+ ],
+ [
+ "▁",
+ "machine"
+ ],
+ [
+ "xf",
+ "f"
+ ],
+ [
+ "x",
+ "ff"
+ ],
+ [
+ "br",
+ "o"
+ ],
+ [
+ "b",
+ "ro"
+ ],
+ [
+ "▁s",
+ "ymbol"
+ ],
+ [
+ "▁sym",
+ "bol"
+ ],
+ [
+ "▁",
+ "symbol"
+ ],
+ [
+ "land",
+ "s"
+ ],
+ [
+ "lan",
+ "ds"
+ ],
+ [
+ "l",
+ "ands"
+ ],
+ [
+ "Ac",
+ "c"
+ ],
+ [
+ "A",
+ "cc"
+ ],
+ [
+ "d",
+ "i"
+ ],
+ [
+ "▁Rober",
+ "t"
+ ],
+ [
+ "▁Ro",
+ "bert"
+ ],
+ [
+ "▁Rob",
+ "ert"
+ ],
+ [
+ "pro",
+ "p"
+ ],
+ [
+ "pr",
+ "op"
+ ],
+ [
+ "p",
+ "rop"
+ ],
+ [
+ "ur",
+ "ity"
+ ],
+ [
+ "uri",
+ "ty"
+ ],
+ [
+ "▁#",
+ "####"
+ ],
+ [
+ "▁##",
+ "###"
+ ],
+ [
+ "▁###",
+ "##"
+ ],
+ [
+ "▁####",
+ "#"
+ ],
+ [
+ "▁walk",
+ "ed"
+ ],
+ [
+ "▁wal",
+ "ked"
+ ],
+ [
+ "▁intern",
+ "ational"
+ ],
+ [
+ "▁internation",
+ "al"
+ ],
+ [
+ "▁",
+ "Е"
+ ],
+ [
+ "Y",
+ "es"
+ ],
+ [
+ "▁re",
+ "lease"
+ ],
+ [
+ "▁rele",
+ "ase"
+ ],
+ [
+ "▁",
+ "release"
+ ],
+ [
+ "▁start",
+ "ing"
+ ],
+ [
+ "▁star",
+ "ting"
+ ],
+ [
+ "st",
+ "atic"
+ ],
+ [
+ "stat",
+ "ic"
+ ],
+ [
+ "▁b",
+ "ei"
+ ],
+ [
+ "▁be",
+ "i"
+ ],
+ [
+ "al",
+ "low"
+ ],
+ [
+ "all",
+ "ow"
+ ],
+ [
+ "allo",
+ "w"
+ ],
+ [
+ "▁Pe",
+ "ople"
+ ],
+ [
+ "▁",
+ "People"
+ ],
+ [
+ "e",
+ "z"
+ ],
+ [
+ "▁param",
+ "eter"
+ ],
+ [
+ "▁",
+ "parameter"
+ ],
+ [
+ "C",
+ "ache"
+ ],
+ [
+ "▁$",
+ "$"
+ ],
+ [
+ "▁",
+ "$$"
+ ],
+ [
+ "amp",
+ "ions"
+ ],
+ [
+ "ampion",
+ "s"
+ ],
+ [
+ "▁M",
+ "er"
+ ],
+ [
+ "▁Me",
+ "r"
+ ],
+ [
+ "▁",
+ "Mer"
+ ],
+ [
+ "▁k",
+ "om"
+ ],
+ [
+ "▁ko",
+ "m"
+ ],
+ [
+ "▁",
+ "kom"
+ ],
+ [
+ "le",
+ "ted"
+ ],
+ [
+ "let",
+ "ed"
+ ],
+ [
+ "lete",
+ "d"
+ ],
+ [
+ "l",
+ "eted"
+ ],
+ [
+ "oi",
+ "s"
+ ],
+ [
+ "o",
+ "is"
+ ],
+ [
+ "▁O",
+ "pen"
+ ],
+ [
+ "▁Op",
+ "en"
+ ],
+ [
+ "▁",
+ "Open"
+ ],
+ [
+ "ty",
+ "pes"
+ ],
+ [
+ "type",
+ "s"
+ ],
+ [
+ "typ",
+ "es"
+ ],
+ [
+ "t",
+ "ypes"
+ ],
+ [
+ "▁f",
+ "ue"
+ ],
+ [
+ "▁fu",
+ "e"
+ ],
+ [
+ "ac",
+ "ters"
+ ],
+ [
+ "act",
+ "ers"
+ ],
+ [
+ "acter",
+ "s"
+ ],
+ [
+ "▁re",
+ "ference"
+ ],
+ [
+ "▁refer",
+ "ence"
+ ],
+ [
+ "▁",
+ "reference"
+ ],
+ [
+ "Equ",
+ "als"
+ ],
+ [
+ "Equal",
+ "s"
+ ],
+ [
+ "Eq",
+ "uals"
+ ],
+ [
+ "▁a",
+ "ware"
+ ],
+ [
+ "▁aw",
+ "are"
+ ],
+ [
+ "▁",
+ "aware"
+ ],
+ [
+ "▁h",
+ "ol"
+ ],
+ [
+ "▁ho",
+ "l"
+ ],
+ [
+ "▁",
+ "hol"
+ ],
+ [
+ "▁de",
+ "mand"
+ ],
+ [
+ "▁dem",
+ "and"
+ ],
+ [
+ "lo",
+ "r"
+ ],
+ [
+ "l",
+ "or"
+ ],
+ [
+ "▁v",
+ "eh"
+ ],
+ [
+ "▁ve",
+ "h"
+ ],
+ [
+ "▁",
+ "veh"
+ ],
+ [
+ "▁not",
+ "ice"
+ ],
+ [
+ "▁",
+ "notice"
+ ],
+ [
+ "▁com",
+ "ponent"
+ ],
+ [
+ "▁compon",
+ "ent"
+ ],
+ [
+ "▁",
+ "component"
+ ],
+ [
+ "f",
+ "n"
+ ],
+ [
+ "▁anal",
+ "ysis"
+ ],
+ [
+ "▁analy",
+ "sis"
+ ],
+ [
+ "▁analys",
+ "is"
+ ],
+ [
+ "▁",
+ "analysis"
+ ],
+ [
+ "mat",
+ "ch"
+ ],
+ [
+ "m",
+ "atch"
+ ],
+ [
+ "▁effect",
+ "ive"
+ ],
+ [
+ "▁",
+ "effective"
+ ],
+ [
+ "pro",
+ "duct"
+ ],
+ [
+ "produ",
+ "ct"
+ ],
+ [
+ "prod",
+ "uct"
+ ],
+ [
+ "ни",
+ "к"
+ ],
+ [
+ "▁le",
+ "gal"
+ ],
+ [
+ "▁leg",
+ "al"
+ ],
+ [
+ "▁",
+ "legal"
+ ],
+ [
+ "е",
+ "й"
+ ],
+ [
+ "se",
+ "mb"
+ ],
+ [
+ "sem",
+ "b"
+ ],
+ [
+ "s",
+ "emb"
+ ],
+ [
+ "▁loc",
+ "ated"
+ ],
+ [
+ "▁locate",
+ "d"
+ ],
+ [
+ "▁с",
+ "у"
+ ],
+ [
+ "▁",
+ "су"
+ ],
+ [
+ "Q",
+ "L"
+ ],
+ [
+ "in",
+ "ct"
+ ],
+ [
+ "inc",
+ "t"
+ ],
+ [
+ "et",
+ "o"
+ ],
+ [
+ "e",
+ "to"
+ ],
+ [
+ "Dr",
+ "aw"
+ ],
+ [
+ "D",
+ "raw"
+ ],
+ [
+ "▁sc",
+ "ale"
+ ],
+ [
+ "▁scal",
+ "e"
+ ],
+ [
+ "▁",
+ "scale"
+ ],
+ [
+ "ро",
+ "в"
+ ],
+ [
+ "р",
+ "ов"
+ ],
+ [
+ "▁w",
+ "ants"
+ ],
+ [
+ "▁want",
+ "s"
+ ],
+ [
+ "H",
+ "ow"
+ ],
+ [
+ "▁w",
+ "el"
+ ],
+ [
+ "▁we",
+ "l"
+ ],
+ [
+ "is",
+ "ions"
+ ],
+ [
+ "ision",
+ "s"
+ ],
+ [
+ "isi",
+ "ons"
+ ],
+ [
+ "▁de",
+ "liver"
+ ],
+ [
+ "▁del",
+ "iver"
+ ],
+ [
+ "un",
+ "der"
+ ],
+ [
+ "und",
+ "er"
+ ],
+ [
+ "unde",
+ "r"
+ ],
+ [
+ "u",
+ "nder"
+ ],
+ [
+ "▁d",
+ "eb"
+ ],
+ [
+ "▁de",
+ "b"
+ ],
+ [
+ "▁j",
+ "u"
+ ],
+ [
+ "▁",
+ "ju"
+ ],
+ [
+ "val",
+ "ues"
+ ],
+ [
+ "value",
+ "s"
+ ],
+ [
+ "▁s",
+ "ister"
+ ],
+ [
+ "▁si",
+ "ster"
+ ],
+ [
+ "▁sist",
+ "er"
+ ],
+ [
+ "ко",
+ "в"
+ ],
+ [
+ "к",
+ "ов"
+ ],
+ [
+ "▁C",
+ "reate"
+ ],
+ [
+ "▁Creat",
+ "e"
+ ],
+ [
+ "▁Cre",
+ "ate"
+ ],
+ [
+ "▁",
+ "Create"
+ ],
+ [
+ "▁I",
+ "nc"
+ ],
+ [
+ "▁In",
+ "c"
+ ],
+ [
+ "▁a",
+ "ux"
+ ],
+ [
+ "▁au",
+ "x"
+ ],
+ [
+ "▁",
+ "aux"
+ ],
+ [
+ "▁Wh",
+ "ite"
+ ],
+ [
+ "▁Whit",
+ "e"
+ ],
+ [
+ "▁",
+ "White"
+ ],
+ [
+ "Me",
+ "nu"
+ ],
+ [
+ "Men",
+ "u"
+ ],
+ [
+ "M",
+ "enu"
+ ],
+ [
+ "au",
+ "d"
+ ],
+ [
+ "a",
+ "ud"
+ ],
+ [
+ "re",
+ "source"
+ ],
+ [
+ "res",
+ "ource"
+ ],
+ [
+ "▁c",
+ "ab"
+ ],
+ [
+ "▁ca",
+ "b"
+ ],
+ [
+ "▁l",
+ "if"
+ ],
+ [
+ "▁li",
+ "f"
+ ],
+ [
+ "▁",
+ "lif"
+ ],
+ [
+ "▁c",
+ "ulture"
+ ],
+ [
+ "▁cult",
+ "ure"
+ ],
+ [
+ "ic",
+ "he"
+ ],
+ [
+ "ich",
+ "e"
+ ],
+ [
+ "i",
+ "che"
+ ],
+ [
+ "▁wh",
+ "atever"
+ ],
+ [
+ "▁what",
+ "ever"
+ ],
+ [
+ "▁de",
+ "signed"
+ ],
+ [
+ "▁des",
+ "igned"
+ ],
+ [
+ "▁design",
+ "ed"
+ ],
+ [
+ "▁re",
+ "pe"
+ ],
+ [
+ "▁rep",
+ "e"
+ ],
+ [
+ "▁M",
+ "ont"
+ ],
+ [
+ "▁Mon",
+ "t"
+ ],
+ [
+ "▁Mo",
+ "nt"
+ ],
+ [
+ "▁",
+ "Mont"
+ ],
+ [
+ "▁ch",
+ "arge"
+ ],
+ [
+ "▁char",
+ "ge"
+ ],
+ [
+ "▁charg",
+ "e"
+ ],
+ [
+ "▁",
+ "charge"
+ ],
+ [
+ "Name",
+ "s"
+ ],
+ [
+ "Na",
+ "mes"
+ ],
+ [
+ "N",
+ "ames"
+ ],
+ [
+ "▁in",
+ "sp"
+ ],
+ [
+ "▁ins",
+ "p"
+ ],
+ [
+ "▁custom",
+ "ers"
+ ],
+ [
+ "▁customer",
+ "s"
+ ],
+ [
+ "os",
+ "a"
+ ],
+ [
+ "o",
+ "sa"
+ ],
+ [
+ "▁d",
+ "aughter"
+ ],
+ [
+ "▁E",
+ "ast"
+ ],
+ [
+ "E",
+ "Q"
+ ],
+ [
+ "▁o",
+ "pin"
+ ],
+ [
+ "▁op",
+ "in"
+ ],
+ [
+ "▁F",
+ "re"
+ ],
+ [
+ "▁Fr",
+ "e"
+ ],
+ [
+ "▁se",
+ "ek"
+ ],
+ [
+ "▁see",
+ "k"
+ ],
+ [
+ "▁",
+ "seek"
+ ],
+ [
+ "▁p",
+ "ush"
+ ],
+ [
+ "▁pu",
+ "sh"
+ ],
+ [
+ "▁",
+ "push"
+ ],
+ [
+ "▁n",
+ "av"
+ ],
+ [
+ "▁na",
+ "v"
+ ],
+ [
+ "▁",
+ "nav"
+ ],
+ [
+ "▁b",
+ "urn"
+ ],
+ [
+ "▁bu",
+ "rn"
+ ],
+ [
+ "▁bur",
+ "n"
+ ],
+ [
+ "▁",
+ "burn"
+ ],
+ [
+ "ar",
+ "den"
+ ],
+ [
+ "ard",
+ "en"
+ ],
+ [
+ "arde",
+ "n"
+ ],
+ [
+ "ha",
+ "sh"
+ ],
+ [
+ "has",
+ "h"
+ ],
+ [
+ "h",
+ "ash"
+ ],
+ [
+ "▁opportun",
+ "ity"
+ ],
+ [
+ "▁M",
+ "at"
+ ],
+ [
+ "▁Ma",
+ "t"
+ ],
+ [
+ "▁",
+ "Mat"
+ ],
+ [
+ "oy",
+ "al"
+ ],
+ [
+ "oya",
+ "l"
+ ],
+ [
+ "o",
+ "yal"
+ ],
+ [
+ "▁p",
+ "un"
+ ],
+ [
+ "▁pu",
+ "n"
+ ],
+ [
+ "sc",
+ "ale"
+ ],
+ [
+ "scal",
+ "e"
+ ],
+ [
+ "yn",
+ "amic"
+ ],
+ [
+ "ynam",
+ "ic"
+ ],
+ [
+ "yna",
+ "mic"
+ ],
+ [
+ "▁T",
+ "ype"
+ ],
+ [
+ "▁Ty",
+ "pe"
+ ],
+ [
+ "▁Typ",
+ "e"
+ ],
+ [
+ "▁",
+ "Type"
+ ],
+ [
+ "il",
+ "ing"
+ ],
+ [
+ "ili",
+ "ng"
+ ],
+ [
+ "i",
+ "ling"
+ ],
+ [
+ "▁qu",
+ "ery"
+ ],
+ [
+ "▁que",
+ "ry"
+ ],
+ [
+ "▁quer",
+ "y"
+ ],
+ [
+ "▁",
+ "query"
+ ],
+ [
+ "▁m",
+ "ist"
+ ],
+ [
+ "▁mis",
+ "t"
+ ],
+ [
+ "▁mi",
+ "st"
+ ],
+ [
+ "ro",
+ "r"
+ ],
+ [
+ "r",
+ "or"
+ ],
+ [
+ "for",
+ "ce"
+ ],
+ [
+ "▁On",
+ "ce"
+ ],
+ [
+ "▁",
+ "Once"
+ ],
+ [
+ "▁med",
+ "ical"
+ ],
+ [
+ "▁medic",
+ "al"
+ ],
+ [
+ "▁medi",
+ "cal"
+ ],
+ [
+ "li",
+ "e"
+ ],
+ [
+ "l",
+ "ie"
+ ],
+ [
+ "▁stud",
+ "ent"
+ ],
+ [
+ "▁",
+ "student"
+ ],
+ [
+ "ed",
+ "eral"
+ ],
+ [
+ "eder",
+ "al"
+ ],
+ [
+ "ede",
+ "ral"
+ ],
+ [
+ "▁l",
+ "ov"
+ ],
+ [
+ "▁lo",
+ "v"
+ ],
+ [
+ "▁",
+ "lov"
+ ],
+ [
+ "if",
+ "orm"
+ ],
+ [
+ "i",
+ "form"
+ ],
+ [
+ "▁al",
+ "tern"
+ ],
+ [
+ "▁alt",
+ "ern"
+ ],
+ [
+ "▁alter",
+ "n"
+ ],
+ [
+ "▁",
+ "altern"
+ ],
+ [
+ "bi",
+ "n"
+ ],
+ [
+ "b",
+ "in"
+ ],
+ [
+ "od",
+ "er"
+ ],
+ [
+ "ode",
+ "r"
+ ],
+ [
+ "o",
+ "der"
+ ],
+ [
+ "▁return",
+ "s"
+ ],
+ [
+ "▁",
+ "returns"
+ ],
+ [
+ "reg",
+ "ister"
+ ],
+ [
+ "ut",
+ "s"
+ ],
+ [
+ "u",
+ "ts"
+ ],
+ [
+ "C",
+ "I"
+ ],
+ [
+ "▁T",
+ "or"
+ ],
+ [
+ "▁To",
+ "r"
+ ],
+ [
+ "▁",
+ "Tor"
+ ],
+ [
+ "C",
+ "R"
+ ],
+ [
+ "▁L",
+ "os"
+ ],
+ [
+ "▁Lo",
+ "s"
+ ],
+ [
+ "▁",
+ "Los"
+ ],
+ [
+ "am",
+ "ily"
+ ],
+ [
+ "ami",
+ "ly"
+ ],
+ [
+ "amil",
+ "y"
+ ],
+ [
+ "air",
+ "e"
+ ],
+ [
+ "ai",
+ "re"
+ ],
+ [
+ "a",
+ "ire"
+ ],
+ [
+ "++",
+ ";"
+ ],
+ [
+ "Cont",
+ "roller"
+ ],
+ [
+ "Control",
+ "ler"
+ ],
+ [
+ "wi",
+ "de"
+ ],
+ [
+ "wid",
+ "e"
+ ],
+ [
+ "w",
+ "ide"
+ ],
+ [
+ "x",
+ "x"
+ ],
+ [
+ "row",
+ "ser"
+ ],
+ [
+ "rows",
+ "er"
+ ],
+ [
+ "▁B",
+ "ook"
+ ],
+ [
+ "▁Bo",
+ "ok"
+ ],
+ [
+ "▁",
+ "Book"
+ ],
+ [
+ "Cont",
+ "ainer"
+ ],
+ [
+ "pl",
+ "oad"
+ ],
+ [
+ "plo",
+ "ad"
+ ],
+ [
+ "p",
+ "load"
+ ],
+ [
+ "▁E",
+ "v"
+ ],
+ [
+ "▁",
+ "Ev"
+ ],
+ [
+ "▁t",
+ "al"
+ ],
+ [
+ "▁ta",
+ "l"
+ ],
+ [
+ "▁",
+ "tal"
+ ],
+ [
+ "▁the",
+ "ory"
+ ],
+ [
+ "eqn",
+ "array"
+ ],
+ [
+ "б",
+ "е"
+ ],
+ [
+ "▁rep",
+ "orted"
+ ],
+ [
+ "▁report",
+ "ed"
+ ],
+ [
+ "▁me",
+ "aning"
+ ],
+ [
+ "▁mean",
+ "ing"
+ ],
+ [
+ "▁s",
+ "y"
+ ],
+ [
+ "▁",
+ "sy"
+ ],
+ [
+ "ri",
+ "be"
+ ],
+ [
+ "rib",
+ "e"
+ ],
+ [
+ "r",
+ "ibe"
+ ],
+ [
+ "ic",
+ "ate"
+ ],
+ [
+ "ica",
+ "te"
+ ],
+ [
+ "ho",
+ "ld"
+ ],
+ [
+ "hol",
+ "d"
+ ],
+ [
+ "h",
+ "old"
+ ],
+ [
+ "▁of",
+ "fers"
+ ],
+ [
+ "▁off",
+ "ers"
+ ],
+ [
+ "▁offer",
+ "s"
+ ],
+ [
+ "▁t",
+ "empl"
+ ],
+ [
+ "▁tem",
+ "pl"
+ ],
+ [
+ "▁temp",
+ "l"
+ ],
+ [
+ "cs",
+ "s"
+ ],
+ [
+ "c",
+ "ss"
+ ],
+ [
+ "▁p",
+ "icture"
+ ],
+ [
+ "▁pict",
+ "ure"
+ ],
+ [
+ "▁",
+ "picture"
+ ],
+ [
+ "▁a",
+ "sync"
+ ],
+ [
+ "▁as",
+ "ync"
+ ],
+ [
+ "▁",
+ "async"
+ ],
+ [
+ "▁st",
+ "ock"
+ ],
+ [
+ "▁sto",
+ "ck"
+ ],
+ [
+ "▁",
+ "stock"
+ ],
+ [
+ "▁in",
+ "ternal"
+ ],
+ [
+ "▁inter",
+ "nal"
+ ],
+ [
+ "▁intern",
+ "al"
+ ],
+ [
+ "▁",
+ "internal"
+ ],
+ [
+ "t",
+ "i"
+ ],
+ [
+ "B",
+ "O"
+ ],
+ [
+ "V",
+ "er"
+ ],
+ [
+ "с",
+ "по"
+ ],
+ [
+ "▁d",
+ "emon"
+ ],
+ [
+ "▁de",
+ "mon"
+ ],
+ [
+ "▁dem",
+ "on"
+ ],
+ [
+ "▁demo",
+ "n"
+ ],
+ [
+ "▁l",
+ "augh"
+ ],
+ [
+ "▁la",
+ "ugh"
+ ],
+ [
+ "▁laug",
+ "h"
+ ],
+ [
+ "▁E",
+ "nd"
+ ],
+ [
+ "▁En",
+ "d"
+ ],
+ [
+ "▁",
+ "End"
+ ],
+ [
+ "▁k",
+ "on"
+ ],
+ [
+ "▁ko",
+ "n"
+ ],
+ [
+ "▁",
+ "kon"
+ ],
+ [
+ "▁ide",
+ "as"
+ ],
+ [
+ "▁idea",
+ "s"
+ ],
+ [
+ "▁c",
+ "andid"
+ ],
+ [
+ "▁can",
+ "did"
+ ],
+ [
+ "▁cand",
+ "id"
+ ],
+ [
+ "Me",
+ "m"
+ ],
+ [
+ "M",
+ "em"
+ ],
+ [
+ "iz",
+ "z"
+ ],
+ [
+ "i",
+ "zz"
+ ],
+ [
+ "re",
+ "fix"
+ ],
+ [
+ "ref",
+ "ix"
+ ],
+ [
+ "▁A",
+ "ND"
+ ],
+ [
+ "▁AN",
+ "D"
+ ],
+ [
+ "▁",
+ "AND"
+ ],
+ [
+ "eg",
+ "en"
+ ],
+ [
+ "e",
+ "gen"
+ ],
+ [
+ "E",
+ "l"
+ ],
+ [
+ "▁camp",
+ "aign"
+ ],
+ [
+ "H",
+ "ttp"
+ ],
+ [
+ "▁R",
+ "ob"
+ ],
+ [
+ "▁Ro",
+ "b"
+ ],
+ [
+ "▁",
+ "Rob"
+ ],
+ [
+ "д",
+ "і"
+ ],
+ [
+ "▁b",
+ "ul"
+ ],
+ [
+ "▁bu",
+ "l"
+ ],
+ [
+ "▁",
+ "bul"
+ ],
+ [
+ "▁К",
+ "о"
+ ],
+ [
+ "▁",
+ "Ко"
+ ],
+ [
+ "▁count",
+ "ries"
+ ],
+ [
+ "▁countr",
+ "ies"
+ ],
+ [
+ "»",
+ "."
+ ],
+ [
+ "▁ex",
+ "pression"
+ ],
+ [
+ "▁exp",
+ "ression"
+ ],
+ [
+ "▁express",
+ "ion"
+ ],
+ [
+ "▁expr",
+ "ession"
+ ],
+ [
+ "▁",
+ "expression"
+ ],
+ [
+ "▁Eng",
+ "land"
+ ],
+ [
+ "s",
+ "f"
+ ],
+ [
+ "▁certain",
+ "ly"
+ ],
+ [
+ "ag",
+ "en"
+ ],
+ [
+ "age",
+ "n"
+ ],
+ [
+ "a",
+ "gen"
+ ],
+ [
+ "▁ч",
+ "а"
+ ],
+ [
+ "▁",
+ "ча"
+ ],
+ [
+ "▁A",
+ "NY"
+ ],
+ [
+ "▁AN",
+ "Y"
+ ],
+ [
+ "▁",
+ "ANY"
+ ],
+ [
+ "▁conne",
+ "ct"
+ ],
+ [
+ "▁conn",
+ "ect"
+ ],
+ [
+ "▁",
+ "connect"
+ ],
+ [
+ "F",
+ "E"
+ ],
+ [
+ "▁and",
+ "roid"
+ ],
+ [
+ "▁",
+ "android"
+ ],
+ [
+ "▁G",
+ "old"
+ ],
+ [
+ "▁Go",
+ "ld"
+ ],
+ [
+ "▁Gol",
+ "d"
+ ],
+ [
+ "▁",
+ "Gold"
+ ],
+ [
+ "▁op",
+ "pos"
+ ],
+ [
+ "▁opp",
+ "os"
+ ],
+ [
+ "ov",
+ "ern"
+ ],
+ [
+ "ove",
+ "rn"
+ ],
+ [
+ "over",
+ "n"
+ ],
+ [
+ "o",
+ "vern"
+ ],
+ [
+ "▁Com",
+ "mun"
+ ],
+ [
+ "▁Comm",
+ "un"
+ ],
+ [
+ ",",
+ "_"
+ ],
+ [
+ "as",
+ "ion"
+ ],
+ [
+ "asi",
+ "on"
+ ],
+ [
+ "L",
+ "a"
+ ],
+ [
+ "▁f",
+ "irm"
+ ],
+ [
+ "▁fi",
+ "rm"
+ ],
+ [
+ "▁fir",
+ "m"
+ ],
+ [
+ "▁Al",
+ "though"
+ ],
+ [
+ "▁G",
+ "ood"
+ ],
+ [
+ "▁Go",
+ "od"
+ ],
+ [
+ "▁",
+ "Good"
+ ],
+ [
+ "▁L",
+ "aw"
+ ],
+ [
+ "▁La",
+ "w"
+ ],
+ [
+ "er",
+ "ve"
+ ],
+ [
+ "erv",
+ "e"
+ ],
+ [
+ "▁b",
+ "rand"
+ ],
+ [
+ "▁br",
+ "and"
+ ],
+ [
+ "▁bra",
+ "nd"
+ ],
+ [
+ "▁",
+ "brand"
+ ],
+ [
+ "M",
+ "in"
+ ],
+ [
+ "fil",
+ "l"
+ ],
+ [
+ "fi",
+ "ll"
+ ],
+ [
+ "f",
+ "ill"
+ ],
+ [
+ "']",
+ ","
+ ],
+ [
+ "'",
+ "],"
+ ],
+ [
+ "▁J",
+ "ew"
+ ],
+ [
+ "▁Je",
+ "w"
+ ],
+ [
+ "il",
+ "er"
+ ],
+ [
+ "ile",
+ "r"
+ ],
+ [
+ "i",
+ "ler"
+ ],
+ [
+ "in",
+ "gle"
+ ],
+ [
+ "ing",
+ "le"
+ ],
+ [
+ "it",
+ "hub"
+ ],
+ [
+ "ith",
+ "ub"
+ ],
+ [
+ "▁D",
+ "iv"
+ ],
+ [
+ "▁Di",
+ "v"
+ ],
+ [
+ "▁",
+ "Div"
+ ],
+ [
+ "▁c",
+ "ert"
+ ],
+ [
+ "▁ce",
+ "rt"
+ ],
+ [
+ "▁cer",
+ "t"
+ ],
+ [
+ "▁",
+ "cert"
+ ],
+ [
+ "He",
+ "ight"
+ ],
+ [
+ "H",
+ "eight"
+ ],
+ [
+ "ra",
+ "el"
+ ],
+ [
+ "r",
+ "ael"
+ ],
+ [
+ "The",
+ "re"
+ ],
+ [
+ "Th",
+ "ere"
+ ],
+ [
+ "T",
+ "here"
+ ],
+ [
+ "it",
+ "ute"
+ ],
+ [
+ "itut",
+ "e"
+ ],
+ [
+ "itu",
+ "te"
+ ],
+ [
+ "▁a",
+ "maz"
+ ],
+ [
+ "▁am",
+ "az"
+ ],
+ [
+ "▁",
+ "amaz"
+ ],
+ [
+ "lo",
+ "ok"
+ ],
+ [
+ "l",
+ "ook"
+ ],
+ [
+ "▁S",
+ "E"
+ ],
+ [
+ "▁",
+ "SE"
+ ],
+ [
+ "▁j",
+ "o"
+ ],
+ [
+ "▁",
+ "jo"
+ ],
+ [
+ "▁pull",
+ "ed"
+ ],
+ [
+ "▁pul",
+ "led"
+ ],
+ [
+ "▁re",
+ "sources"
+ ],
+ [
+ "▁res",
+ "ources"
+ ],
+ [
+ "▁resource",
+ "s"
+ ],
+ [
+ "▁",
+ "resources"
+ ],
+ [
+ "▁M",
+ "ax"
+ ],
+ [
+ "▁Ma",
+ "x"
+ ],
+ [
+ "▁",
+ "Max"
+ ],
+ [
+ "▁ag",
+ "reed"
+ ],
+ [
+ "▁agree",
+ "d"
+ ],
+ [
+ "▁agre",
+ "ed"
+ ],
+ [
+ "as",
+ "y"
+ ],
+ [
+ "a",
+ "sy"
+ ],
+ [
+ "▁treat",
+ "ment"
+ ],
+ [
+ "\">",
+ ""
+ ],
+ [
+ "\"><",
+ "/"
+ ],
+ [
+ "\"",
+ ">"
+ ],
+ [
+ "ма",
+ "н"
+ ],
+ [
+ "м",
+ "ан"
+ ],
+ [
+ "▁E",
+ "rr"
+ ],
+ [
+ "▁Er",
+ "r"
+ ],
+ [
+ "▁",
+ "Err"
+ ],
+ [
+ "or",
+ "ig"
+ ],
+ [
+ "ori",
+ "g"
+ ],
+ [
+ "o",
+ "rig"
+ ],
+ [
+ "co",
+ "s"
+ ],
+ [
+ "c",
+ "os"
+ ],
+ [
+ "▁May",
+ "be"
+ ],
+ [
+ "▁",
+ "Maybe"
+ ],
+ [
+ "ot",
+ "al"
+ ],
+ [
+ "ota",
+ "l"
+ ],
+ [
+ "o",
+ "tal"
+ ],
+ [
+ "▁tr",
+ "ain"
+ ],
+ [
+ "▁tra",
+ "in"
+ ],
+ [
+ "▁",
+ "train"
+ ],
+ [
+ "▁S",
+ "ervice"
+ ],
+ [
+ "▁Serv",
+ "ice"
+ ],
+ [
+ "▁",
+ "Service"
+ ],
+ [
+ "▁i",
+ "h"
+ ],
+ [
+ "▁",
+ "ih"
+ ],
+ [
+ "▁sp",
+ "irit"
+ ],
+ [
+ "▁spir",
+ "it"
+ ],
+ [
+ "Com",
+ "p"
+ ],
+ [
+ "Co",
+ "mp"
+ ],
+ [
+ "C",
+ "omp"
+ ],
+ [
+ "sq",
+ "rt"
+ ],
+ [
+ "▁b",
+ "road"
+ ],
+ [
+ "▁br",
+ "oad"
+ ],
+ [
+ "▁bro",
+ "ad"
+ ],
+ [
+ "▁",
+ "broad"
+ ],
+ [
+ "}",
+ "["
+ ],
+ [
+ "▁sh",
+ "ape"
+ ],
+ [
+ "▁sha",
+ "pe"
+ ],
+ [
+ "▁",
+ "shape"
+ ],
+ [
+ "▁d",
+ "oc"
+ ],
+ [
+ "▁do",
+ "c"
+ ],
+ [
+ "▁",
+ "doc"
+ ],
+ [
+ "ho",
+ "w"
+ ],
+ [
+ "h",
+ "ow"
+ ],
+ [
+ "▁t",
+ "ag"
+ ],
+ [
+ "▁ta",
+ "g"
+ ],
+ [
+ "▁",
+ "tag"
+ ],
+ [
+ "ata",
+ "log"
+ ],
+ [
+ "atal",
+ "og"
+ ],
+ [
+ "s",
+ "d"
+ ],
+ [
+ "▁me",
+ "as"
+ ],
+ [
+ "▁Р",
+ "о"
+ ],
+ [
+ "▁ex",
+ "ception"
+ ],
+ [
+ "▁except",
+ "ion"
+ ],
+ [
+ "▁",
+ "exception"
+ ],
+ [
+ "▁T",
+ "w"
+ ],
+ [
+ "▁",
+ "Tw"
+ ],
+ [
+ "▁interest",
+ "ing"
+ ],
+ [
+ "AT",
+ "A"
+ ],
+ [
+ "A",
+ "TA"
+ ],
+ [
+ "▁R",
+ "el"
+ ],
+ [
+ "▁Re",
+ "l"
+ ],
+ [
+ "▁",
+ "Rel"
+ ],
+ [
+ "á",
+ "r"
+ ],
+ [
+ "▁use",
+ "ful"
+ ],
+ [
+ "use",
+ "um"
+ ],
+ [
+ "▁b",
+ "ottom"
+ ],
+ [
+ "▁bott",
+ "om"
+ ],
+ [
+ "▁bot",
+ "tom"
+ ],
+ [
+ "▁",
+ "bottom"
+ ],
+ [
+ "▁other",
+ "wise"
+ ],
+ [
+ "▁ag",
+ "ree"
+ ],
+ [
+ "▁agre",
+ "e"
+ ],
+ [
+ "ch",
+ "t"
+ ],
+ [
+ "c",
+ "ht"
+ ],
+ [
+ "th",
+ "en"
+ ],
+ [
+ "the",
+ "n"
+ ],
+ [
+ "t",
+ "hen"
+ ],
+ [
+ "▁signific",
+ "ant"
+ ],
+ [
+ "}",
+ "/"
+ ],
+ [
+ "▁ch",
+ "annel"
+ ],
+ [
+ "▁",
+ "channel"
+ ],
+ [
+ "ic",
+ "ial"
+ ],
+ [
+ "ici",
+ "al"
+ ],
+ [
+ "icia",
+ "l"
+ ],
+ [
+ "i",
+ "cial"
+ ],
+ [
+ "ти",
+ "в"
+ ],
+ [
+ "var",
+ "e"
+ ],
+ [
+ "va",
+ "re"
+ ],
+ [
+ "v",
+ "are"
+ ],
+ [
+ "▁en",
+ "ter"
+ ],
+ [
+ "▁ent",
+ "er"
+ ],
+ [
+ "▁",
+ "enter"
+ ],
+ [
+ "En",
+ "g"
+ ],
+ [
+ "E",
+ "ng"
+ ],
+ [
+ "u",
+ "j"
+ ],
+ [
+ "UR",
+ "E"
+ ],
+ [
+ "U",
+ "RE"
+ ],
+ [
+ "que",
+ "ue"
+ ],
+ [
+ "on",
+ "o"
+ ],
+ [
+ "o",
+ "no"
+ ],
+ [
+ "▁cont",
+ "ains"
+ ],
+ [
+ "▁contain",
+ "s"
+ ],
+ [
+ "▁",
+ "contains"
+ ],
+ [
+ "M",
+ "I"
+ ],
+ [
+ "▁n",
+ "ation"
+ ],
+ [
+ "▁nat",
+ "ion"
+ ],
+ [
+ "▁r",
+ "ules"
+ ],
+ [
+ "▁rule",
+ "s"
+ ],
+ [
+ "▁ru",
+ "les"
+ ],
+ [
+ "▁rul",
+ "es"
+ ],
+ [
+ "▁",
+ "rules"
+ ],
+ [
+ "fo",
+ "l"
+ ],
+ [
+ "f",
+ "ol"
+ ],
+ [
+ "▁p",
+ "a"
+ ],
+ [
+ "▁",
+ "pa"
+ ],
+ [
+ "ar",
+ "p"
+ ],
+ [
+ "a",
+ "rp"
+ ],
+ [
+ "▁qu",
+ "iet"
+ ],
+ [
+ "▁qui",
+ "et"
+ ],
+ [
+ "▁t",
+ "hus"
+ ],
+ [
+ "▁th",
+ "us"
+ ],
+ [
+ "ip",
+ "ped"
+ ],
+ [
+ "ipp",
+ "ed"
+ ],
+ [
+ "i",
+ "pped"
+ ],
+ [
+ "an",
+ "not"
+ ],
+ [
+ "ann",
+ "ot"
+ ],
+ [
+ "anno",
+ "t"
+ ],
+ [
+ "ud",
+ "es"
+ ],
+ [
+ "ude",
+ "s"
+ ],
+ [
+ "u",
+ "des"
+ ],
+ [
+ "()",
+ ":"
+ ],
+ [
+ "(",
+ "):"
+ ],
+ [
+ "name",
+ "s"
+ ],
+ [
+ "na",
+ "mes"
+ ],
+ [
+ "nam",
+ "es"
+ ],
+ [
+ "n",
+ "ames"
+ ],
+ [
+ "▁com",
+ "pos"
+ ],
+ [
+ "▁comp",
+ "os"
+ ],
+ [
+ "▁in",
+ "j"
+ ],
+ [
+ "un",
+ "a"
+ ],
+ [
+ "u",
+ "na"
+ ],
+ [
+ "bin",
+ "d"
+ ],
+ [
+ "bi",
+ "nd"
+ ],
+ [
+ "b",
+ "ind"
+ ],
+ [
+ "▁f",
+ "ully"
+ ],
+ [
+ "▁full",
+ "y"
+ ],
+ [
+ "▁ful",
+ "ly"
+ ],
+ [
+ "▁",
+ "fully"
+ ],
+ [
+ "ra",
+ "s"
+ ],
+ [
+ "r",
+ "as"
+ ],
+ [
+ "Util",
+ "s"
+ ],
+ [
+ "Ut",
+ "ils"
+ ],
+ [
+ "an",
+ "ges"
+ ],
+ [
+ "ang",
+ "es"
+ ],
+ [
+ "ange",
+ "s"
+ ],
+ [
+ "du",
+ "le"
+ ],
+ [
+ "d",
+ "ule"
+ ],
+ [
+ "▁Christ",
+ "ian"
+ ],
+ [
+ "▁re",
+ "ve"
+ ],
+ [
+ "▁r",
+ "eve"
+ ],
+ [
+ "▁rev",
+ "e"
+ ],
+ [
+ "än",
+ "d"
+ ],
+ [
+ "ä",
+ "nd"
+ ],
+ [
+ "▁col",
+ "lect"
+ ],
+ [
+ "▁coll",
+ "ect"
+ ],
+ [
+ "▁colle",
+ "ct"
+ ],
+ [
+ "▁",
+ "collect"
+ ],
+ [
+ "▁cele",
+ "br"
+ ],
+ [
+ "an",
+ "da"
+ ],
+ [
+ "and",
+ "a"
+ ],
+ [
+ "í",
+ "n"
+ ],
+ [
+ "jo",
+ "in"
+ ],
+ [
+ "j",
+ "oin"
+ ],
+ [
+ "▁p",
+ "aid"
+ ],
+ [
+ "▁pa",
+ "id"
+ ],
+ [
+ "▁",
+ "paid"
+ ],
+ [
+ "Co",
+ "re"
+ ],
+ [
+ "Cor",
+ "e"
+ ],
+ [
+ "C",
+ "ore"
+ ],
+ [
+ "G",
+ "e"
+ ],
+ [
+ ".",
+ "$"
+ ],
+ [
+ "▁f",
+ "if"
+ ],
+ [
+ "▁fi",
+ "f"
+ ],
+ [
+ "▁",
+ "fif"
+ ],
+ [
+ "▁u",
+ "ma"
+ ],
+ [
+ "▁um",
+ "a"
+ ],
+ [
+ "▁",
+ "uma"
+ ],
+ [
+ "▁",
+ "~"
+ ],
+ [
+ "erv",
+ "ices"
+ ],
+ [
+ "ervice",
+ "s"
+ ],
+ [
+ "▁rec",
+ "ently"
+ ],
+ [
+ "▁recent",
+ "ly"
+ ],
+ [
+ "de",
+ "sc"
+ ],
+ [
+ "des",
+ "c"
+ ],
+ [
+ "d",
+ "esc"
+ ],
+ [
+ "▁he",
+ "avy"
+ ],
+ [
+ "▁heav",
+ "y"
+ ],
+ [
+ "▁r",
+ "ule"
+ ],
+ [
+ "▁ru",
+ "le"
+ ],
+ [
+ "▁rul",
+ "e"
+ ],
+ [
+ "▁",
+ "rule"
+ ],
+ [
+ "▁P",
+ "lease"
+ ],
+ [
+ "▁Ple",
+ "ase"
+ ],
+ [
+ "▁",
+ "Please"
+ ],
+ [
+ "ps",
+ "i"
+ ],
+ [
+ "p",
+ "si"
+ ],
+ [
+ "▁con",
+ "sole"
+ ],
+ [
+ "▁cons",
+ "ole"
+ ],
+ [
+ "▁",
+ "console"
+ ],
+ [
+ "▁f",
+ "ort"
+ ],
+ [
+ "▁for",
+ "t"
+ ],
+ [
+ "▁fo",
+ "rt"
+ ],
+ [
+ "▁",
+ "fort"
+ ],
+ [
+ ".",
+ "\\"
+ ],
+ [
+ "▁W",
+ "ashington"
+ ],
+ [
+ "▁g",
+ "ar"
+ ],
+ [
+ "▁ga",
+ "r"
+ ],
+ [
+ "▁",
+ "gar"
+ ],
+ [
+ "▁G",
+ "roup"
+ ],
+ [
+ "▁Gr",
+ "oup"
+ ],
+ [
+ "▁Gro",
+ "up"
+ ],
+ [
+ "▁",
+ "Group"
+ ],
+ [
+ "▁inter",
+ "view"
+ ],
+ [
+ "an",
+ "ned"
+ ],
+ [
+ "ann",
+ "ed"
+ ],
+ [
+ "anne",
+ "d"
+ ],
+ [
+ "sq",
+ "l"
+ ],
+ [
+ "s",
+ "ql"
+ ],
+ [
+ "▁a",
+ "nc"
+ ],
+ [
+ "▁an",
+ "c"
+ ],
+ [
+ "▁",
+ "anc"
+ ],
+ [
+ "ј",
+ "а"
+ ],
+ [
+ "P",
+ "ack"
+ ],
+ [
+ "▁Cl",
+ "ub"
+ ],
+ [
+ "▁m",
+ "ask"
+ ],
+ [
+ "▁ma",
+ "sk"
+ ],
+ [
+ "▁mas",
+ "k"
+ ],
+ [
+ "▁",
+ "mask"
+ ],
+ [
+ "▁con",
+ "cept"
+ ],
+ [
+ "▁conce",
+ "pt"
+ ],
+ [
+ "▁[",
+ "'"
+ ],
+ [
+ "▁",
+ "['"
+ ],
+ [
+ "▁se",
+ "lected"
+ ],
+ [
+ "▁select",
+ "ed"
+ ],
+ [
+ "▁sele",
+ "cted"
+ ],
+ [
+ "▁",
+ "selected"
+ ],
+ [
+ "▁U",
+ "se"
+ ],
+ [
+ "▁Us",
+ "e"
+ ],
+ [
+ "▁",
+ "Use"
+ ],
+ [
+ "▁e",
+ "le"
+ ],
+ [
+ "▁el",
+ "e"
+ ],
+ [
+ "▁",
+ "ele"
+ ],
+ [
+ "ear",
+ "s"
+ ],
+ [
+ "ea",
+ "rs"
+ ],
+ [
+ "e",
+ "ars"
+ ],
+ [
+ "▁r",
+ "ace"
+ ],
+ [
+ "▁rac",
+ "e"
+ ],
+ [
+ "▁ra",
+ "ce"
+ ],
+ [
+ "h",
+ "y"
+ ],
+ [
+ "O",
+ "m"
+ ],
+ [
+ "▁st",
+ "eps"
+ ],
+ [
+ "▁ste",
+ "ps"
+ ],
+ [
+ "▁step",
+ "s"
+ ],
+ [
+ "▁",
+ "steps"
+ ],
+ [
+ "il",
+ "a"
+ ],
+ [
+ "i",
+ "la"
+ ],
+ [
+ "es",
+ "ts"
+ ],
+ [
+ "est",
+ "s"
+ ],
+ [
+ "e",
+ "sts"
+ ],
+ [
+ "ed",
+ "s"
+ ],
+ [
+ "e",
+ "ds"
+ ],
+ [
+ "▁stre",
+ "et"
+ ],
+ [
+ "ne",
+ "rs"
+ ],
+ [
+ "ner",
+ "s"
+ ],
+ [
+ "n",
+ "ers"
+ ],
+ [
+ "▁b",
+ "irth"
+ ],
+ [
+ "po",
+ "p"
+ ],
+ [
+ "p",
+ "op"
+ ],
+ [
+ "▁",
+ "ли"
+ ],
+ [
+ "M",
+ "B"
+ ],
+ [
+ "к",
+ "ра"
+ ],
+ [
+ "ci",
+ "r"
+ ],
+ [
+ "c",
+ "ir"
+ ],
+ [
+ "eps",
+ "ilon"
+ ],
+ [
+ "e",
+ "psilon"
+ ],
+ [
+ "▁con",
+ "stant"
+ ],
+ [
+ "▁const",
+ "ant"
+ ],
+ [
+ "▁",
+ "constant"
+ ],
+ [
+ "qu",
+ "es"
+ ],
+ [
+ "que",
+ "s"
+ ],
+ [
+ "q",
+ "ues"
+ ],
+ [
+ "ad",
+ "as"
+ ],
+ [
+ "ada",
+ "s"
+ ],
+ [
+ "a",
+ "das"
+ ],
+ [
+ "▁kn",
+ "ows"
+ ],
+ [
+ "▁know",
+ "s"
+ ],
+ [
+ "▁P",
+ "y"
+ ],
+ [
+ "▁",
+ "Py"
+ ],
+ [
+ "cl",
+ "es"
+ ],
+ [
+ "cle",
+ "s"
+ ],
+ [
+ "c",
+ "les"
+ ],
+ [
+ "▁c",
+ "it"
+ ],
+ [
+ "▁ci",
+ "t"
+ ],
+ [
+ "▁",
+ "cit"
+ ],
+ [
+ "▁p",
+ "air"
+ ],
+ [
+ "▁pa",
+ "ir"
+ ],
+ [
+ "▁",
+ "pair"
+ ],
+ [
+ "in",
+ "ese"
+ ],
+ [
+ "ine",
+ "se"
+ ],
+ [
+ "ines",
+ "e"
+ ],
+ [
+ "▁P",
+ "eter"
+ ],
+ [
+ "▁Pe",
+ "ter"
+ ],
+ [
+ "▁Pet",
+ "er"
+ ],
+ [
+ "▁Pete",
+ "r"
+ ],
+ [
+ "▁fin",
+ "ished"
+ ],
+ [
+ "▁finish",
+ "ed"
+ ],
+ [
+ "▁",
+ "finished"
+ ],
+ [
+ "▁m",
+ "aster"
+ ],
+ [
+ "▁ma",
+ "ster"
+ ],
+ [
+ "▁mas",
+ "ter"
+ ],
+ [
+ "▁mast",
+ "er"
+ ],
+ [
+ "▁",
+ "master"
+ ],
+ [
+ "▁tw",
+ "enty"
+ ],
+ [
+ "▁f",
+ "ell"
+ ],
+ [
+ "▁fe",
+ "ll"
+ ],
+ [
+ "▁fel",
+ "l"
+ ],
+ [
+ "▁cent",
+ "ral"
+ ],
+ [
+ "▁m",
+ "es"
+ ],
+ [
+ "▁me",
+ "s"
+ ],
+ [
+ "▁",
+ "mes"
+ ],
+ [
+ "re",
+ "v"
+ ],
+ [
+ "r",
+ "ev"
+ ],
+ [
+ "ST",
+ "AT"
+ ],
+ [
+ "st",
+ "at"
+ ],
+ [
+ "sta",
+ "t"
+ ],
+ [
+ "s",
+ "tat"
+ ],
+ [
+ "▁all",
+ "ows"
+ ],
+ [
+ "▁allow",
+ "s"
+ ],
+ [
+ "▁g",
+ "ro"
+ ],
+ [
+ "▁gr",
+ "o"
+ ],
+ [
+ "▁",
+ "gro"
+ ],
+ [
+ "Cl",
+ "ick"
+ ],
+ [
+ "C",
+ "lick"
+ ],
+ [
+ "▁st",
+ "ories"
+ ],
+ [
+ "▁stor",
+ "ies"
+ ],
+ [
+ "▁sto",
+ "ries"
+ ],
+ [
+ "F",
+ "e"
+ ],
+ [
+ "å",
+ "r"
+ ],
+ [
+ "▁b",
+ "aby"
+ ],
+ [
+ "▁bab",
+ "y"
+ ],
+ [
+ "▁ba",
+ "by"
+ ],
+ [
+ "en",
+ "cia"
+ ],
+ [
+ "enc",
+ "ia"
+ ],
+ [
+ "enci",
+ "a"
+ ],
+ [
+ "e",
+ "ncia"
+ ],
+ [
+ "▁e",
+ "iner"
+ ],
+ [
+ "▁ein",
+ "er"
+ ],
+ [
+ "▁eine",
+ "r"
+ ],
+ [
+ "Ar",
+ "e"
+ ],
+ [
+ "A",
+ "re"
+ ],
+ [
+ "eb",
+ "ug"
+ ],
+ [
+ "e",
+ "bug"
+ ],
+ [
+ "st",
+ "ore"
+ ],
+ [
+ "sto",
+ "re"
+ ],
+ [
+ "\",",
+ "\""
+ ],
+ [
+ "\"",
+ ",\""
+ ],
+ [
+ "la",
+ "m"
+ ],
+ [
+ "l",
+ "am"
+ ],
+ [
+ "▁s",
+ "v"
+ ],
+ [
+ "▁",
+ "sv"
+ ],
+ [
+ "ци",
+ "и"
+ ],
+ [
+ "NU",
+ "LL"
+ ],
+ [
+ "N",
+ "ULL"
+ ],
+ [
+ "▁L",
+ "eg"
+ ],
+ [
+ "▁Le",
+ "g"
+ ],
+ [
+ "▁",
+ "Leg"
+ ],
+ [
+ "▁m",
+ "ovie"
+ ],
+ [
+ "▁mov",
+ "ie"
+ ],
+ [
+ "▁h",
+ "ous"
+ ],
+ [
+ "▁ho",
+ "us"
+ ],
+ [
+ "▁learn",
+ "ed"
+ ],
+ [
+ "▁lear",
+ "ned"
+ ],
+ [
+ "bo",
+ "n"
+ ],
+ [
+ "b",
+ "on"
+ ],
+ [
+ "▁trans",
+ "fer"
+ ],
+ [
+ "▁",
+ "transfer"
+ ],
+ [
+ "iforn",
+ "ia"
+ ],
+ [
+ "ps",
+ "ilon"
+ ],
+ [
+ "psi",
+ "lon"
+ ],
+ [
+ "▁S",
+ "oft"
+ ],
+ [
+ "▁So",
+ "ft"
+ ],
+ [
+ "▁Sof",
+ "t"
+ ],
+ [
+ "▁",
+ "Soft"
+ ],
+ [
+ "▁com",
+ "mer"
+ ],
+ [
+ "▁comm",
+ "er"
+ ],
+ [
+ "▁comme",
+ "r"
+ ],
+ [
+ "▁had",
+ "n"
+ ],
+ [
+ "▁ha",
+ "dn"
+ ],
+ [
+ "▁E",
+ "in"
+ ],
+ [
+ "▁T",
+ "wo"
+ ],
+ [
+ "▁Tw",
+ "o"
+ ],
+ [
+ "▁",
+ "Two"
+ ],
+ [
+ "cr",
+ "aft"
+ ],
+ [
+ "c",
+ "raft"
+ ],
+ [
+ "Pro",
+ "cess"
+ ],
+ [
+ "Proc",
+ "ess"
+ ],
+ [
+ "▁по",
+ "д"
+ ],
+ [
+ "ar",
+ "gin"
+ ],
+ [
+ "arg",
+ "in"
+ ],
+ [
+ "▁est",
+ "im"
+ ],
+ [
+ "▁es",
+ "tim"
+ ],
+ [
+ "▁M",
+ "em"
+ ],
+ [
+ "▁Me",
+ "m"
+ ],
+ [
+ "▁",
+ "Mem"
+ ],
+ [
+ "ik",
+ "a"
+ ],
+ [
+ "i",
+ "ka"
+ ],
+ [
+ "▁T",
+ "od"
+ ],
+ [
+ "▁To",
+ "d"
+ ],
+ [
+ "du",
+ "c"
+ ],
+ [
+ "d",
+ "uc"
+ ],
+ [
+ "▁d",
+ "anger"
+ ],
+ [
+ "▁dan",
+ "ger"
+ ],
+ [
+ "ri",
+ "ve"
+ ],
+ [
+ "riv",
+ "e"
+ ],
+ [
+ "r",
+ "ive"
+ ],
+ [
+ "Do",
+ "n"
+ ],
+ [
+ "D",
+ "on"
+ ],
+ [
+ "▁Q",
+ "ue"
+ ],
+ [
+ "▁Qu",
+ "e"
+ ],
+ [
+ "▁",
+ "Que"
+ ],
+ [
+ "ha",
+ "l"
+ ],
+ [
+ "h",
+ "al"
+ ],
+ [
+ "▁m",
+ "m"
+ ],
+ [
+ "▁",
+ "mm"
+ ],
+ [
+ "▁S",
+ "ur"
+ ],
+ [
+ "▁Su",
+ "r"
+ ],
+ [
+ "▁",
+ "Sur"
+ ],
+ [
+ "Or",
+ "der"
+ ],
+ [
+ "Ord",
+ "er"
+ ],
+ [
+ "▁d",
+ "istribution"
+ ],
+ [
+ "▁distribut",
+ "ion"
+ ],
+ [
+ "f",
+ "a"
+ ],
+ [
+ "▁M",
+ "any"
+ ],
+ [
+ "▁Man",
+ "y"
+ ],
+ [
+ "▁Ma",
+ "ny"
+ ],
+ [
+ "▁",
+ "Many"
+ ],
+ [
+ "pl",
+ "icit"
+ ],
+ [
+ "plic",
+ "it"
+ ],
+ [
+ "Em",
+ "pty"
+ ],
+ [
+ "Emp",
+ "ty"
+ ],
+ [
+ "Hand",
+ "le"
+ ],
+ [
+ "▁t",
+ "oken"
+ ],
+ [
+ "▁to",
+ "ken"
+ ],
+ [
+ "▁tok",
+ "en"
+ ],
+ [
+ "▁",
+ "token"
+ ],
+ [
+ "▁e",
+ "pis"
+ ],
+ [
+ "▁ep",
+ "is"
+ ],
+ [
+ "▁ass",
+ "ist"
+ ],
+ [
+ "▁pur",
+ "pose"
+ ],
+ [
+ "▁",
+ "ц"
+ ],
+ [
+ "N",
+ "U"
+ ],
+ [
+ "id",
+ "ers"
+ ],
+ [
+ "ide",
+ "rs"
+ ],
+ [
+ "ider",
+ "s"
+ ],
+ [
+ "i",
+ "ders"
+ ],
+ [
+ "ra",
+ "te"
+ ],
+ [
+ "rat",
+ "e"
+ ],
+ [
+ "r",
+ "ate"
+ ],
+ [
+ "The",
+ "y"
+ ],
+ [
+ "Th",
+ "ey"
+ ],
+ [
+ "Param",
+ "eter"
+ ],
+ [
+ "De",
+ "c"
+ ],
+ [
+ "D",
+ "ec"
+ ],
+ [
+ "▁str",
+ "ugg"
+ ],
+ [
+ "▁stru",
+ "gg"
+ ],
+ [
+ "▁sh",
+ "oot"
+ ],
+ [
+ "I",
+ "V"
+ ],
+ [
+ "▁G",
+ "reat"
+ ],
+ [
+ "▁Gre",
+ "at"
+ ],
+ [
+ "▁",
+ "Great"
+ ],
+ [
+ "▁S",
+ "il"
+ ],
+ [
+ "▁Si",
+ "l"
+ ],
+ [
+ "▁",
+ "Sil"
+ ],
+ [
+ "▁l",
+ "oved"
+ ],
+ [
+ "▁lo",
+ "ved"
+ ],
+ [
+ "▁love",
+ "d"
+ ],
+ [
+ "▁lov",
+ "ed"
+ ],
+ [
+ "▁c",
+ "lick"
+ ],
+ [
+ "▁cl",
+ "ick"
+ ],
+ [
+ "▁",
+ "click"
+ ],
+ [
+ "▁re",
+ "serv"
+ ],
+ [
+ "▁res",
+ "erv"
+ ],
+ [
+ "▁в",
+ "е"
+ ],
+ [
+ "▁",
+ "ве"
+ ],
+ [
+ "▁s",
+ "pread"
+ ],
+ [
+ "▁sp",
+ "read"
+ ],
+ [
+ "▁spr",
+ "ead"
+ ],
+ [
+ "▁o",
+ "g"
+ ],
+ [
+ "▁",
+ "og"
+ ],
+ [
+ "▁$",
+ "{"
+ ],
+ [
+ "▁",
+ "${"
+ ],
+ [
+ "▁m",
+ "iles"
+ ],
+ [
+ "▁mil",
+ "es"
+ ],
+ [
+ "▁mi",
+ "les"
+ ],
+ [
+ "▁mile",
+ "s"
+ ],
+ [
+ "▁success",
+ "ful"
+ ],
+ [
+ "▁",
+ "successful"
+ ],
+ [
+ "o",
+ "j"
+ ],
+ [
+ "▁D",
+ "irect"
+ ],
+ [
+ "▁Di",
+ "rect"
+ ],
+ [
+ "▁Dire",
+ "ct"
+ ],
+ [
+ "▁Dir",
+ "ect"
+ ],
+ [
+ "▁",
+ "Direct"
+ ],
+ [
+ "▁a",
+ "x"
+ ],
+ [
+ "▁",
+ "ax"
+ ],
+ [
+ "▁grow",
+ "th"
+ ],
+ [
+ "W",
+ "ork"
+ ],
+ [
+ "▁ch",
+ "urch"
+ ],
+ [
+ "In",
+ "st"
+ ],
+ [
+ "Ins",
+ "t"
+ ],
+ [
+ "IC",
+ "E"
+ ],
+ [
+ "I",
+ "CE"
+ ],
+ [
+ "st",
+ "en"
+ ],
+ [
+ "ste",
+ "n"
+ ],
+ [
+ "s",
+ "ten"
+ ],
+ [
+ "ро",
+ "д"
+ ],
+ [
+ "▁C",
+ "enter"
+ ],
+ [
+ "▁Cent",
+ "er"
+ ],
+ [
+ "▁",
+ "Center"
+ ],
+ [
+ "se",
+ "s"
+ ],
+ [
+ "s",
+ "es"
+ ],
+ [
+ "go",
+ "t"
+ ],
+ [
+ "g",
+ "ot"
+ ],
+ [
+ "de",
+ "lete"
+ ],
+ [
+ "del",
+ "ete"
+ ],
+ [
+ "▁M",
+ "a"
+ ],
+ [
+ "▁",
+ "Ma"
+ ],
+ [
+ "%",
+ "%"
+ ],
+ [
+ "▁c",
+ "row"
+ ],
+ [
+ "▁cr",
+ "ow"
+ ],
+ [
+ "▁cro",
+ "w"
+ ],
+ [
+ "D",
+ "F"
+ ],
+ [
+ "fr",
+ "ont"
+ ],
+ [
+ "▁b",
+ "log"
+ ],
+ [
+ "▁bl",
+ "og"
+ ],
+ [
+ "▁blo",
+ "g"
+ ],
+ [
+ "▁",
+ "blog"
+ ],
+ [
+ "▁comp",
+ "uter"
+ ],
+ [
+ "▁comput",
+ "er"
+ ],
+ [
+ "▁compute",
+ "r"
+ ],
+ [
+ "на",
+ "я"
+ ],
+ [
+ "▁m",
+ "ir"
+ ],
+ [
+ "▁mi",
+ "r"
+ ],
+ [
+ "▁",
+ "mir"
+ ],
+ [
+ "▁S",
+ "uper"
+ ],
+ [
+ "▁Su",
+ "per"
+ ],
+ [
+ "▁Sup",
+ "er"
+ ],
+ [
+ "▁",
+ "Super"
+ ],
+ [
+ "',",
+ "'"
+ ],
+ [
+ "'",
+ ",'"
+ ],
+ [
+ "▁mult",
+ "i"
+ ],
+ [
+ "▁mul",
+ "ti"
+ ],
+ [
+ "▁",
+ "multi"
+ ],
+ [
+ "▁g",
+ "ru"
+ ],
+ [
+ "▁gr",
+ "u"
+ ],
+ [
+ "▁",
+ "gru"
+ ],
+ [
+ "▁J",
+ "o"
+ ],
+ [
+ "▁",
+ "Jo"
+ ],
+ [
+ "▁Can",
+ "ada"
+ ],
+ [
+ "▁Canad",
+ "a"
+ ],
+ [
+ "▁Th",
+ "omas"
+ ],
+ [
+ "▁Thom",
+ "as"
+ ],
+ [
+ "▁large",
+ "r"
+ ],
+ [
+ "▁larg",
+ "er"
+ ],
+ [
+ "▁com",
+ "par"
+ ],
+ [
+ "▁comp",
+ "ar"
+ ],
+ [
+ "▁",
+ "compar"
+ ],
+ [
+ "Cur",
+ "rent"
+ ],
+ [
+ "th",
+ "at"
+ ],
+ [
+ "tha",
+ "t"
+ ],
+ [
+ "t",
+ "hat"
+ ],
+ [
+ "▁d",
+ "rop"
+ ],
+ [
+ "▁dr",
+ "op"
+ ],
+ [
+ "▁dro",
+ "p"
+ ],
+ [
+ "▁",
+ "drop"
+ ],
+ [
+ "ен",
+ "т"
+ ],
+ [
+ "▁Re",
+ "public"
+ ],
+ [
+ "▁Rep",
+ "ublic"
+ ],
+ [
+ "▁Repub",
+ "lic"
+ ],
+ [
+ "▁d",
+ "ise"
+ ],
+ [
+ "▁dis",
+ "e"
+ ],
+ [
+ "▁di",
+ "se"
+ ],
+ [
+ "▁effect",
+ "s"
+ ],
+ [
+ "▁girl",
+ "s"
+ ],
+ [
+ "▁gir",
+ "ls"
+ ],
+ [
+ "en",
+ "cies"
+ ],
+ [
+ "enc",
+ "ies"
+ ],
+ [
+ "enci",
+ "es"
+ ],
+ [
+ "el",
+ "lig"
+ ],
+ [
+ "ell",
+ "ig"
+ ],
+ [
+ "elli",
+ "g"
+ ],
+ [
+ "▁N",
+ "ote"
+ ],
+ [
+ "▁No",
+ "te"
+ ],
+ [
+ "▁Not",
+ "e"
+ ],
+ [
+ "▁",
+ "Note"
+ ],
+ [
+ "▁Ass",
+ "oci"
+ ],
+ [
+ "▁",
+ "Associ"
+ ],
+ [
+ "▁u",
+ "ses"
+ ],
+ [
+ "▁us",
+ "es"
+ ],
+ [
+ "▁use",
+ "s"
+ ],
+ [
+ "▁",
+ "uses"
+ ],
+ [
+ "el",
+ "led"
+ ],
+ [
+ "ell",
+ "ed"
+ ],
+ [
+ "elle",
+ "d"
+ ],
+ [
+ "▁w",
+ "arm"
+ ],
+ [
+ "▁war",
+ "m"
+ ],
+ [
+ "▁wa",
+ "rm"
+ ],
+ [
+ "th",
+ "read"
+ ],
+ [
+ "fo",
+ "nt"
+ ],
+ [
+ "fon",
+ "t"
+ ],
+ [
+ "f",
+ "ont"
+ ],
+ [
+ "▁z",
+ "um"
+ ],
+ [
+ "▁zu",
+ "m"
+ ],
+ [
+ "▁follow",
+ "s"
+ ],
+ [
+ "▁w",
+ "hom"
+ ],
+ [
+ "▁wh",
+ "om"
+ ],
+ [
+ "▁who",
+ "m"
+ ],
+ [
+ "T",
+ "A"
+ ],
+ [
+ "▁w",
+ "ild"
+ ],
+ [
+ "▁A",
+ "R"
+ ],
+ [
+ "▁",
+ "AR"
+ ],
+ [
+ "ia",
+ "ble"
+ ],
+ [
+ "i",
+ "able"
+ ],
+ [
+ "▁Tr",
+ "ue"
+ ],
+ [
+ "▁Tru",
+ "e"
+ ],
+ [
+ "▁",
+ "True"
+ ],
+ [
+ "Pos",
+ "ition"
+ ],
+ [
+ "▁s",
+ "ell"
+ ],
+ [
+ "▁se",
+ "ll"
+ ],
+ [
+ "▁sel",
+ "l"
+ ],
+ [
+ "ch",
+ "er"
+ ],
+ [
+ "che",
+ "r"
+ ],
+ [
+ "c",
+ "her"
+ ],
+ [
+ "▁B",
+ "us"
+ ],
+ [
+ "▁Bu",
+ "s"
+ ],
+ [
+ "▁",
+ "Bus"
+ ],
+ [
+ "▁le",
+ "an"
+ ],
+ [
+ "▁",
+ "lean"
+ ],
+ [
+ "AC",
+ "E"
+ ],
+ [
+ "A",
+ "CE"
+ ],
+ [
+ "▁s",
+ "erved"
+ ],
+ [
+ "▁ser",
+ "ved"
+ ],
+ [
+ "▁serv",
+ "ed"
+ ],
+ [
+ "▁serve",
+ "d"
+ ],
+ [
+ "h",
+ "w"
+ ],
+ [
+ "▁C",
+ "ur"
+ ],
+ [
+ "▁Cu",
+ "r"
+ ],
+ [
+ "▁",
+ "Cur"
+ ],
+ [
+ "▁n",
+ "orth"
+ ],
+ [
+ "▁nor",
+ "th"
+ ],
+ [
+ "▁nort",
+ "h"
+ ],
+ [
+ "Da",
+ "t"
+ ],
+ [
+ "D",
+ "at"
+ ],
+ [
+ "▁>",
+ ">"
+ ],
+ [
+ "▁",
+ ">>"
+ ],
+ [
+ "com",
+ "mand"
+ ],
+ [
+ "comm",
+ "and"
+ ],
+ [
+ "at",
+ "z"
+ ],
+ [
+ "a",
+ "tz"
+ ],
+ [
+ "▁m",
+ "al"
+ ],
+ [
+ "▁ma",
+ "l"
+ ],
+ [
+ "▁",
+ "mal"
+ ],
+ [
+ "ста",
+ "в"
+ ],
+ [
+ "▁P",
+ "ress"
+ ],
+ [
+ "▁Pr",
+ "ess"
+ ],
+ [
+ "▁Pres",
+ "s"
+ ],
+ [
+ "▁Pre",
+ "ss"
+ ],
+ [
+ "▁",
+ "Press"
+ ],
+ [
+ "▁char",
+ "acters"
+ ],
+ [
+ "▁character",
+ "s"
+ ],
+ [
+ "▁z",
+ "ero"
+ ],
+ [
+ "▁ze",
+ "ro"
+ ],
+ [
+ "▁",
+ "zero"
+ ],
+ [
+ "AG",
+ "E"
+ ],
+ [
+ "A",
+ "GE"
+ ],
+ [
+ "rap",
+ "per"
+ ],
+ [
+ "▁kit",
+ "chen"
+ ],
+ [
+ "am",
+ "ing"
+ ],
+ [
+ "ami",
+ "ng"
+ ],
+ [
+ "amin",
+ "g"
+ ],
+ [
+ "a",
+ "ming"
+ ],
+ [
+ "▁re",
+ "str"
+ ],
+ [
+ "▁r",
+ "estr"
+ ],
+ [
+ "▁res",
+ "tr"
+ ],
+ [
+ "▁rest",
+ "r"
+ ],
+ [
+ "X",
+ "X"
+ ],
+ [
+ "▁Col",
+ "lege"
+ ],
+ [
+ "▁Ar",
+ "ray"
+ ],
+ [
+ "▁Arr",
+ "ay"
+ ],
+ [
+ "▁",
+ "Array"
+ ],
+ [
+ "▁f",
+ "resh"
+ ],
+ [
+ "▁fr",
+ "esh"
+ ],
+ [
+ "▁fre",
+ "sh"
+ ],
+ [
+ "▁fres",
+ "h"
+ ],
+ [
+ "▁sh",
+ "ift"
+ ],
+ [
+ "▁",
+ "shift"
+ ],
+ [
+ "▁spec",
+ "ified"
+ ],
+ [
+ "pl",
+ "ete"
+ ],
+ [
+ "ple",
+ "te"
+ ],
+ [
+ "plet",
+ "e"
+ ],
+ [
+ "p",
+ "lete"
+ ],
+ [
+ "IT",
+ "E"
+ ],
+ [
+ "I",
+ "TE"
+ ],
+ [
+ "▁C",
+ "amp"
+ ],
+ [
+ "▁Cam",
+ "p"
+ ],
+ [
+ "▁Ca",
+ "mp"
+ ],
+ [
+ "▁",
+ "Camp"
+ ],
+ [
+ "ri",
+ "al"
+ ],
+ [
+ "ria",
+ "l"
+ ],
+ [
+ "r",
+ "ial"
+ ],
+ [
+ "c",
+ "b"
+ ],
+ [
+ "▁T",
+ "H"
+ ],
+ [
+ "▁",
+ "TH"
+ ],
+ [
+ "I",
+ "B"
+ ],
+ [
+ "os",
+ "en"
+ ],
+ [
+ "ose",
+ "n"
+ ],
+ [
+ "o",
+ "sen"
+ ],
+ [
+ "▁",
+ "ú"
+ ],
+ [
+ "▁par",
+ "ams"
+ ],
+ [
+ "▁param",
+ "s"
+ ],
+ [
+ "▁para",
+ "ms"
+ ],
+ [
+ "▁",
+ "params"
+ ],
+ [
+ "ign",
+ "ment"
+ ],
+ [
+ "ad",
+ "ding"
+ ],
+ [
+ "add",
+ "ing"
+ ],
+ [
+ "▁deg",
+ "ree"
+ ],
+ [
+ "▁",
+ "degree"
+ ],
+ [
+ "Loc",
+ "al"
+ ],
+ [
+ "Lo",
+ "cal"
+ ],
+ [
+ "L",
+ "ocal"
+ ],
+ [
+ "O",
+ "h"
+ ],
+ [
+ "▁z",
+ "ur"
+ ],
+ [
+ "▁zu",
+ "r"
+ ],
+ [
+ "▁level",
+ "s"
+ ],
+ [
+ "▁lev",
+ "els"
+ ],
+ [
+ "C",
+ "S"
+ ],
+ [
+ "fin",
+ "ished"
+ ],
+ [
+ "finish",
+ "ed"
+ ],
+ [
+ "C",
+ "ase"
+ ],
+ [
+ "ri",
+ "age"
+ ],
+ [
+ "ria",
+ "ge"
+ ],
+ [
+ "Vec",
+ "tor"
+ ],
+ [
+ "V",
+ "ector"
+ ],
+ [
+ "▁s",
+ "ea"
+ ],
+ [
+ "▁se",
+ "a"
+ ],
+ [
+ "▁",
+ "sea"
+ ],
+ [
+ "ant",
+ "ic"
+ ],
+ [
+ "anti",
+ "c"
+ ],
+ [
+ "▁Le",
+ "ague"
+ ],
+ [
+ "▁there",
+ "fore"
+ ],
+ [
+ "▁ther",
+ "efore"
+ ],
+ [
+ "On",
+ "e"
+ ],
+ [
+ "O",
+ "ne"
+ ],
+ [
+ "Re",
+ "turn"
+ ],
+ [
+ "Ret",
+ "urn"
+ ],
+ [
+ "R",
+ "eturn"
+ ],
+ [
+ "Acc",
+ "ess"
+ ],
+ [
+ "Ac",
+ "cess"
+ ],
+ [
+ "A",
+ "ccess"
+ ],
+ [
+ "va",
+ "s"
+ ],
+ [
+ "v",
+ "as"
+ ],
+ [
+ "▁о",
+ "с"
+ ],
+ [
+ "▁r",
+ "at"
+ ],
+ [
+ "▁ra",
+ "t"
+ ],
+ [
+ "▁",
+ "rat"
+ ],
+ [
+ "Bi",
+ "g"
+ ],
+ [
+ "B",
+ "ig"
+ ],
+ [
+ "▁be",
+ "havior"
+ ],
+ [
+ "▁behav",
+ "ior"
+ ],
+ [
+ "▁behavi",
+ "or"
+ ],
+ [
+ "k",
+ "r"
+ ],
+ [
+ "▁un",
+ "defined"
+ ],
+ [
+ "▁und",
+ "efined"
+ ],
+ [
+ "▁",
+ "undefined"
+ ],
+ [
+ "▁E",
+ "s"
+ ],
+ [
+ "▁",
+ "Es"
+ ],
+ [
+ "▁appe",
+ "ared"
+ ],
+ [
+ "▁appear",
+ "ed"
+ ],
+ [
+ "el",
+ "es"
+ ],
+ [
+ "ele",
+ "s"
+ ],
+ [
+ "e",
+ "les"
+ ],
+ [
+ "▁W",
+ "AR"
+ ],
+ [
+ "▁WA",
+ "R"
+ ],
+ [
+ "▁",
+ "WAR"
+ ],
+ [
+ "St",
+ "at"
+ ],
+ [
+ "S",
+ "tat"
+ ],
+ [
+ "▁Go",
+ "ogle"
+ ],
+ [
+ "▁",
+ "Google"
+ ],
+ [
+ "▁c",
+ "redit"
+ ],
+ [
+ "▁cre",
+ "dit"
+ ],
+ [
+ "▁cr",
+ "edit"
+ ],
+ [
+ "▁cred",
+ "it"
+ ],
+ [
+ "▁F",
+ "ile"
+ ],
+ [
+ "▁Fil",
+ "e"
+ ],
+ [
+ "▁Fi",
+ "le"
+ ],
+ [
+ "▁",
+ "File"
+ ],
+ [
+ "an",
+ "ging"
+ ],
+ [
+ "ang",
+ "ing"
+ ],
+ [
+ "ho",
+ "use"
+ ],
+ [
+ "hou",
+ "se"
+ ],
+ [
+ "h",
+ "ouse"
+ ],
+ [
+ "rom",
+ "ise"
+ ],
+ [
+ "ge",
+ "nt"
+ ],
+ [
+ "gen",
+ "t"
+ ],
+ [
+ "g",
+ "ent"
+ ],
+ [
+ "▁hab",
+ "it"
+ ],
+ [
+ "▁ha",
+ "bit"
+ ],
+ [
+ "▁soc",
+ "iety"
+ ],
+ [
+ "▁soci",
+ "ety"
+ ],
+ [
+ "▁societ",
+ "y"
+ ],
+ [
+ "▁enc",
+ "our"
+ ],
+ [
+ "▁p",
+ "aint"
+ ],
+ [
+ "▁pain",
+ "t"
+ ],
+ [
+ "▁pa",
+ "int"
+ ],
+ [
+ "pe",
+ "t"
+ ],
+ [
+ "p",
+ "et"
+ ],
+ [
+ "▁U",
+ "K"
+ ],
+ [
+ "▁",
+ "UK"
+ ],
+ [
+ "aw",
+ "s"
+ ],
+ [
+ "a",
+ "ws"
+ ],
+ [
+ "on",
+ "om"
+ ],
+ [
+ "ono",
+ "m"
+ ],
+ [
+ "o",
+ "nom"
+ ],
+ [
+ "G",
+ "l"
+ ],
+ [
+ "}_",
+ "{\\"
+ ],
+ [
+ "}_{",
+ "\\"
+ ],
+ [
+ "}",
+ "_{\\"
+ ],
+ [
+ "el",
+ "ess"
+ ],
+ [
+ "ele",
+ "ss"
+ ],
+ [
+ "eles",
+ "s"
+ ],
+ [
+ "e",
+ "less"
+ ],
+ [
+ "em",
+ "y"
+ ],
+ [
+ "e",
+ "my"
+ ],
+ [
+ "▁C",
+ "ong"
+ ],
+ [
+ "▁Con",
+ "g"
+ ],
+ [
+ "▁Co",
+ "ng"
+ ],
+ [
+ "▁develop",
+ "ed"
+ ],
+ [
+ "▁im",
+ "ages"
+ ],
+ [
+ "▁image",
+ "s"
+ ],
+ [
+ "▁imag",
+ "es"
+ ],
+ [
+ "▁",
+ "images"
+ ],
+ [
+ "▁",
+ "ö"
+ ],
+ [
+ "▁f",
+ "ont"
+ ],
+ [
+ "▁fo",
+ "nt"
+ ],
+ [
+ "▁fon",
+ "t"
+ ],
+ [
+ "▁",
+ "font"
+ ],
+ [
+ "cl",
+ "ear"
+ ],
+ [
+ "cle",
+ "ar"
+ ],
+ [
+ "c",
+ "lear"
+ ],
+ [
+ "gi",
+ "n"
+ ],
+ [
+ "g",
+ "in"
+ ],
+ [
+ "▁L",
+ "ord"
+ ],
+ [
+ "▁Lo",
+ "rd"
+ ],
+ [
+ "▁Lor",
+ "d"
+ ],
+ [
+ "▁trans",
+ "port"
+ ],
+ [
+ "▁",
+ "transport"
+ ],
+ [
+ "▁:",
+ ":"
+ ],
+ [
+ "▁",
+ "::"
+ ],
+ [
+ "▁c",
+ "up"
+ ],
+ [
+ "▁cu",
+ "p"
+ ],
+ [
+ "▁",
+ "cup"
+ ],
+ [
+ "ul",
+ "ate"
+ ],
+ [
+ "ula",
+ "te"
+ ],
+ [
+ "u",
+ "late"
+ ],
+ [
+ "▁D",
+ "uring"
+ ],
+ [
+ "▁Du",
+ "ring"
+ ],
+ [
+ "▁Dur",
+ "ing"
+ ],
+ [
+ "pr",
+ "iv"
+ ],
+ [
+ "p",
+ "riv"
+ ],
+ [
+ "▁ext",
+ "rem"
+ ],
+ [
+ "▁extr",
+ "em"
+ ],
+ [
+ "▁D",
+ "i"
+ ],
+ [
+ "▁",
+ "Di"
+ ],
+ [
+ "▁d",
+ "oubt"
+ ],
+ [
+ "▁dou",
+ "bt"
+ ],
+ [
+ "▁doub",
+ "t"
+ ],
+ [
+ "P",
+ "y"
+ ],
+ [
+ "if",
+ "ying"
+ ],
+ [
+ "ify",
+ "ing"
+ ],
+ [
+ "sp",
+ "lit"
+ ],
+ [
+ "spl",
+ "it"
+ ],
+ [
+ "s",
+ "plit"
+ ],
+ [
+ "eg",
+ "o"
+ ],
+ [
+ "e",
+ "go"
+ ],
+ [
+ "git",
+ "hub"
+ ],
+ [
+ "g",
+ "ithub"
+ ],
+ [
+ "▁)",
+ ","
+ ],
+ [
+ "▁",
+ "),"
+ ],
+ [
+ "RO",
+ "M"
+ ],
+ [
+ "R",
+ "OM"
+ ],
+ [
+ "▁ch",
+ "air"
+ ],
+ [
+ "▁cha",
+ "ir"
+ ],
+ [
+ "▁",
+ "chair"
+ ],
+ [
+ "▁t",
+ "rade"
+ ],
+ [
+ "▁tr",
+ "ade"
+ ],
+ [
+ "▁trad",
+ "e"
+ ],
+ [
+ "▁tra",
+ "de"
+ ],
+ [
+ "▁n",
+ "icht"
+ ],
+ [
+ "▁ni",
+ "cht"
+ ],
+ [
+ "▁nic",
+ "ht"
+ ],
+ [
+ "To",
+ "p"
+ ],
+ [
+ "T",
+ "op"
+ ],
+ [
+ "St",
+ "ore"
+ ],
+ [
+ "▁p",
+ "arte"
+ ],
+ [
+ "▁part",
+ "e"
+ ],
+ [
+ "▁par",
+ "te"
+ ],
+ [
+ "pro",
+ "ject"
+ ],
+ [
+ "ni",
+ "a"
+ ],
+ [
+ "n",
+ "ia"
+ ],
+ [
+ "▁в",
+ "ід"
+ ],
+ [
+ "▁ві",
+ "д"
+ ],
+ [
+ "wa",
+ "r"
+ ],
+ [
+ "w",
+ "ar"
+ ],
+ [
+ "▁Pro",
+ "f"
+ ],
+ [
+ "▁Pr",
+ "of"
+ ],
+ [
+ "▁c",
+ "aught"
+ ],
+ [
+ "Th",
+ "read"
+ ],
+ [
+ "ст",
+ "ва"
+ ],
+ [
+ "ств",
+ "а"
+ ],
+ [
+ "с",
+ "тва"
+ ],
+ [
+ "aut",
+ "hor"
+ ],
+ [
+ "auth",
+ "or"
+ ],
+ [
+ "▁d",
+ "oll"
+ ],
+ [
+ "▁do",
+ "ll"
+ ],
+ [
+ "▁dol",
+ "l"
+ ],
+ [
+ "▁h",
+ "arm"
+ ],
+ [
+ "▁ha",
+ "rm"
+ ],
+ [
+ "▁har",
+ "m"
+ ],
+ [
+ "▁",
+ "harm"
+ ],
+ [
+ "▁G",
+ "en"
+ ],
+ [
+ "▁Ge",
+ "n"
+ ],
+ [
+ "▁",
+ "Gen"
+ ],
+ [
+ "tr",
+ "ee"
+ ],
+ [
+ "tre",
+ "e"
+ ],
+ [
+ "t",
+ "ree"
+ ],
+ [
+ "et",
+ "ime"
+ ],
+ [
+ "eti",
+ "me"
+ ],
+ [
+ "e",
+ "time"
+ ],
+ [
+ "cf",
+ "g"
+ ],
+ [
+ "c",
+ "fg"
+ ],
+ [
+ "▁gu",
+ "ys"
+ ],
+ [
+ "▁guy",
+ "s"
+ ],
+ [
+ "▁Cal",
+ "ifornia"
+ ],
+ [
+ "▁G",
+ "reen"
+ ],
+ [
+ "▁Gr",
+ "een"
+ ],
+ [
+ "▁Gre",
+ "en"
+ ],
+ [
+ "▁Gree",
+ "n"
+ ],
+ [
+ "▁",
+ "Green"
+ ],
+ [
+ "▁mov",
+ "ement"
+ ],
+ [
+ "▁move",
+ "ment"
+ ],
+ [
+ "▁mo",
+ "vement"
+ ],
+ [
+ "ie",
+ "j"
+ ],
+ [
+ "i",
+ "ej"
+ ],
+ [
+ "▁stat",
+ "ement"
+ ],
+ [
+ "▁state",
+ "ment"
+ ],
+ [
+ "▁",
+ "statement"
+ ],
+ [
+ "▁se",
+ "eing"
+ ],
+ [
+ "▁see",
+ "ing"
+ ],
+ [
+ "▁h",
+ "aven"
+ ],
+ [
+ "▁have",
+ "n"
+ ],
+ [
+ "▁ha",
+ "ven"
+ ],
+ [
+ "▁hav",
+ "en"
+ ],
+ [
+ "vent",
+ "ion"
+ ],
+ [
+ "v",
+ "ention"
+ ],
+ [
+ "S",
+ "L"
+ ],
+ [
+ "ched",
+ "ul"
+ ],
+ [
+ "ie",
+ "rt"
+ ],
+ [
+ "ier",
+ "t"
+ ],
+ [
+ "i",
+ "ert"
+ ],
+ [
+ "▁pr",
+ "imary"
+ ],
+ [
+ "▁prim",
+ "ary"
+ ],
+ [
+ "▁pri",
+ "mary"
+ ],
+ [
+ "▁prima",
+ "ry"
+ ],
+ [
+ "▁",
+ "primary"
+ ],
+ [
+ "▁c",
+ "ivil"
+ ],
+ [
+ "▁ci",
+ "vil"
+ ],
+ [
+ "▁civ",
+ "il"
+ ],
+ [
+ "ri",
+ "an"
+ ],
+ [
+ "ria",
+ "n"
+ ],
+ [
+ "r",
+ "ian"
+ ],
+ [
+ "▁b",
+ "utton"
+ ],
+ [
+ "▁but",
+ "ton"
+ ],
+ [
+ "▁butt",
+ "on"
+ ],
+ [
+ "▁",
+ "button"
+ ],
+ [
+ "▁l",
+ "ived"
+ ],
+ [
+ "▁li",
+ "ved"
+ ],
+ [
+ "▁live",
+ "d"
+ ],
+ [
+ "▁liv",
+ "ed"
+ ],
+ [
+ "P",
+ "ass"
+ ],
+ [
+ "so",
+ "r"
+ ],
+ [
+ "s",
+ "or"
+ ],
+ [
+ "▁watch",
+ "ing"
+ ],
+ [
+ "▁wat",
+ "ching"
+ ],
+ [
+ "▁sk",
+ "ills"
+ ],
+ [
+ "▁skill",
+ "s"
+ ],
+ [
+ "te",
+ "e"
+ ],
+ [
+ "t",
+ "ee"
+ ],
+ [
+ "Le",
+ "vel"
+ ],
+ [
+ "L",
+ "evel"
+ ],
+ [
+ "▁sc",
+ "ient"
+ ],
+ [
+ "h",
+ "s"
+ ],
+ [
+ "▁a",
+ "gre"
+ ],
+ [
+ "▁ag",
+ "re"
+ ],
+ [
+ "ca",
+ "t"
+ ],
+ [
+ "c",
+ "at"
+ ],
+ [
+ "▁t",
+ "end"
+ ],
+ [
+ "▁te",
+ "nd"
+ ],
+ [
+ "▁ten",
+ "d"
+ ],
+ [
+ "▁M",
+ "ill"
+ ],
+ [
+ "▁Mil",
+ "l"
+ ],
+ [
+ "▁Mi",
+ "ll"
+ ],
+ [
+ "▁",
+ "Mill"
+ ],
+ [
+ "▁C",
+ "ap"
+ ],
+ [
+ "▁Ca",
+ "p"
+ ],
+ [
+ "▁",
+ "Cap"
+ ],
+ [
+ "OR",
+ "D"
+ ],
+ [
+ "O",
+ "RD"
+ ],
+ [
+ "gl",
+ "e"
+ ],
+ [
+ "g",
+ "le"
+ ],
+ [
+ "▁с",
+ "во"
+ ],
+ [
+ "»",
+ ","
+ ],
+ [
+ "▁a",
+ "head"
+ ],
+ [
+ "▁ah",
+ "ead"
+ ],
+ [
+ "ve",
+ "st"
+ ],
+ [
+ "ves",
+ "t"
+ ],
+ [
+ "v",
+ "est"
+ ],
+ [
+ "▁J",
+ "ose"
+ ],
+ [
+ "▁Jo",
+ "se"
+ ],
+ [
+ "▁Jos",
+ "e"
+ ],
+ [
+ "is",
+ "cher"
+ ],
+ [
+ "isch",
+ "er"
+ ],
+ [
+ "ische",
+ "r"
+ ],
+ [
+ "isc",
+ "her"
+ ],
+ [
+ "ș",
+ "i"
+ ],
+ [
+ "▁le",
+ "aving"
+ ],
+ [
+ "▁д",
+ "ля"
+ ],
+ [
+ "▁s",
+ "outh"
+ ],
+ [
+ "▁so",
+ "uth"
+ ],
+ [
+ "▁sou",
+ "th"
+ ],
+ [
+ "▁sout",
+ "h"
+ ],
+ [
+ "▁con",
+ "sum"
+ ],
+ [
+ "▁cons",
+ "um"
+ ],
+ [
+ "▁",
+ "consum"
+ ],
+ [
+ "R",
+ "ange"
+ ],
+ [
+ "▁activ",
+ "ities"
+ ],
+ [
+ "Se",
+ "c"
+ ],
+ [
+ "S",
+ "ec"
+ ],
+ [
+ "▁s",
+ "ales"
+ ],
+ [
+ "▁sa",
+ "les"
+ ],
+ [
+ "▁sal",
+ "es"
+ ],
+ [
+ "▁sale",
+ "s"
+ ],
+ [
+ "▁f",
+ "ix"
+ ],
+ [
+ "▁fi",
+ "x"
+ ],
+ [
+ "▁",
+ "fix"
+ ],
+ [
+ "▁j",
+ "ed"
+ ],
+ [
+ "▁je",
+ "d"
+ ],
+ [
+ "▁",
+ "jed"
+ ],
+ [
+ "ru",
+ "m"
+ ],
+ [
+ "r",
+ "um"
+ ],
+ [
+ "ve",
+ "ctor"
+ ],
+ [
+ "vec",
+ "tor"
+ ],
+ [
+ "v",
+ "ector"
+ ],
+ [
+ "▁s",
+ "pot"
+ ],
+ [
+ "▁sp",
+ "ot"
+ ],
+ [
+ "▁spo",
+ "t"
+ ],
+ [
+ "▁",
+ "spot"
+ ],
+ [
+ "▁man",
+ "ufact"
+ ],
+ [
+ "к",
+ "т"
+ ],
+ [
+ "or",
+ "row"
+ ],
+ [
+ "orr",
+ "ow"
+ ],
+ [
+ "si",
+ "gn"
+ ],
+ [
+ "sig",
+ "n"
+ ],
+ [
+ "s",
+ "ign"
+ ],
+ [
+ "▁col",
+ "lege"
+ ],
+ [
+ "▁colle",
+ "ge"
+ ],
+ [
+ "▁colleg",
+ "e"
+ ],
+ [
+ "▁d",
+ "river"
+ ],
+ [
+ "▁dr",
+ "iver"
+ ],
+ [
+ "▁dri",
+ "ver"
+ ],
+ [
+ "▁driv",
+ "er"
+ ],
+ [
+ "▁drive",
+ "r"
+ ],
+ [
+ "▁",
+ "driver"
+ ],
+ [
+ "▁def",
+ "initely"
+ ],
+ [
+ "▁definit",
+ "ely"
+ ],
+ [
+ "▁s",
+ "pend"
+ ],
+ [
+ "▁sp",
+ "end"
+ ],
+ [
+ "▁spe",
+ "nd"
+ ],
+ [
+ "miss",
+ "ion"
+ ],
+ [
+ "m",
+ "ission"
+ ],
+ [
+ "з",
+ "у"
+ ],
+ [
+ "at",
+ "ively"
+ ],
+ [
+ "ative",
+ "ly"
+ ],
+ [
+ "ativ",
+ "ely"
+ ],
+ [
+ "b",
+ "i"
+ ],
+ [
+ "Call",
+ "back"
+ ],
+ [
+ "▁particular",
+ "ly"
+ ],
+ [
+ "▁particul",
+ "arly"
+ ],
+ [
+ "▁h",
+ "ell"
+ ],
+ [
+ "▁he",
+ "ll"
+ ],
+ [
+ "▁hel",
+ "l"
+ ],
+ [
+ "▁",
+ "hell"
+ ],
+ [
+ "▁p",
+ "ool"
+ ],
+ [
+ "▁po",
+ "ol"
+ ],
+ [
+ "▁",
+ "pool"
+ ],
+ [
+ "PR",
+ "E"
+ ],
+ [
+ "P",
+ "RE"
+ ],
+ [
+ "▁cle",
+ "arly"
+ ],
+ [
+ "▁clear",
+ "ly"
+ ],
+ [
+ "P",
+ "T"
+ ],
+ [
+ "ot",
+ "hes"
+ ],
+ [
+ "oth",
+ "es"
+ ],
+ [
+ "othe",
+ "s"
+ ],
+ [
+ "▁I",
+ "d"
+ ],
+ [
+ "▁",
+ "Id"
+ ],
+ [
+ "Loc",
+ "ation"
+ ],
+ [
+ "L",
+ "ocation"
+ ],
+ [
+ "▁R",
+ "un"
+ ],
+ [
+ "▁Ru",
+ "n"
+ ],
+ [
+ "▁",
+ "Run"
+ ],
+ [
+ "▁f",
+ "ixed"
+ ],
+ [
+ "▁fix",
+ "ed"
+ ],
+ [
+ "▁",
+ "fixed"
+ ],
+ [
+ "▁H",
+ "and"
+ ],
+ [
+ "▁Ha",
+ "nd"
+ ],
+ [
+ "▁Han",
+ "d"
+ ],
+ [
+ "▁",
+ "Hand"
+ ],
+ [
+ "ba",
+ "l"
+ ],
+ [
+ "b",
+ "al"
+ ],
+ [
+ "d",
+ "ouble"
+ ],
+ [
+ "C",
+ "an"
+ ],
+ [
+ "Om",
+ "ega"
+ ],
+ [
+ "▁chall",
+ "eng"
+ ],
+ [
+ "▁stand",
+ "ing"
+ ],
+ [
+ "▁stan",
+ "ding"
+ ],
+ [
+ "▁",
+ "standing"
+ ],
+ [
+ "it",
+ "en"
+ ],
+ [
+ "ite",
+ "n"
+ ],
+ [
+ "i",
+ "ten"
+ ],
+ [
+ "▁me",
+ "chan"
+ ],
+ [
+ "▁d",
+ "urch"
+ ],
+ [
+ "▁dur",
+ "ch"
+ ],
+ [
+ "▁d",
+ "ell"
+ ],
+ [
+ "▁de",
+ "ll"
+ ],
+ [
+ "▁del",
+ "l"
+ ],
+ [
+ "▁rais",
+ "ed"
+ ],
+ [
+ "▁raise",
+ "d"
+ ],
+ [
+ "▁ra",
+ "ised"
+ ],
+ [
+ "▁we",
+ "ak"
+ ],
+ [
+ "▁",
+ "weak"
+ ],
+ [
+ "▁D",
+ "u"
+ ],
+ [
+ "▁",
+ "Du"
+ ],
+ [
+ "gr",
+ "ad"
+ ],
+ [
+ "gra",
+ "d"
+ ],
+ [
+ "g",
+ "rad"
+ ],
+ [
+ "▁sc",
+ "ene"
+ ],
+ [
+ "▁scen",
+ "e"
+ ],
+ [
+ "▁",
+ "scene"
+ ],
+ [
+ "pos",
+ "s"
+ ],
+ [
+ "po",
+ "ss"
+ ],
+ [
+ "p",
+ "oss"
+ ],
+ [
+ "▁t",
+ "on"
+ ],
+ [
+ "▁to",
+ "n"
+ ],
+ [
+ "▁",
+ "ton"
+ ],
+ [
+ "▁e",
+ "arth"
+ ],
+ [
+ "▁ear",
+ "th"
+ ],
+ [
+ "ul",
+ "ations"
+ ],
+ [
+ "ulation",
+ "s"
+ ],
+ [
+ "▁str",
+ "ength"
+ ],
+ [
+ "▁stre",
+ "ngth"
+ ],
+ [
+ "▁streng",
+ "th"
+ ],
+ [
+ "ak",
+ "ed"
+ ],
+ [
+ "ake",
+ "d"
+ ],
+ [
+ "a",
+ "ked"
+ ],
+ [
+ "▁re",
+ "main"
+ ],
+ [
+ "▁rem",
+ "ain"
+ ],
+ [
+ "▁B",
+ "i"
+ ],
+ [
+ "▁",
+ "Bi"
+ ],
+ [
+ "▁custom",
+ "er"
+ ],
+ [
+ "▁cust",
+ "omer"
+ ],
+ [
+ "▁",
+ "customer"
+ ],
+ [
+ "ran",
+ "ge"
+ ],
+ [
+ "r",
+ "ange"
+ ],
+ [
+ "▁inter",
+ "ested"
+ ],
+ [
+ "▁interest",
+ "ed"
+ ],
+ [
+ "ON",
+ "E"
+ ],
+ [
+ "O",
+ "NE"
+ ],
+ [
+ "▁c",
+ "off"
+ ],
+ [
+ "▁co",
+ "ff"
+ ],
+ [
+ "re",
+ "quire"
+ ],
+ [
+ "requ",
+ "ire"
+ ],
+ [
+ "▁On",
+ "ly"
+ ],
+ [
+ "▁",
+ "Only"
+ ],
+ [
+ "▁W",
+ "eb"
+ ],
+ [
+ "▁We",
+ "b"
+ ],
+ [
+ "▁",
+ "Web"
+ ],
+ [
+ "▁f",
+ "arm"
+ ],
+ [
+ "▁far",
+ "m"
+ ],
+ [
+ "▁fa",
+ "rm"
+ ],
+ [
+ "▁act",
+ "ivity"
+ ],
+ [
+ "▁activ",
+ "ity"
+ ],
+ [
+ "▁",
+ "activity"
+ ],
+ [
+ "▁r",
+ "out"
+ ],
+ [
+ "▁ro",
+ "ut"
+ ],
+ [
+ "▁rou",
+ "t"
+ ],
+ [
+ "bl",
+ "ing"
+ ],
+ [
+ "b",
+ "ling"
+ ],
+ [
+ "S",
+ "Y"
+ ],
+ [
+ "▁Rich",
+ "ard"
+ ],
+ [
+ "▁Ric",
+ "hard"
+ ],
+ [
+ "▁R",
+ "ef"
+ ],
+ [
+ "▁Re",
+ "f"
+ ],
+ [
+ "▁",
+ "Ref"
+ ],
+ [
+ "▁ко",
+ "н"
+ ],
+ [
+ "▁к",
+ "он"
+ ],
+ [
+ "▁",
+ "кон"
+ ],
+ [
+ "▁j",
+ "un"
+ ],
+ [
+ "▁ju",
+ "n"
+ ],
+ [
+ "bo",
+ "rn"
+ ],
+ [
+ "bor",
+ "n"
+ ],
+ [
+ "b",
+ "orn"
+ ],
+ [
+ "ij",
+ "n"
+ ],
+ [
+ "Config",
+ "uration"
+ ],
+ [
+ "um",
+ "an"
+ ],
+ [
+ "uma",
+ "n"
+ ],
+ [
+ "u",
+ "man"
+ ],
+ [
+ "E",
+ "E"
+ ],
+ [
+ "▁mar",
+ "ried"
+ ],
+ [
+ "▁З",
+ "а"
+ ],
+ [
+ "▁",
+ "За"
+ ],
+ [
+ "▁f",
+ "at"
+ ],
+ [
+ "▁fa",
+ "t"
+ ],
+ [
+ "▁k",
+ "id"
+ ],
+ [
+ "▁ki",
+ "d"
+ ],
+ [
+ "▁T",
+ "ur"
+ ],
+ [
+ "▁Tu",
+ "r"
+ ],
+ [
+ "▁",
+ "Tur"
+ ],
+ [
+ "▁off",
+ "ered"
+ ],
+ [
+ "▁offer",
+ "ed"
+ ],
+ [
+ "ni",
+ "c"
+ ],
+ [
+ "n",
+ "ic"
+ ],
+ [
+ "▁B",
+ "ig"
+ ],
+ [
+ "▁Bi",
+ "g"
+ ],
+ [
+ "▁",
+ "Big"
+ ],
+ [
+ "Ga",
+ "mma"
+ ],
+ [
+ "G",
+ "amma"
+ ],
+ [
+ "▁He",
+ "alth"
+ ],
+ [
+ "▁",
+ "Health"
+ ],
+ [
+ "▁T",
+ "R"
+ ],
+ [
+ "▁",
+ "TR"
+ ],
+ [
+ "▁s",
+ "ię"
+ ],
+ [
+ "▁si",
+ "ę"
+ ],
+ [
+ "▁const",
+ "ruction"
+ ],
+ [
+ "▁construct",
+ "ion"
+ ],
+ [
+ "▁constr",
+ "uction"
+ ],
+ [
+ "▁constru",
+ "ction"
+ ],
+ [
+ "▁",
+ "construction"
+ ],
+ [
+ "▁Ch",
+ "urch"
+ ],
+ [
+ "▁B",
+ "et"
+ ],
+ [
+ "▁Be",
+ "t"
+ ],
+ [
+ "▁",
+ "Bet"
+ ],
+ [
+ "bu",
+ "s"
+ ],
+ [
+ "b",
+ "us"
+ ],
+ [
+ "▁e",
+ "arn"
+ ],
+ [
+ "▁ear",
+ "n"
+ ],
+ [
+ "ri",
+ "ct"
+ ],
+ [
+ "ric",
+ "t"
+ ],
+ [
+ "r",
+ "ict"
+ ],
+ [
+ "▁п",
+ "ра"
+ ],
+ [
+ "▁пр",
+ "а"
+ ],
+ [
+ "▁",
+ "пра"
+ ],
+ [
+ "▁br",
+ "ain"
+ ],
+ [
+ "▁bra",
+ "in"
+ ],
+ [
+ "▁f",
+ "ra"
+ ],
+ [
+ "▁fr",
+ "a"
+ ],
+ [
+ "▁O",
+ "p"
+ ],
+ [
+ "▁",
+ "Op"
+ ],
+ [
+ "FI",
+ "G"
+ ],
+ [
+ "F",
+ "IG"
+ ],
+ [
+ "em",
+ "a"
+ ],
+ [
+ "e",
+ "ma"
+ ],
+ [
+ "▁Europe",
+ "an"
+ ],
+ [
+ "▁S",
+ "aint"
+ ],
+ [
+ "▁Sa",
+ "int"
+ ],
+ [
+ "▁",
+ "Saint"
+ ],
+ [
+ "AR",
+ "E"
+ ],
+ [
+ "A",
+ "RE"
+ ],
+ [
+ "ur",
+ "i"
+ ],
+ [
+ "u",
+ "ri"
+ ],
+ [
+ "▁R",
+ "iver"
+ ],
+ [
+ "{",
+ "}"
+ ],
+ [
+ "▁s",
+ "itting"
+ ],
+ [
+ "▁sit",
+ "ting"
+ ],
+ [
+ "▁under",
+ "standing"
+ ],
+ [
+ "▁understand",
+ "ing"
+ ],
+ [
+ "▁pl",
+ "ans"
+ ],
+ [
+ "▁plan",
+ "s"
+ ],
+ [
+ "rop",
+ "ri"
+ ],
+ [
+ "▁old",
+ "er"
+ ],
+ [
+ "▁ol",
+ "der"
+ ],
+ [
+ "▁",
+ "older"
+ ],
+ [
+ "▁pres",
+ "sure"
+ ],
+ [
+ "▁press",
+ "ure"
+ ],
+ [
+ "Im",
+ "pl"
+ ],
+ [
+ "Imp",
+ "l"
+ ],
+ [
+ "▁pe",
+ "ace"
+ ],
+ [
+ "Conne",
+ "ction"
+ ],
+ [
+ "Conn",
+ "ection"
+ ],
+ [
+ "Connect",
+ "ion"
+ ],
+ [
+ "▁f",
+ "i"
+ ],
+ [
+ "▁",
+ "fi"
+ ],
+ [
+ "ri",
+ "ch"
+ ],
+ [
+ "ric",
+ "h"
+ ],
+ [
+ "r",
+ "ich"
+ ],
+ [
+ "▁sh",
+ "ut"
+ ],
+ [
+ "ap",
+ "ers"
+ ],
+ [
+ "ape",
+ "rs"
+ ],
+ [
+ "aper",
+ "s"
+ ],
+ [
+ "a",
+ "pers"
+ ],
+ [
+ "Po",
+ "rt"
+ ],
+ [
+ "P",
+ "ort"
+ ],
+ [
+ "▁L",
+ "ook"
+ ],
+ [
+ "▁Lo",
+ "ok"
+ ],
+ [
+ "▁",
+ "Look"
+ ],
+ [
+ "ri",
+ "m"
+ ],
+ [
+ "r",
+ "im"
+ ],
+ [
+ "au",
+ "th"
+ ],
+ [
+ "aut",
+ "h"
+ ],
+ [
+ "a",
+ "uth"
+ ],
+ [
+ "au",
+ "to"
+ ],
+ [
+ "aut",
+ "o"
+ ],
+ [
+ "a",
+ "uto"
+ ],
+ [
+ "▁high",
+ "ly"
+ ],
+ [
+ "▁un",
+ "less"
+ ],
+ [
+ "▁W",
+ "al"
+ ],
+ [
+ "▁Wa",
+ "l"
+ ],
+ [
+ "▁re",
+ "n"
+ ],
+ [
+ "▁r",
+ "en"
+ ],
+ [
+ "▁",
+ "ren"
+ ],
+ [
+ "w",
+ "s"
+ ],
+ [
+ "▁c",
+ "ore"
+ ],
+ [
+ "▁co",
+ "re"
+ ],
+ [
+ "▁cor",
+ "e"
+ ],
+ [
+ "▁",
+ "core"
+ ],
+ [
+ "(",
+ "-"
+ ],
+ [
+ "▁c",
+ "lim"
+ ],
+ [
+ "▁cl",
+ "im"
+ ],
+ [
+ "ru",
+ "it"
+ ],
+ [
+ "r",
+ "uit"
+ ],
+ [
+ "▁call",
+ "back"
+ ],
+ [
+ "▁",
+ "callback"
+ ],
+ [
+ "he",
+ "st"
+ ],
+ [
+ "hes",
+ "t"
+ ],
+ [
+ "h",
+ "est"
+ ],
+ [
+ "▁Char",
+ "les"
+ ],
+ [
+ "▁Charl",
+ "es"
+ ],
+ [
+ "▁L",
+ "ong"
+ ],
+ [
+ "▁Lo",
+ "ng"
+ ],
+ [
+ "▁",
+ "Long"
+ ],
+ [
+ "}",
+ "="
+ ],
+ [
+ "ъ",
+ "р"
+ ],
+ [
+ "▁sh",
+ "ared"
+ ],
+ [
+ "▁share",
+ "d"
+ ],
+ [
+ "▁shar",
+ "ed"
+ ],
+ [
+ "▁sha",
+ "red"
+ ],
+ [
+ "▁",
+ "shared"
+ ],
+ [
+ "ul",
+ "ated"
+ ],
+ [
+ "ula",
+ "ted"
+ ],
+ [
+ "ulate",
+ "d"
+ ],
+ [
+ "gorith",
+ "m"
+ ],
+ [
+ "▁H",
+ "ome"
+ ],
+ [
+ "▁Ho",
+ "me"
+ ],
+ [
+ "▁Hom",
+ "e"
+ ],
+ [
+ "▁",
+ "Home"
+ ],
+ [
+ "▁vill",
+ "age"
+ ],
+ [
+ "▁vil",
+ "lage"
+ ],
+ [
+ "ee",
+ "s"
+ ],
+ [
+ "e",
+ "es"
+ ],
+ [
+ "s",
+ "v"
+ ],
+ [
+ "▁rest",
+ "aur"
+ ],
+ [
+ "re",
+ "y"
+ ],
+ [
+ "r",
+ "ey"
+ ],
+ [
+ "▁C",
+ "ast"
+ ],
+ [
+ "▁Cas",
+ "t"
+ ],
+ [
+ "▁Ca",
+ "st"
+ ],
+ [
+ "▁",
+ "Cast"
+ ],
+ [
+ "▁P",
+ "erson"
+ ],
+ [
+ "▁Per",
+ "son"
+ ],
+ [
+ "▁Pers",
+ "on"
+ ],
+ [
+ "▁",
+ "Person"
+ ],
+ [
+ "ки",
+ "й"
+ ],
+ [
+ "▁organ",
+ "iz"
+ ],
+ [
+ "▁R",
+ "ad"
+ ],
+ [
+ "▁Ra",
+ "d"
+ ],
+ [
+ "▁",
+ "Rad"
+ ],
+ [
+ "pon",
+ "ents"
+ ],
+ [
+ "ponent",
+ "s"
+ ],
+ [
+ "▁wer",
+ "den"
+ ],
+ [
+ "▁werd",
+ "en"
+ ],
+ [
+ "▁b",
+ "ow"
+ ],
+ [
+ "▁bo",
+ "w"
+ ],
+ [
+ "▁",
+ "bow"
+ ],
+ [
+ "se",
+ "n"
+ ],
+ [
+ "s",
+ "en"
+ ],
+ [
+ "am",
+ "i"
+ ],
+ [
+ "a",
+ "mi"
+ ],
+ [
+ "Inter",
+ "face"
+ ],
+ [
+ "▁b",
+ "asis"
+ ],
+ [
+ "▁bas",
+ "is"
+ ],
+ [
+ "▁ba",
+ "sis"
+ ],
+ [
+ "▁Comp",
+ "any"
+ ],
+ [
+ "▁Compan",
+ "y"
+ ],
+ [
+ "▁",
+ "Company"
+ ],
+ [
+ "er",
+ "nel"
+ ],
+ [
+ "ern",
+ "el"
+ ],
+ [
+ "erne",
+ "l"
+ ],
+ [
+ "it",
+ "u"
+ ],
+ [
+ "i",
+ "tu"
+ ],
+ [
+ "Has",
+ "h"
+ ],
+ [
+ "Ha",
+ "sh"
+ ],
+ [
+ "H",
+ "ash"
+ ],
+ [
+ "▁a",
+ "an"
+ ],
+ [
+ "▁",
+ "х"
+ ],
+ [
+ "▁s",
+ "mile"
+ ],
+ [
+ "▁sm",
+ "ile"
+ ],
+ [
+ "x",
+ "ml"
+ ],
+ [
+ "▁s",
+ "cen"
+ ],
+ [
+ "▁sc",
+ "en"
+ ],
+ [
+ "am",
+ "m"
+ ],
+ [
+ "a",
+ "mm"
+ ],
+ [
+ "to",
+ "ol"
+ ],
+ [
+ "too",
+ "l"
+ ],
+ [
+ "t",
+ "ool"
+ ],
+ [
+ "ar",
+ "ia"
+ ],
+ [
+ "ari",
+ "a"
+ ],
+ [
+ "a",
+ "ria"
+ ],
+ [
+ "▁acc",
+ "ur"
+ ],
+ [
+ "▁ac",
+ "cur"
+ ],
+ [
+ "▁",
+ "accur"
+ ],
+ [
+ "set",
+ "tings"
+ ],
+ [
+ "setting",
+ "s"
+ ],
+ [
+ "▁Jes",
+ "us"
+ ],
+ [
+ "ac",
+ "ement"
+ ],
+ [
+ "ace",
+ "ment"
+ ],
+ [
+ "po",
+ "wer"
+ ],
+ [
+ "pow",
+ "er"
+ ],
+ [
+ "p",
+ "ower"
+ ],
+ [
+ "(",
+ "!"
+ ],
+ [
+ "▁c",
+ "alls"
+ ],
+ [
+ "▁call",
+ "s"
+ ],
+ [
+ "▁cal",
+ "ls"
+ ],
+ [
+ "▁",
+ "calls"
+ ],
+ [
+ "▁bas",
+ "ic"
+ ],
+ [
+ "▁",
+ "basic"
+ ],
+ [
+ "▁set",
+ "tings"
+ ],
+ [
+ "▁sett",
+ "ings"
+ ],
+ [
+ "▁setting",
+ "s"
+ ],
+ [
+ "▁",
+ "settings"
+ ],
+ [
+ "ri",
+ "pt"
+ ],
+ [
+ "rip",
+ "t"
+ ],
+ [
+ "r",
+ "ipt"
+ ],
+ [
+ "po",
+ "ol"
+ ],
+ [
+ "p",
+ "ool"
+ ],
+ [
+ "ct",
+ "ors"
+ ],
+ [
+ "ctor",
+ "s"
+ ],
+ [
+ "▁Found",
+ "ation"
+ ],
+ [
+ "▁",
+ "Foundation"
+ ],
+ [
+ "▁we",
+ "ap"
+ ],
+ [
+ "KE",
+ "Y"
+ ],
+ [
+ "K",
+ "EY"
+ ],
+ [
+ "fo",
+ "ot"
+ ],
+ [
+ "foo",
+ "t"
+ ],
+ [
+ "f",
+ "oot"
+ ],
+ [
+ "▁r",
+ "adio"
+ ],
+ [
+ "▁rad",
+ "io"
+ ],
+ [
+ "▁radi",
+ "o"
+ ],
+ [
+ "▁",
+ "radio"
+ ],
+ [
+ "▁hel",
+ "ped"
+ ],
+ [
+ "▁help",
+ "ed"
+ ],
+ [
+ "ma",
+ "nn"
+ ],
+ [
+ "man",
+ "n"
+ ],
+ [
+ "m",
+ "ann"
+ ],
+ [
+ "▁j",
+ "ump"
+ ],
+ [
+ "▁ju",
+ "mp"
+ ],
+ [
+ "▁t",
+ "ick"
+ ],
+ [
+ "▁ti",
+ "ck"
+ ],
+ [
+ "▁",
+ "tick"
+ ],
+ [
+ "▁gr",
+ "owing"
+ ],
+ [
+ "▁grow",
+ "ing"
+ ],
+ [
+ "▁gro",
+ "wing"
+ ],
+ [
+ "at",
+ "en"
+ ],
+ [
+ "ate",
+ "n"
+ ],
+ [
+ "a",
+ "ten"
+ ],
+ [
+ "re",
+ "al"
+ ],
+ [
+ "rea",
+ "l"
+ ],
+ [
+ "▁incre",
+ "asing"
+ ],
+ [
+ "Dev",
+ "ice"
+ ],
+ [
+ "var",
+ "epsilon"
+ ],
+ [
+ "vare",
+ "psilon"
+ ],
+ [
+ "▁s",
+ "ets"
+ ],
+ [
+ "▁se",
+ "ts"
+ ],
+ [
+ "▁set",
+ "s"
+ ],
+ [
+ "▁",
+ "sets"
+ ],
+ [
+ "▁adv",
+ "ant"
+ ],
+ [
+ "Op",
+ "en"
+ ],
+ [
+ "O",
+ "pen"
+ ],
+ [
+ "▁re",
+ "asons"
+ ],
+ [
+ "▁reason",
+ "s"
+ ],
+ [
+ "▁sup",
+ "posed"
+ ],
+ [
+ "▁supp",
+ "osed"
+ ],
+ [
+ "▁suppose",
+ "d"
+ ],
+ [
+ "oe",
+ "s"
+ ],
+ [
+ "o",
+ "es"
+ ],
+ [
+ "ed",
+ "e"
+ ],
+ [
+ "e",
+ "de"
+ ],
+ [
+ "te",
+ "en"
+ ],
+ [
+ "tee",
+ "n"
+ ],
+ [
+ "t",
+ "een"
+ ],
+ [
+ "if",
+ "def"
+ ],
+ [
+ "▁de",
+ "lete"
+ ],
+ [
+ "▁del",
+ "ete"
+ ],
+ [
+ "▁delet",
+ "e"
+ ],
+ [
+ "▁",
+ "delete"
+ ],
+ [
+ "▁&",
+ "="
+ ],
+ [
+ "▁",
+ "&="
+ ],
+ [
+ "▁B",
+ "ill"
+ ],
+ [
+ "▁Bi",
+ "ll"
+ ],
+ [
+ "▁Bil",
+ "l"
+ ],
+ [
+ "▁",
+ "Bill"
+ ],
+ [
+ "▁a",
+ "im"
+ ],
+ [
+ "▁ai",
+ "m"
+ ],
+ [
+ "▁",
+ "aim"
+ ],
+ [
+ "▁O",
+ "k"
+ ],
+ [
+ "▁",
+ "Ok"
+ ],
+ [
+ "▁A",
+ "v"
+ ],
+ [
+ "▁",
+ "Av"
+ ],
+ [
+ "re",
+ "ci"
+ ],
+ [
+ "rec",
+ "i"
+ ],
+ [
+ "ac",
+ "ks"
+ ],
+ [
+ "ack",
+ "s"
+ ],
+ [
+ "a",
+ "cks"
+ ],
+ [
+ "is",
+ "te"
+ ],
+ [
+ "ist",
+ "e"
+ ],
+ [
+ "i",
+ "ste"
+ ],
+ [
+ "Pro",
+ "perties"
+ ],
+ [
+ "▁t",
+ "mp"
+ ],
+ [
+ "▁tm",
+ "p"
+ ],
+ [
+ "▁",
+ "tmp"
+ ],
+ [
+ "▁d",
+ "ei"
+ ],
+ [
+ "▁de",
+ "i"
+ ],
+ [
+ "PE",
+ "R"
+ ],
+ [
+ "P",
+ "ER"
+ ],
+ [
+ "D",
+ "C"
+ ],
+ [
+ "st",
+ "a"
+ ],
+ [
+ "s",
+ "ta"
+ ],
+ [
+ "ни",
+ "и"
+ ],
+ [
+ "▁lim",
+ "ited"
+ ],
+ [
+ "▁limit",
+ "ed"
+ ],
+ [
+ "▁",
+ "limited"
+ ],
+ [
+ "▁great",
+ "er"
+ ],
+ [
+ "▁gre",
+ "ater"
+ ],
+ [
+ "de",
+ "scription"
+ ],
+ [
+ "des",
+ "cription"
+ ],
+ [
+ "or",
+ "i"
+ ],
+ [
+ "o",
+ "ri"
+ ],
+ [
+ "ain",
+ "ts"
+ ],
+ [
+ "aint",
+ "s"
+ ],
+ [
+ "▁h",
+ "y"
+ ],
+ [
+ "▁",
+ "hy"
+ ],
+ [
+ "▁M",
+ "el"
+ ],
+ [
+ "▁Me",
+ "l"
+ ],
+ [
+ "▁C",
+ "H"
+ ],
+ [
+ "▁",
+ "CH"
+ ],
+ [
+ "con",
+ "s"
+ ],
+ [
+ "co",
+ "ns"
+ ],
+ [
+ "c",
+ "ons"
+ ],
+ [
+ "▁sur",
+ "round"
+ ],
+ [
+ "▁W",
+ "ho"
+ ],
+ [
+ "▁Wh",
+ "o"
+ ],
+ [
+ "▁",
+ "Who"
+ ],
+ [
+ "ar",
+ "c"
+ ],
+ [
+ "a",
+ "rc"
+ ],
+ [
+ "▁te",
+ "lev"
+ ],
+ [
+ "▁tele",
+ "v"
+ ],
+ [
+ "▁tel",
+ "ev"
+ ],
+ [
+ "it",
+ "ution"
+ ],
+ [
+ "itut",
+ "ion"
+ ],
+ [
+ "▁e",
+ "qual"
+ ],
+ [
+ "▁equ",
+ "al"
+ ],
+ [
+ "▁eq",
+ "ual"
+ ],
+ [
+ "▁",
+ "equal"
+ ],
+ [
+ "к",
+ "і"
+ ],
+ [
+ "▁Is",
+ "rael"
+ ],
+ [
+ "ä",
+ "h"
+ ],
+ [
+ "▁C",
+ "aption"
+ ],
+ [
+ "▁Capt",
+ "ion"
+ ],
+ [
+ "▁Ca",
+ "ption"
+ ],
+ [
+ "▁ex",
+ "erc"
+ ],
+ [
+ "em",
+ "por"
+ ],
+ [
+ "emp",
+ "or"
+ ],
+ [
+ "▁+",
+ "+"
+ ],
+ [
+ "▁",
+ "++"
+ ],
+ [
+ "▁l",
+ "ib"
+ ],
+ [
+ "▁li",
+ "b"
+ ],
+ [
+ "▁",
+ "lib"
+ ],
+ [
+ "ma",
+ "ke"
+ ],
+ [
+ "m",
+ "ake"
+ ],
+ [
+ "▁M",
+ "A"
+ ],
+ [
+ "▁",
+ "MA"
+ ],
+ [
+ "co",
+ "py"
+ ],
+ [
+ "cop",
+ "y"
+ ],
+ [
+ "c",
+ "opy"
+ ],
+ [
+ "f",
+ "riend"
+ ],
+ [
+ "▁ко",
+ "то"
+ ],
+ [
+ "▁",
+ "кото"
+ ],
+ [
+ "▁dam",
+ "age"
+ ],
+ [
+ "▁\\",
+ ","
+ ],
+ [
+ "▁",
+ "\\,"
+ ],
+ [
+ "od",
+ "ed"
+ ],
+ [
+ "ode",
+ "d"
+ ],
+ [
+ "o",
+ "ded"
+ ],
+ [
+ "▁n",
+ "one"
+ ],
+ [
+ "▁no",
+ "ne"
+ ],
+ [
+ "▁non",
+ "e"
+ ],
+ [
+ "▁",
+ "none"
+ ],
+ [
+ "▁ev",
+ "alu"
+ ],
+ [
+ "▁eval",
+ "u"
+ ],
+ [
+ "▁",
+ "evalu"
+ ],
+ [
+ "st",
+ "on"
+ ],
+ [
+ "sto",
+ "n"
+ ],
+ [
+ "s",
+ "ton"
+ ],
+ [
+ ">",
+ ","
+ ],
+ [
+ "FO",
+ "R"
+ ],
+ [
+ "F",
+ "OR"
+ ],
+ [
+ "▁n",
+ "orm"
+ ],
+ [
+ "▁no",
+ "rm"
+ ],
+ [
+ "▁nor",
+ "m"
+ ],
+ [
+ "▁",
+ "norm"
+ ],
+ [
+ "ap",
+ "pe"
+ ],
+ [
+ "app",
+ "e"
+ ],
+ [
+ "a",
+ "ppe"
+ ],
+ [
+ "S",
+ "ession"
+ ],
+ [
+ "▁ad",
+ "ult"
+ ],
+ [
+ "▁h",
+ "ospital"
+ ],
+ [
+ "▁hosp",
+ "ital"
+ ],
+ [
+ "▁recomm",
+ "end"
+ ],
+ [
+ "pro",
+ "perty"
+ ],
+ [
+ "ste",
+ "in"
+ ],
+ [
+ "fin",
+ "al"
+ ],
+ [
+ "fi",
+ "nal"
+ ],
+ [
+ "f",
+ "inal"
+ ],
+ [
+ "▁n",
+ "u"
+ ],
+ [
+ "▁",
+ "nu"
+ ],
+ [
+ "se",
+ "cond"
+ ],
+ [
+ "sec",
+ "ond"
+ ],
+ [
+ "▁a",
+ "spect"
+ ],
+ [
+ "▁as",
+ "pect"
+ ],
+ [
+ "▁asp",
+ "ect"
+ ],
+ [
+ "\")",
+ "]"
+ ],
+ [
+ "\"",
+ ")]"
+ ],
+ [
+ "же",
+ "н"
+ ],
+ [
+ "ж",
+ "ен"
+ ],
+ [
+ "am",
+ "ento"
+ ],
+ [
+ "ament",
+ "o"
+ ],
+ [
+ "amen",
+ "to"
+ ],
+ [
+ "▁r",
+ "ac"
+ ],
+ [
+ "▁ra",
+ "c"
+ ],
+ [
+ "▁",
+ "rac"
+ ],
+ [
+ "sa",
+ "ve"
+ ],
+ [
+ "s",
+ "ave"
+ ],
+ [
+ "▁foot",
+ "ball"
+ ],
+ [
+ "A",
+ "b"
+ ],
+ [
+ "un",
+ "gs"
+ ],
+ [
+ "ung",
+ "s"
+ ],
+ [
+ "ab",
+ "il"
+ ],
+ [
+ "abi",
+ "l"
+ ],
+ [
+ "a",
+ "bil"
+ ],
+ [
+ "▁Ar",
+ "ch"
+ ],
+ [
+ "▁Arc",
+ "h"
+ ],
+ [
+ "▁",
+ "Arch"
+ ],
+ [
+ "sys",
+ "tem"
+ ],
+ [
+ "s",
+ "ystem"
+ ],
+ [
+ "hi",
+ "st"
+ ],
+ [
+ "his",
+ "t"
+ ],
+ [
+ "h",
+ "ist"
+ ],
+ [
+ "▁l",
+ "uck"
+ ],
+ [
+ "▁lu",
+ "ck"
+ ],
+ [
+ "▁luc",
+ "k"
+ ],
+ [
+ "re",
+ "nder"
+ ],
+ [
+ "ren",
+ "der"
+ ],
+ [
+ "rend",
+ "er"
+ ],
+ [
+ "r",
+ "ender"
+ ],
+ [
+ "▁se",
+ "in"
+ ],
+ [
+ "▁sei",
+ "n"
+ ],
+ [
+ "ion",
+ "i"
+ ],
+ [
+ "io",
+ "ni"
+ ],
+ [
+ "i",
+ "oni"
+ ],
+ [
+ "▁r",
+ "ot"
+ ],
+ [
+ "▁ro",
+ "t"
+ ],
+ [
+ "▁",
+ "rot"
+ ],
+ [
+ "▁cor",
+ "ner"
+ ],
+ [
+ "▁corn",
+ "er"
+ ],
+ [
+ "▁app",
+ "ropri"
+ ],
+ [
+ "▁ap",
+ "propri"
+ ],
+ [
+ "▁",
+ "appropri"
+ ],
+ [
+ "▁Soft",
+ "ware"
+ ],
+ [
+ "▁t",
+ "ele"
+ ],
+ [
+ "▁te",
+ "le"
+ ],
+ [
+ "▁tel",
+ "e"
+ ],
+ [
+ "▁",
+ "tele"
+ ],
+ [
+ "De",
+ "lete"
+ ],
+ [
+ "Dele",
+ "te"
+ ],
+ [
+ "Del",
+ "ete"
+ ],
+ [
+ "▁Acc",
+ "ording"
+ ],
+ [
+ "▁pr",
+ "ison"
+ ],
+ [
+ "▁pri",
+ "son"
+ ],
+ [
+ "▁",
+ "prison"
+ ],
+ [
+ "▁l",
+ "ic"
+ ],
+ [
+ "▁li",
+ "c"
+ ],
+ [
+ "▁",
+ "lic"
+ ],
+ [
+ "▁м",
+ "и"
+ ],
+ [
+ "▁",
+ "ми"
+ ],
+ [
+ "ter",
+ "m"
+ ],
+ [
+ "te",
+ "rm"
+ ],
+ [
+ "t",
+ "erm"
+ ],
+ [
+ "se",
+ "ts"
+ ],
+ [
+ "set",
+ "s"
+ ],
+ [
+ "s",
+ "ets"
+ ],
+ [
+ "▁v",
+ "el"
+ ],
+ [
+ "▁ve",
+ "l"
+ ],
+ [
+ "▁",
+ "vel"
+ ],
+ [
+ "▁r",
+ "ank"
+ ],
+ [
+ "▁ran",
+ "k"
+ ],
+ [
+ "▁",
+ "rank"
+ ],
+ [
+ "▁ex",
+ "isting"
+ ],
+ [
+ "▁exist",
+ "ing"
+ ],
+ [
+ "▁",
+ "existing"
+ ],
+ [
+ "▁V",
+ "ir"
+ ],
+ [
+ "▁Vi",
+ "r"
+ ],
+ [
+ "▁t",
+ "rip"
+ ],
+ [
+ "▁tr",
+ "ip"
+ ],
+ [
+ "▁tri",
+ "p"
+ ],
+ [
+ "▁м",
+ "у"
+ ],
+ [
+ "▁",
+ "му"
+ ],
+ [
+ "av",
+ "ax"
+ ],
+ [
+ "ava",
+ "x"
+ ],
+ [
+ "▁r",
+ "is"
+ ],
+ [
+ "▁ri",
+ "s"
+ ],
+ [
+ "▁",
+ "ris"
+ ],
+ [
+ "▁def",
+ "ine"
+ ],
+ [
+ "▁defin",
+ "e"
+ ],
+ [
+ "▁",
+ "define"
+ ],
+ [
+ "▁he",
+ "at"
+ ],
+ [
+ "ca",
+ "r"
+ ],
+ [
+ "c",
+ "ar"
+ ],
+ [
+ "▁con",
+ "vert"
+ ],
+ [
+ "▁conv",
+ "ert"
+ ],
+ [
+ "▁conver",
+ "t"
+ ],
+ [
+ "▁conve",
+ "rt"
+ ],
+ [
+ "▁",
+ "convert"
+ ],
+ [
+ "em",
+ "ail"
+ ],
+ [
+ "ema",
+ "il"
+ ],
+ [
+ "e",
+ "mail"
+ ],
+ [
+ "▁U",
+ "nder"
+ ],
+ [
+ "▁Un",
+ "der"
+ ],
+ [
+ "▁Und",
+ "er"
+ ],
+ [
+ "▁",
+ "Under"
+ ],
+ [
+ "▁",
+ "Ш"
+ ],
+ [
+ "▁G",
+ "rand"
+ ],
+ [
+ "▁Gr",
+ "and"
+ ],
+ [
+ "▁Gran",
+ "d"
+ ],
+ [
+ "▁Gra",
+ "nd"
+ ],
+ [
+ "▁ex",
+ "ists"
+ ],
+ [
+ "▁exist",
+ "s"
+ ],
+ [
+ "▁",
+ "exists"
+ ],
+ [
+ "sy",
+ "s"
+ ],
+ [
+ "s",
+ "ys"
+ ],
+ [
+ "ef",
+ "f"
+ ],
+ [
+ "e",
+ "ff"
+ ],
+ [
+ "▁T",
+ "op"
+ ],
+ [
+ "▁To",
+ "p"
+ ],
+ [
+ "▁",
+ "Top"
+ ],
+ [
+ "▁",
+ "č"
+ ],
+ [
+ "▁t",
+ "empor"
+ ],
+ [
+ "▁tem",
+ "por"
+ ],
+ [
+ "▁temp",
+ "or"
+ ],
+ [
+ "▁tempo",
+ "r"
+ ],
+ [
+ "▁arg",
+ "uments"
+ ],
+ [
+ "▁argument",
+ "s"
+ ],
+ [
+ "▁",
+ "arguments"
+ ],
+ [
+ "▁support",
+ "ed"
+ ],
+ [
+ "▁supp",
+ "orted"
+ ],
+ [
+ "▁",
+ "supported"
+ ],
+ [
+ "en",
+ "sed"
+ ],
+ [
+ "ens",
+ "ed"
+ ],
+ [
+ "ense",
+ "d"
+ ],
+ [
+ "▁Franc",
+ "is"
+ ],
+ [
+ "▁co",
+ "ord"
+ ],
+ [
+ "▁",
+ "coord"
+ ],
+ [
+ "▁achie",
+ "ve"
+ ],
+ [
+ "▁N",
+ "ame"
+ ],
+ [
+ "▁Na",
+ "me"
+ ],
+ [
+ "▁Nam",
+ "e"
+ ],
+ [
+ "▁",
+ "Name"
+ ],
+ [
+ "▁J",
+ "ahr"
+ ],
+ [
+ "▁Jah",
+ "r"
+ ],
+ [
+ "▁Ja",
+ "hr"
+ ],
+ [
+ "▁G",
+ "i"
+ ],
+ [
+ "sh",
+ "e"
+ ],
+ [
+ "s",
+ "he"
+ ],
+ [
+ "▁D",
+ "ev"
+ ],
+ [
+ "▁De",
+ "v"
+ ],
+ [
+ "▁",
+ "Dev"
+ ],
+ [
+ "▁a",
+ "lla"
+ ],
+ [
+ "▁al",
+ "la"
+ ],
+ [
+ "▁all",
+ "a"
+ ],
+ [
+ "▁",
+ "alla"
+ ],
+ [
+ "▁W",
+ "IT"
+ ],
+ [
+ "ag",
+ "ment"
+ ],
+ [
+ "c",
+ "ustom"
+ ],
+ [
+ "al",
+ "ls"
+ ],
+ [
+ "all",
+ "s"
+ ],
+ [
+ "&",
+ "&"
+ ],
+ [
+ "W",
+ "E"
+ ],
+ [
+ "▁h",
+ "olding"
+ ],
+ [
+ "▁hold",
+ "ing"
+ ],
+ [
+ "▁hol",
+ "ding"
+ ],
+ [
+ "pro",
+ "totype"
+ ],
+ [
+ "proto",
+ "type"
+ ],
+ [
+ "prot",
+ "otype"
+ ],
+ [
+ "▁f",
+ "ing"
+ ],
+ [
+ "▁fin",
+ "g"
+ ],
+ [
+ "▁fi",
+ "ng"
+ ],
+ [
+ "▁b",
+ "ag"
+ ],
+ [
+ "▁ba",
+ "g"
+ ],
+ [
+ "▁",
+ "bag"
+ ],
+ [
+ "▁Par",
+ "ty"
+ ],
+ [
+ "▁Part",
+ "y"
+ ],
+ [
+ "st",
+ "ack"
+ ],
+ [
+ "sta",
+ "ck"
+ ],
+ [
+ "▁econom",
+ "ic"
+ ],
+ [
+ "▁G",
+ "al"
+ ],
+ [
+ "▁Ga",
+ "l"
+ ],
+ [
+ "id",
+ "ents"
+ ],
+ [
+ "ident",
+ "s"
+ ],
+ [
+ "iden",
+ "ts"
+ ],
+ [
+ "▁J",
+ "un"
+ ],
+ [
+ "▁Ju",
+ "n"
+ ],
+ [
+ "▁sh",
+ "owed"
+ ],
+ [
+ "▁show",
+ "ed"
+ ],
+ [
+ "os",
+ "h"
+ ],
+ [
+ "o",
+ "sh"
+ ],
+ [
+ "▁B",
+ "ay"
+ ],
+ [
+ "▁Ba",
+ "y"
+ ],
+ [
+ "▁",
+ "Bay"
+ ],
+ [
+ "ma",
+ "il"
+ ],
+ [
+ "m",
+ "ail"
+ ],
+ [
+ "▁S",
+ "O"
+ ],
+ [
+ "▁",
+ "SO"
+ ],
+ [
+ "▁\"",
+ "<"
+ ],
+ [
+ "graph",
+ "ics"
+ ],
+ [
+ "▁f",
+ "u"
+ ],
+ [
+ "▁",
+ "fu"
+ ],
+ [
+ "cl",
+ "ick"
+ ],
+ [
+ "cli",
+ "ck"
+ ],
+ [
+ "c",
+ "lick"
+ ],
+ [
+ "▁b",
+ "attle"
+ ],
+ [
+ "▁batt",
+ "le"
+ ],
+ [
+ "▁bat",
+ "tle"
+ ],
+ [
+ "{",
+ "{"
+ ],
+ [
+ "▁E",
+ "vent"
+ ],
+ [
+ "▁Even",
+ "t"
+ ],
+ [
+ "▁Ev",
+ "ent"
+ ],
+ [
+ "▁Eve",
+ "nt"
+ ],
+ [
+ "▁",
+ "Event"
+ ],
+ [
+ "ri",
+ "or"
+ ],
+ [
+ "rio",
+ "r"
+ ],
+ [
+ "r",
+ "ior"
+ ],
+ [
+ "ch",
+ "aft"
+ ],
+ [
+ "cha",
+ "ft"
+ ],
+ [
+ "▁f",
+ "avorite"
+ ],
+ [
+ "▁favor",
+ "ite"
+ ],
+ [
+ "us",
+ "ive"
+ ],
+ [
+ "sup",
+ "port"
+ ],
+ [
+ "supp",
+ "ort"
+ ],
+ [
+ "s",
+ "upport"
+ ],
+ [
+ "b",
+ "m"
+ ],
+ [
+ "K",
+ "ind"
+ ],
+ [
+ "▁saf",
+ "ety"
+ ],
+ [
+ "▁safe",
+ "ty"
+ ],
+ [
+ "▁E",
+ "nt"
+ ],
+ [
+ "▁En",
+ "t"
+ ],
+ [
+ "▁",
+ "Ent"
+ ],
+ [
+ "cu",
+ "p"
+ ],
+ [
+ "c",
+ "up"
+ ],
+ [
+ "▁Austral",
+ "ia"
+ ],
+ [
+ "▁dest",
+ "roy"
+ ],
+ [
+ "▁destro",
+ "y"
+ ],
+ [
+ "▁",
+ "destroy"
+ ],
+ [
+ "▁organ",
+ "ization"
+ ],
+ [
+ "▁organiz",
+ "ation"
+ ],
+ [
+ "id",
+ "en"
+ ],
+ [
+ "ide",
+ "n"
+ ],
+ [
+ "i",
+ "den"
+ ],
+ [
+ "########",
+ "########"
+ ],
+ [
+ "de",
+ "c"
+ ],
+ [
+ "d",
+ "ec"
+ ],
+ [
+ "▁z",
+ "a"
+ ],
+ [
+ "▁",
+ "za"
+ ],
+ [
+ "▁s",
+ "even"
+ ],
+ [
+ "▁se",
+ "ven"
+ ],
+ [
+ "▁",
+ "seven"
+ ],
+ [
+ "ar",
+ "ely"
+ ],
+ [
+ "are",
+ "ly"
+ ],
+ [
+ "arel",
+ "y"
+ ],
+ [
+ "▁f",
+ "lag"
+ ],
+ [
+ "▁fl",
+ "ag"
+ ],
+ [
+ "▁",
+ "flag"
+ ],
+ [
+ "Di",
+ "r"
+ ],
+ [
+ "D",
+ "ir"
+ ],
+ [
+ "▁C",
+ "arl"
+ ],
+ [
+ "▁Car",
+ "l"
+ ],
+ [
+ "▁Ca",
+ "rl"
+ ],
+ [
+ "▁do",
+ "ctor"
+ ],
+ [
+ "▁doc",
+ "tor"
+ ],
+ [
+ "▁var",
+ "iety"
+ ],
+ [
+ "▁vari",
+ "ety"
+ ],
+ [
+ "▁L",
+ "in"
+ ],
+ [
+ "▁Li",
+ "n"
+ ],
+ [
+ "▁",
+ "Lin"
+ ],
+ [
+ "▁t",
+ "om"
+ ],
+ [
+ "▁to",
+ "m"
+ ],
+ [
+ "▁",
+ "tom"
+ ],
+ [
+ "^{",
+ "("
+ ],
+ [
+ "^",
+ "{("
+ ],
+ [
+ "B",
+ "o"
+ ],
+ [
+ "an",
+ "tes"
+ ],
+ [
+ "ant",
+ "es"
+ ],
+ [
+ "ante",
+ "s"
+ ],
+ [
+ "▁m",
+ "ine"
+ ],
+ [
+ "▁min",
+ "e"
+ ],
+ [
+ "▁mi",
+ "ne"
+ ],
+ [
+ "▁",
+ "mine"
+ ],
+ [
+ "▁M",
+ "it"
+ ],
+ [
+ "▁Mi",
+ "t"
+ ],
+ [
+ "▁de",
+ "scribe"
+ ],
+ [
+ "▁desc",
+ "ribe"
+ ],
+ [
+ "▁describ",
+ "e"
+ ],
+ [
+ "Ar",
+ "gs"
+ ],
+ [
+ "Arg",
+ "s"
+ ],
+ [
+ "L",
+ "S"
+ ],
+ [
+ "AP",
+ "I"
+ ],
+ [
+ "A",
+ "PI"
+ ],
+ [
+ "▁L",
+ "uc"
+ ],
+ [
+ "▁Lu",
+ "c"
+ ],
+ [
+ "▁",
+ "Luc"
+ ],
+ [
+ "ph",
+ "one"
+ ],
+ [
+ "▁sc",
+ "ience"
+ ],
+ [
+ "▁",
+ "science"
+ ],
+ [
+ "▁O",
+ "per"
+ ],
+ [
+ "▁Op",
+ "er"
+ ],
+ [
+ "▁",
+ "Oper"
+ ],
+ [
+ "Ne",
+ "xt"
+ ],
+ [
+ "N",
+ "ext"
+ ],
+ [
+ "▁invest",
+ "ig"
+ ],
+ [
+ "▁demon",
+ "str"
+ ],
+ [
+ "▁G",
+ "overn"
+ ],
+ [
+ "▁Go",
+ "vern"
+ ],
+ [
+ "▁object",
+ "s"
+ ],
+ [
+ "▁",
+ "objects"
+ ],
+ [
+ "▁Lou",
+ "is"
+ ],
+ [
+ "▁Lo",
+ "uis"
+ ],
+ [
+ "▁Return",
+ "s"
+ ],
+ [
+ "▁",
+ "Returns"
+ ],
+ [
+ "▁h",
+ "an"
+ ],
+ [
+ "▁ha",
+ "n"
+ ],
+ [
+ "▁",
+ "han"
+ ],
+ [
+ "na",
+ "m"
+ ],
+ [
+ "n",
+ "am"
+ ],
+ [
+ "▁com",
+ "me"
+ ],
+ [
+ "▁comm",
+ "e"
+ ],
+ [
+ "▁pres",
+ "ence"
+ ],
+ [
+ "▁p",
+ "el"
+ ],
+ [
+ "▁pe",
+ "l"
+ ],
+ [
+ "▁",
+ "pel"
+ ],
+ [
+ "▁det",
+ "ect"
+ ],
+ [
+ "▁",
+ "detect"
+ ],
+ [
+ ")",
+ "="
+ ],
+ [
+ "▁Ch",
+ "inese"
+ ],
+ [
+ "▁r",
+ "ich"
+ ],
+ [
+ "▁ri",
+ "ch"
+ ],
+ [
+ "▁ric",
+ "h"
+ ],
+ [
+ "▁",
+ "rich"
+ ],
+ [
+ "▁class",
+ "es"
+ ],
+ [
+ "▁classe",
+ "s"
+ ],
+ [
+ "▁clas",
+ "ses"
+ ],
+ [
+ "▁",
+ "classes"
+ ],
+ [
+ "▁exp",
+ "and"
+ ],
+ [
+ "▁",
+ "expand"
+ ],
+ [
+ "▁D",
+ "om"
+ ],
+ [
+ "▁Do",
+ "m"
+ ],
+ [
+ "▁",
+ "Dom"
+ ],
+ [
+ "▁D",
+ "ec"
+ ],
+ [
+ "▁De",
+ "c"
+ ],
+ [
+ "▁",
+ "Dec"
+ ],
+ [
+ "s",
+ "n"
+ ],
+ [
+ "pe",
+ "ed"
+ ],
+ [
+ "p",
+ "eed"
+ ],
+ [
+ "▁J",
+ "im"
+ ],
+ [
+ "▁Ji",
+ "m"
+ ],
+ [
+ "sh",
+ "ould"
+ ],
+ [
+ "▁Sm",
+ "ith"
+ ],
+ [
+ "▁p",
+ "ages"
+ ],
+ [
+ "▁page",
+ "s"
+ ],
+ [
+ "▁pa",
+ "ges"
+ ],
+ [
+ "▁pag",
+ "es"
+ ],
+ [
+ "▁",
+ "pages"
+ ],
+ [
+ "▁Je",
+ "an"
+ ],
+ [
+ "ri",
+ "cs"
+ ],
+ [
+ "ric",
+ "s"
+ ],
+ [
+ "r",
+ "ics"
+ ],
+ [
+ "▁S",
+ "und"
+ ],
+ [
+ "▁Su",
+ "nd"
+ ],
+ [
+ "▁Sun",
+ "d"
+ ],
+ [
+ "ad",
+ "s"
+ ],
+ [
+ "a",
+ "ds"
+ ],
+ [
+ "▁The",
+ "ir"
+ ],
+ [
+ "un",
+ "icip"
+ ],
+ [
+ "uni",
+ "cip"
+ ],
+ [
+ "unic",
+ "ip"
+ ],
+ [
+ "в",
+ "у"
+ ],
+ [
+ "▁down",
+ "load"
+ ],
+ [
+ "▁",
+ "download"
+ ],
+ [
+ "▁st",
+ "ress"
+ ],
+ [
+ "▁str",
+ "ess"
+ ],
+ [
+ "▁stre",
+ "ss"
+ ],
+ [
+ "▁P",
+ "et"
+ ],
+ [
+ "▁Pe",
+ "t"
+ ],
+ [
+ "▁",
+ "Pet"
+ ],
+ [
+ "me",
+ "nu"
+ ],
+ [
+ "men",
+ "u"
+ ],
+ [
+ "m",
+ "enu"
+ ],
+ [
+ "re",
+ "me"
+ ],
+ [
+ "rem",
+ "e"
+ ],
+ [
+ "r",
+ "eme"
+ ],
+ [
+ "▁com",
+ "pared"
+ ],
+ [
+ "▁comp",
+ "ared"
+ ],
+ [
+ "▁compar",
+ "ed"
+ ],
+ [
+ "▁compare",
+ "d"
+ ],
+ [
+ "St",
+ "e"
+ ],
+ [
+ "S",
+ "te"
+ ],
+ [
+ "IN",
+ "D"
+ ],
+ [
+ "I",
+ "ND"
+ ],
+ [
+ "cont",
+ "ainer"
+ ],
+ [
+ "▁Ind",
+ "ian"
+ ],
+ [
+ "▁India",
+ "n"
+ ],
+ [
+ "or",
+ "en"
+ ],
+ [
+ "ore",
+ "n"
+ ],
+ [
+ "o",
+ "ren"
+ ],
+ [
+ "▁s",
+ "es"
+ ],
+ [
+ "▁se",
+ "s"
+ ],
+ [
+ "▁",
+ "ses"
+ ],
+ [
+ "▁W",
+ "he"
+ ],
+ [
+ "▁Wh",
+ "e"
+ ],
+ [
+ "▁",
+ "Whe"
+ ],
+ [
+ "▁r",
+ "oku"
+ ],
+ [
+ "▁ro",
+ "ku"
+ ],
+ [
+ "▁estab",
+ "lished"
+ ],
+ [
+ "▁establish",
+ "ed"
+ ],
+ [
+ "▁gener",
+ "ally"
+ ],
+ [
+ "▁general",
+ "ly"
+ ],
+ [
+ "▁f",
+ "le"
+ ],
+ [
+ "▁fl",
+ "e"
+ ],
+ [
+ "__",
+ "("
+ ],
+ [
+ "_",
+ "_("
+ ],
+ [
+ "=\"",
+ "+"
+ ],
+ [
+ "=",
+ "\"+"
+ ],
+ [
+ "V",
+ "ar"
+ ],
+ [
+ "▁M",
+ "ake"
+ ],
+ [
+ "▁Ma",
+ "ke"
+ ],
+ [
+ "▁Mak",
+ "e"
+ ],
+ [
+ "▁",
+ "Make"
+ ],
+ [
+ "▁rem",
+ "oved"
+ ],
+ [
+ "▁remove",
+ "d"
+ ],
+ [
+ "▁",
+ "removed"
+ ],
+ [
+ "z",
+ "z"
+ ],
+ [
+ "ü",
+ "n"
+ ],
+ [
+ "▁m",
+ "ix"
+ ],
+ [
+ "▁mi",
+ "x"
+ ],
+ [
+ "▁",
+ "mix"
+ ],
+ [
+ "er",
+ "k"
+ ],
+ [
+ "iat",
+ "ion"
+ ],
+ [
+ "i",
+ "ation"
+ ],
+ [
+ "ou",
+ "ter"
+ ],
+ [
+ "out",
+ "er"
+ ],
+ [
+ "oute",
+ "r"
+ ],
+ [
+ "o",
+ "uter"
+ ],
+ [
+ "S",
+ "K"
+ ],
+ [
+ "▁be",
+ "comes"
+ ],
+ [
+ "▁bec",
+ "omes"
+ ],
+ [
+ "▁become",
+ "s"
+ ],
+ [
+ "▁H",
+ "all"
+ ],
+ [
+ "▁Ha",
+ "ll"
+ ],
+ [
+ "▁Hal",
+ "l"
+ ],
+ [
+ "sc",
+ "ious"
+ ],
+ [
+ "▁w",
+ "atched"
+ ],
+ [
+ "▁watch",
+ "ed"
+ ],
+ [
+ "▁wat",
+ "ched"
+ ],
+ [
+ "▁g",
+ "ather"
+ ],
+ [
+ "▁ga",
+ "ther"
+ ],
+ [
+ "▁",
+ "gather"
+ ],
+ [
+ "▁Res",
+ "ult"
+ ],
+ [
+ "▁",
+ "Result"
+ ],
+ [
+ "pro",
+ "of"
+ ],
+ [
+ "pa",
+ "y"
+ ],
+ [
+ "p",
+ "ay"
+ ],
+ [
+ "▁produ",
+ "ced"
+ ],
+ [
+ "▁produce",
+ "d"
+ ],
+ [
+ "▁prod",
+ "uced"
+ ],
+ [
+ "▁|",
+ "="
+ ],
+ [
+ "▁b",
+ "order"
+ ],
+ [
+ "▁bord",
+ "er"
+ ],
+ [
+ "▁bor",
+ "der"
+ ],
+ [
+ "▁",
+ "border"
+ ],
+ [
+ "▁d",
+ "in"
+ ],
+ [
+ "▁di",
+ "n"
+ ],
+ [
+ "▁s",
+ "cript"
+ ],
+ [
+ "▁sc",
+ "ript"
+ ],
+ [
+ "▁scr",
+ "ipt"
+ ],
+ [
+ "▁",
+ "script"
+ ],
+ [
+ "▁a",
+ "ctions"
+ ],
+ [
+ "▁act",
+ "ions"
+ ],
+ [
+ "▁action",
+ "s"
+ ],
+ [
+ "▁",
+ "actions"
+ ],
+ [
+ "▁m",
+ "as"
+ ],
+ [
+ "▁ma",
+ "s"
+ ],
+ [
+ "▁",
+ "mas"
+ ],
+ [
+ "щ",
+ "а"
+ ],
+ [
+ "oot",
+ "h"
+ ],
+ [
+ "oo",
+ "th"
+ ],
+ [
+ "o",
+ "oth"
+ ],
+ [
+ "▁Te",
+ "chn"
+ ],
+ [
+ "▁Tech",
+ "n"
+ ],
+ [
+ "Js",
+ "on"
+ ],
+ [
+ "J",
+ "son"
+ ],
+ [
+ "▁f",
+ "illed"
+ ],
+ [
+ "▁fil",
+ "led"
+ ],
+ [
+ "▁fill",
+ "ed"
+ ],
+ [
+ "▁",
+ "filled"
+ ],
+ [
+ "де",
+ "н"
+ ],
+ [
+ "д",
+ "ен"
+ ],
+ [
+ "und",
+ "le"
+ ],
+ [
+ "ст",
+ "у"
+ ],
+ [
+ "с",
+ "ту"
+ ],
+ [
+ "To",
+ "ol"
+ ],
+ [
+ "Too",
+ "l"
+ ],
+ [
+ "T",
+ "ool"
+ ],
+ [
+ "▁k",
+ "ing"
+ ],
+ [
+ "▁ki",
+ "ng"
+ ],
+ [
+ "▁kin",
+ "g"
+ ],
+ [
+ "▁",
+ "king"
+ ],
+ [
+ "▁v",
+ "en"
+ ],
+ [
+ "▁ve",
+ "n"
+ ],
+ [
+ "▁",
+ "ven"
+ ],
+ [
+ "st",
+ "ra"
+ ],
+ [
+ "str",
+ "a"
+ ],
+ [
+ "s",
+ "tra"
+ ],
+ [
+ "▁pre",
+ "dict"
+ ],
+ [
+ "▁pred",
+ "ict"
+ ],
+ [
+ "▁",
+ "predict"
+ ],
+ [
+ "▁l",
+ "ui"
+ ],
+ [
+ "▁lu",
+ "i"
+ ],
+ [
+ "▁WAR",
+ "RAN"
+ ],
+ [
+ "▁F",
+ "un"
+ ],
+ [
+ "▁Fu",
+ "n"
+ ],
+ [
+ "▁",
+ "Fun"
+ ],
+ [
+ "Sc",
+ "ript"
+ ],
+ [
+ "S",
+ "cript"
+ ],
+ [
+ "▁power",
+ "ful"
+ ],
+ [
+ "▁l",
+ "ose"
+ ],
+ [
+ "▁lo",
+ "se"
+ ],
+ [
+ "▁los",
+ "e"
+ ],
+ [
+ "at",
+ "ically"
+ ],
+ [
+ "atic",
+ "ally"
+ ],
+ [
+ "▁d",
+ "aily"
+ ],
+ [
+ "▁da",
+ "ily"
+ ],
+ [
+ "▁dai",
+ "ly"
+ ],
+ [
+ "▁r",
+ "ing"
+ ],
+ [
+ "▁ri",
+ "ng"
+ ],
+ [
+ "▁",
+ "ring"
+ ],
+ [
+ "▁ar",
+ "rived"
+ ],
+ [
+ "▁arriv",
+ "ed"
+ ],
+ [
+ "▁arr",
+ "ived"
+ ],
+ [
+ "▁arrive",
+ "d"
+ ],
+ [
+ "St",
+ "ack"
+ ],
+ [
+ "sc",
+ "ope"
+ ],
+ [
+ "s",
+ "cope"
+ ],
+ [
+ "▁B",
+ "ack"
+ ],
+ [
+ "▁Ba",
+ "ck"
+ ],
+ [
+ "▁",
+ "Back"
+ ],
+ [
+ "el",
+ "ij"
+ ],
+ [
+ "eli",
+ "j"
+ ],
+ [
+ "e",
+ "lij"
+ ],
+ [
+ "▁z",
+ "e"
+ ],
+ [
+ "▁",
+ "ze"
+ ],
+ [
+ "ke",
+ "ys"
+ ],
+ [
+ "key",
+ "s"
+ ],
+ [
+ "{",
+ "\""
+ ],
+ [
+ "VI",
+ "D"
+ ],
+ [
+ "V",
+ "ID"
+ ],
+ [
+ "▁l",
+ "icense"
+ ],
+ [
+ "▁lic",
+ "ense"
+ ],
+ [
+ "▁",
+ "license"
+ ],
+ [
+ "wh",
+ "at"
+ ],
+ [
+ "w",
+ "hat"
+ ],
+ [
+ "▁pro",
+ "ced"
+ ],
+ [
+ "▁proc",
+ "ed"
+ ],
+ [
+ "ra",
+ "nt"
+ ],
+ [
+ "ran",
+ "t"
+ ],
+ [
+ "r",
+ "ant"
+ ],
+ [
+ "est",
+ "ival"
+ ],
+ [
+ "ag",
+ "ram"
+ ],
+ [
+ "agr",
+ "am"
+ ],
+ [
+ "agra",
+ "m"
+ ],
+ [
+ "a",
+ "gram"
+ ],
+ [
+ "▁L",
+ "O"
+ ],
+ [
+ "▁",
+ "LO"
+ ],
+ [
+ "▁Hen",
+ "ry"
+ ],
+ [
+ "▁fl",
+ "ags"
+ ],
+ [
+ "▁flag",
+ "s"
+ ],
+ [
+ "▁",
+ "flags"
+ ],
+ [
+ "Do",
+ "wn"
+ ],
+ [
+ "D",
+ "own"
+ ],
+ [
+ "scri",
+ "ption"
+ ],
+ [
+ "script",
+ "ion"
+ ],
+ [
+ "s",
+ "cription"
+ ],
+ [
+ "▁famil",
+ "ies"
+ ],
+ [
+ "▁familie",
+ "s"
+ ],
+ [
+ "is",
+ "se"
+ ],
+ [
+ "iss",
+ "e"
+ ],
+ [
+ "bo",
+ "ur"
+ ],
+ [
+ "b",
+ "our"
+ ],
+ [
+ "▁B",
+ "ur"
+ ],
+ [
+ "▁Bu",
+ "r"
+ ],
+ [
+ "—",
+ "\""
+ ],
+ [
+ "▁b",
+ "rief"
+ ],
+ [
+ "▁br",
+ "ief"
+ ],
+ [
+ "▁",
+ "brief"
+ ],
+ [
+ "▁cre",
+ "ating"
+ ],
+ [
+ "▁creat",
+ "ing"
+ ],
+ [
+ "▁cl",
+ "ients"
+ ],
+ [
+ "▁client",
+ "s"
+ ],
+ [
+ "ran",
+ "gle"
+ ],
+ [
+ "r",
+ "angle"
+ ],
+ [
+ "▁amaz",
+ "ing"
+ ],
+ [
+ "▁s",
+ "ind"
+ ],
+ [
+ "▁si",
+ "nd"
+ ],
+ [
+ "▁sin",
+ "d"
+ ],
+ [
+ "▁cover",
+ "ed"
+ ],
+ [
+ "▁cov",
+ "ered"
+ ],
+ [
+ "▁",
+ "covered"
+ ],
+ [
+ "We",
+ "ll"
+ ],
+ [
+ "W",
+ "ell"
+ ],
+ [
+ "ст",
+ "е"
+ ],
+ [
+ "с",
+ "те"
+ ],
+ [
+ "то",
+ "р"
+ ],
+ [
+ "т",
+ "ор"
+ ],
+ [
+ "▁B",
+ "as"
+ ],
+ [
+ "▁Ba",
+ "s"
+ ],
+ [
+ "▁",
+ "Bas"
+ ],
+ [
+ "to",
+ "tal"
+ ],
+ [
+ "tot",
+ "al"
+ ],
+ [
+ "t",
+ "otal"
+ ],
+ [
+ "▁I",
+ "nit"
+ ],
+ [
+ "▁In",
+ "it"
+ ],
+ [
+ "▁",
+ "Init"
+ ],
+ [
+ "▁s",
+ "and"
+ ],
+ [
+ "▁sa",
+ "nd"
+ ],
+ [
+ "▁san",
+ "d"
+ ],
+ [
+ "Un",
+ "it"
+ ],
+ [
+ "U",
+ "nit"
+ ],
+ [
+ "▁mur",
+ "der"
+ ],
+ [
+ "▁b",
+ "right"
+ ],
+ [
+ "▁br",
+ "ight"
+ ],
+ [
+ "▁brig",
+ "ht"
+ ],
+ [
+ "▁t",
+ "rav"
+ ],
+ [
+ "▁tr",
+ "av"
+ ],
+ [
+ "▁tra",
+ "v"
+ ],
+ [
+ "ic",
+ "ans"
+ ],
+ [
+ "ica",
+ "ns"
+ ],
+ [
+ "ican",
+ "s"
+ ],
+ [
+ "▁att",
+ "ribute"
+ ],
+ [
+ "▁attribut",
+ "e"
+ ],
+ [
+ "▁",
+ "attribute"
+ ],
+ [
+ "f",
+ "c"
+ ],
+ [
+ "▁pl",
+ "aced"
+ ],
+ [
+ "▁place",
+ "d"
+ ],
+ [
+ "▁plac",
+ "ed"
+ ],
+ [
+ "ES",
+ "T"
+ ],
+ [
+ "E",
+ "ST"
+ ],
+ [
+ "Var",
+ "i"
+ ],
+ [
+ "V",
+ "ari"
+ ],
+ [
+ "▁c",
+ "os"
+ ],
+ [
+ "▁co",
+ "s"
+ ],
+ [
+ "▁",
+ "cos"
+ ],
+ [
+ "▁at",
+ "tract"
+ ],
+ [
+ "▁att",
+ "ract"
+ ],
+ [
+ "▁attr",
+ "act"
+ ],
+ [
+ "▁attra",
+ "ct"
+ ],
+ [
+ "an",
+ "el"
+ ],
+ [
+ "ane",
+ "l"
+ ],
+ [
+ "a",
+ "nel"
+ ],
+ [
+ "})",
+ "."
+ ],
+ [
+ "}",
+ ")."
+ ],
+ [
+ "by",
+ "tes"
+ ],
+ [
+ "byte",
+ "s"
+ ],
+ [
+ "▁p",
+ "arse"
+ ],
+ [
+ "▁par",
+ "se"
+ ],
+ [
+ "▁",
+ "parse"
+ ],
+ [
+ "▁be",
+ "long"
+ ],
+ [
+ "▁bel",
+ "ong"
+ ],
+ [
+ "B",
+ "N"
+ ],
+ [
+ "▁S",
+ "ol"
+ ],
+ [
+ "▁So",
+ "l"
+ ],
+ [
+ "P",
+ "o"
+ ],
+ [
+ "`",
+ ","
+ ],
+ [
+ "▁c",
+ "alling"
+ ],
+ [
+ "▁call",
+ "ing"
+ ],
+ [
+ "▁cal",
+ "ling"
+ ],
+ [
+ "▁?",
+ ">"
+ ],
+ [
+ "▁",
+ "?>"
+ ],
+ [
+ "▁it",
+ "er"
+ ],
+ [
+ "▁i",
+ "ter"
+ ],
+ [
+ "▁",
+ "iter"
+ ],
+ [
+ "▁u",
+ "rl"
+ ],
+ [
+ "▁ur",
+ "l"
+ ],
+ [
+ "▁",
+ "url"
+ ],
+ [
+ "▁ev",
+ "ening"
+ ],
+ [
+ "▁even",
+ "ing"
+ ],
+ [
+ "re",
+ "ek"
+ ],
+ [
+ "ree",
+ "k"
+ ],
+ [
+ "▁hon",
+ "est"
+ ],
+ [
+ "▁direct",
+ "or"
+ ],
+ [
+ "▁dire",
+ "ctor"
+ ],
+ [
+ "▁dir",
+ "ector"
+ ],
+ [
+ "R",
+ "C"
+ ],
+ [
+ "▁s",
+ "olid"
+ ],
+ [
+ "▁sol",
+ "id"
+ ],
+ [
+ "▁",
+ "solid"
+ ],
+ [
+ "▁ph",
+ "il"
+ ],
+ [
+ "ie",
+ "ne"
+ ],
+ [
+ "ien",
+ "e"
+ ],
+ [
+ "i",
+ "ene"
+ ],
+ [
+ "FA",
+ "ULT"
+ ],
+ [
+ "co",
+ "pe"
+ ],
+ [
+ "cop",
+ "e"
+ ],
+ [
+ "c",
+ "ope"
+ ],
+ [
+ "▁Hist",
+ "ory"
+ ],
+ [
+ "▁Histor",
+ "y"
+ ],
+ [
+ "▁Hi",
+ "story"
+ ],
+ [
+ "▁",
+ "History"
+ ],
+ [
+ "▁Te",
+ "am"
+ ],
+ [
+ "▁",
+ "Team"
+ ],
+ [
+ "ree",
+ "dom"
+ ],
+ [
+ "reed",
+ "om"
+ ],
+ [
+ "▁r",
+ "u"
+ ],
+ [
+ "▁",
+ "ru"
+ ],
+ [
+ "U",
+ "B"
+ ],
+ [
+ "▁w",
+ "orse"
+ ],
+ [
+ "▁wor",
+ "se"
+ ],
+ [
+ "im",
+ "o"
+ ],
+ [
+ "i",
+ "mo"
+ ],
+ [
+ "Ma",
+ "t"
+ ],
+ [
+ "M",
+ "at"
+ ],
+ [
+ "▁M",
+ "ex"
+ ],
+ [
+ "▁Me",
+ "x"
+ ],
+ [
+ "ac",
+ "tor"
+ ],
+ [
+ "act",
+ "or"
+ ],
+ [
+ "a",
+ "ctor"
+ ],
+ [
+ "▁v",
+ "or"
+ ],
+ [
+ "▁vo",
+ "r"
+ ],
+ [
+ "▁",
+ "vor"
+ ],
+ [
+ "ть",
+ "ся"
+ ],
+ [
+ "▁exper",
+ "iment"
+ ],
+ [
+ "▁experi",
+ "ment"
+ ],
+ [
+ "▁P",
+ "lay"
+ ],
+ [
+ "▁Pl",
+ "ay"
+ ],
+ [
+ "▁",
+ "Play"
+ ],
+ [
+ "▁An",
+ "other"
+ ],
+ [
+ "▁happ",
+ "ens"
+ ],
+ [
+ "▁happen",
+ "s"
+ ],
+ [
+ "ua",
+ "n"
+ ],
+ [
+ "u",
+ "an"
+ ],
+ [
+ "▁pat",
+ "ients"
+ ],
+ [
+ "▁patient",
+ "s"
+ ],
+ [
+ "▁re",
+ "nd"
+ ],
+ [
+ "▁r",
+ "end"
+ ],
+ [
+ "▁ren",
+ "d"
+ ],
+ [
+ "▁",
+ "rend"
+ ],
+ [
+ "▁M",
+ "o"
+ ],
+ [
+ "▁",
+ "Mo"
+ ],
+ [
+ "▁T",
+ "ex"
+ ],
+ [
+ "▁Te",
+ "x"
+ ],
+ [
+ "▁",
+ "Tex"
+ ],
+ [
+ "▁w",
+ "ed"
+ ],
+ [
+ "▁we",
+ "d"
+ ],
+ [
+ "▁",
+ "wed"
+ ],
+ [
+ "t",
+ "n"
+ ],
+ [
+ "in",
+ "sert"
+ ],
+ [
+ "ins",
+ "ert"
+ ],
+ [
+ "▁п",
+ "а"
+ ],
+ [
+ "▁",
+ "па"
+ ],
+ [
+ "▁an",
+ "ti"
+ ],
+ [
+ "▁ant",
+ "i"
+ ],
+ [
+ "▁",
+ "anti"
+ ],
+ [
+ "Mat",
+ "ch"
+ ],
+ [
+ "M",
+ "atch"
+ ],
+ [
+ "ampions",
+ "hip"
+ ],
+ [
+ "ampion",
+ "ship"
+ ],
+ [
+ "▁for",
+ "ces"
+ ],
+ [
+ "▁force",
+ "s"
+ ],
+ [
+ "▁H",
+ "ot"
+ ],
+ [
+ "▁Ho",
+ "t"
+ ],
+ [
+ "▁",
+ "Hot"
+ ],
+ [
+ "▁ph",
+ "ase"
+ ],
+ [
+ "▁",
+ "phase"
+ ],
+ [
+ "▁t",
+ "emplate"
+ ],
+ [
+ "▁templ",
+ "ate"
+ ],
+ [
+ "▁temp",
+ "late"
+ ],
+ [
+ "▁",
+ "template"
+ ],
+ [
+ "st",
+ "op"
+ ],
+ [
+ "sto",
+ "p"
+ ],
+ [
+ "s",
+ "top"
+ ],
+ [
+ "ic",
+ "ated"
+ ],
+ [
+ "ica",
+ "ted"
+ ],
+ [
+ "icate",
+ "d"
+ ],
+ [
+ "▁man",
+ "aged"
+ ],
+ [
+ "▁manage",
+ "d"
+ ],
+ [
+ "▁",
+ "managed"
+ ],
+ [
+ "wa",
+ "it"
+ ],
+ [
+ "w",
+ "ait"
+ ],
+ [
+ "▁*",
+ "("
+ ],
+ [
+ "▁",
+ "*("
+ ],
+ [
+ "G",
+ "B"
+ ],
+ [
+ "▁app",
+ "oint"
+ ],
+ [
+ "▁ap",
+ "point"
+ ],
+ [
+ "▁",
+ "appoint"
+ ],
+ [
+ "ł",
+ "a"
+ ],
+ [
+ "▁s",
+ "tick"
+ ],
+ [
+ "▁st",
+ "ick"
+ ],
+ [
+ "▁",
+ "stick"
+ ],
+ [
+ "▁F",
+ "OR"
+ ],
+ [
+ "▁FO",
+ "R"
+ ],
+ [
+ "▁",
+ "FOR"
+ ],
+ [
+ "▁V",
+ "is"
+ ],
+ [
+ "▁Vi",
+ "s"
+ ],
+ [
+ "▁",
+ "Vis"
+ ],
+ [
+ "to",
+ "r"
+ ],
+ [
+ "t",
+ "or"
+ ],
+ [
+ "▁p",
+ "ř"
+ ],
+ [
+ "qu",
+ "est"
+ ],
+ [
+ "que",
+ "st"
+ ],
+ [
+ "ques",
+ "t"
+ ],
+ [
+ "q",
+ "uest"
+ ],
+ [
+ "us",
+ "es"
+ ],
+ [
+ "use",
+ "s"
+ ],
+ [
+ "u",
+ "ses"
+ ],
+ [
+ "\");",
+ "\r"
+ ],
+ [
+ "\")",
+ ";\r"
+ ],
+ [
+ "\"",
+ ");\r"
+ ],
+ [
+ "▁sudden",
+ "ly"
+ ],
+ [
+ "▁sud",
+ "denly"
+ ],
+ [
+ "é",
+ "c"
+ ],
+ [
+ "N",
+ "D"
+ ],
+ [
+ "ur",
+ "op"
+ ],
+ [
+ "uro",
+ "p"
+ ],
+ [
+ "u",
+ "rop"
+ ],
+ [
+ "ре",
+ "д"
+ ],
+ [
+ "▁ins",
+ "urance"
+ ],
+ [
+ "ac",
+ "cess"
+ ],
+ [
+ "acc",
+ "ess"
+ ],
+ [
+ "a",
+ "ccess"
+ ],
+ [
+ "un",
+ "finished"
+ ],
+ [
+ "▁t",
+ "amb"
+ ],
+ [
+ "▁ta",
+ "mb"
+ ],
+ [
+ "▁tam",
+ "b"
+ ],
+ [
+ "▁s",
+ "ac"
+ ],
+ [
+ "▁sa",
+ "c"
+ ],
+ [
+ "▁C",
+ "ourt"
+ ],
+ [
+ "▁Co",
+ "urt"
+ ],
+ [
+ "▁Cour",
+ "t"
+ ],
+ [
+ "▁Cou",
+ "rt"
+ ],
+ [
+ "▁miss",
+ "ing"
+ ],
+ [
+ "▁mis",
+ "sing"
+ ],
+ [
+ "▁",
+ "missing"
+ ],
+ [
+ "▁W",
+ "here"
+ ],
+ [
+ "▁Wh",
+ "ere"
+ ],
+ [
+ "▁Whe",
+ "re"
+ ],
+ [
+ "▁",
+ "Where"
+ ],
+ [
+ "▁S",
+ "um"
+ ],
+ [
+ "▁Su",
+ "m"
+ ],
+ [
+ "▁",
+ "Sum"
+ ],
+ [
+ "}^",
+ "{\\"
+ ],
+ [
+ "}^{",
+ "\\"
+ ],
+ [
+ "}",
+ "^{\\"
+ ],
+ [
+ "▁s",
+ "ua"
+ ],
+ [
+ "▁su",
+ "a"
+ ],
+ [
+ "_",
+ ","
+ ],
+ [
+ "▁th",
+ "ick"
+ ],
+ [
+ "▁Tr",
+ "ump"
+ ],
+ [
+ "▁Tru",
+ "mp"
+ ],
+ [
+ "▁oper",
+ "ations"
+ ],
+ [
+ "▁operation",
+ "s"
+ ],
+ [
+ "▁",
+ "operations"
+ ],
+ [
+ "F",
+ "S"
+ ],
+ [
+ "▁de",
+ "ux"
+ ],
+ [
+ "d",
+ "z"
+ ],
+ [
+ "Temp",
+ "late"
+ ],
+ [
+ "T",
+ "emplate"
+ ],
+ [
+ "▁\"",
+ "/"
+ ],
+ [
+ "▁o",
+ "dd"
+ ],
+ [
+ "▁od",
+ "d"
+ ],
+ [
+ "▁",
+ "odd"
+ ],
+ [
+ "▁re",
+ "ality"
+ ],
+ [
+ "▁real",
+ "ity"
+ ],
+ [
+ "▁te",
+ "ams"
+ ],
+ [
+ "▁team",
+ "s"
+ ],
+ [
+ "▁tea",
+ "ms"
+ ],
+ [
+ "▁c",
+ "er"
+ ],
+ [
+ "▁ce",
+ "r"
+ ],
+ [
+ "▁",
+ "cer"
+ ],
+ [
+ "om",
+ "a"
+ ],
+ [
+ "o",
+ "ma"
+ ],
+ [
+ "▁",
+ "și"
+ ],
+ [
+ "▁cl",
+ "oud"
+ ],
+ [
+ "▁clo",
+ "ud"
+ ],
+ [
+ "▁",
+ "cloud"
+ ],
+ [
+ "▁Dep",
+ "artment"
+ ],
+ [
+ "N",
+ "e"
+ ],
+ [
+ "▁requ",
+ "ires"
+ ],
+ [
+ "▁require",
+ "s"
+ ],
+ [
+ "it",
+ "ems"
+ ],
+ [
+ "ite",
+ "ms"
+ ],
+ [
+ "item",
+ "s"
+ ],
+ [
+ "▁I",
+ "II"
+ ],
+ [
+ "▁II",
+ "I"
+ ],
+ [
+ "▁",
+ "III"
+ ],
+ [
+ "right",
+ "arrow"
+ ],
+ [
+ ")-",
+ ">"
+ ],
+ [
+ ")",
+ "->"
+ ],
+ [
+ "▁w",
+ "riter"
+ ],
+ [
+ "▁wr",
+ "iter"
+ ],
+ [
+ "▁writ",
+ "er"
+ ],
+ [
+ "▁write",
+ "r"
+ ],
+ [
+ "▁",
+ "writer"
+ ],
+ [
+ "re",
+ "place"
+ ],
+ [
+ "rep",
+ "lace"
+ ],
+ [
+ "▁t",
+ "hr"
+ ],
+ [
+ "▁th",
+ "r"
+ ],
+ [
+ "je",
+ "n"
+ ],
+ [
+ "j",
+ "en"
+ ],
+ [
+ "▁o",
+ "t"
+ ],
+ [
+ "▁",
+ "ot"
+ ],
+ [
+ "▁occ",
+ "up"
+ ],
+ [
+ "▁oc",
+ "cup"
+ ],
+ [
+ "▁",
+ "occup"
+ ],
+ [
+ "▁event",
+ "ually"
+ ],
+ [
+ "▁M",
+ "ath"
+ ],
+ [
+ "▁Mat",
+ "h"
+ ],
+ [
+ "▁Ma",
+ "th"
+ ],
+ [
+ "▁",
+ "Math"
+ ],
+ [
+ "▁con",
+ "serv"
+ ],
+ [
+ "▁cons",
+ "erv"
+ ],
+ [
+ "▁conse",
+ "rv"
+ ],
+ [
+ "am",
+ "er"
+ ],
+ [
+ "ame",
+ "r"
+ ],
+ [
+ "a",
+ "mer"
+ ],
+ [
+ "▁F",
+ "ort"
+ ],
+ [
+ "▁For",
+ "t"
+ ],
+ [
+ "▁Fo",
+ "rt"
+ ],
+ [
+ "▁d",
+ "ry"
+ ],
+ [
+ "▁dr",
+ "y"
+ ],
+ [
+ "▁sex",
+ "ual"
+ ],
+ [
+ "▁co",
+ "sts"
+ ],
+ [
+ "▁cost",
+ "s"
+ ],
+ [
+ "▁cos",
+ "ts"
+ ],
+ [
+ "▁for",
+ "ms"
+ ],
+ [
+ "▁form",
+ "s"
+ ],
+ [
+ "▁",
+ "forms"
+ ],
+ [
+ "▁V",
+ "ict"
+ ],
+ [
+ "▁Vi",
+ "ct"
+ ],
+ [
+ "▁Vic",
+ "t"
+ ],
+ [
+ "PA",
+ "R"
+ ],
+ [
+ "P",
+ "AR"
+ ],
+ [
+ "frame",
+ "work"
+ ],
+ [
+ "▁д",
+ "и"
+ ],
+ [
+ "▁",
+ "ди"
+ ],
+ [
+ "Oper",
+ "ation"
+ ],
+ [
+ "з",
+ "на"
+ ],
+ [
+ "wh",
+ "ich"
+ ],
+ [
+ "▁t",
+ "ight"
+ ],
+ [
+ "▁ti",
+ "ght"
+ ],
+ [
+ "In",
+ "valid"
+ ],
+ [
+ "▁part",
+ "ner"
+ ],
+ [
+ "▁п",
+ "ред"
+ ],
+ [
+ "▁пре",
+ "д"
+ ],
+ [
+ "▁th",
+ "ank"
+ ],
+ [
+ "▁than",
+ "k"
+ ],
+ [
+ "▁gu",
+ "ard"
+ ],
+ [
+ "▁",
+ "guard"
+ ],
+ [
+ "he",
+ "m"
+ ],
+ [
+ "h",
+ "em"
+ ],
+ [
+ "Bo",
+ "dy"
+ ],
+ [
+ "B",
+ "ody"
+ ],
+ [
+ "▁e",
+ "mot"
+ ],
+ [
+ "▁em",
+ "ot"
+ ],
+ [
+ "I",
+ "X"
+ ],
+ [
+ "fa",
+ "st"
+ ],
+ [
+ "fas",
+ "t"
+ ],
+ [
+ "f",
+ "ast"
+ ],
+ [
+ "щ",
+ "о"
+ ],
+ [
+ "ñ",
+ "o"
+ ],
+ [
+ "ni",
+ "ght"
+ ],
+ [
+ "n",
+ "ight"
+ ],
+ [
+ "▁S",
+ "ci"
+ ],
+ [
+ "▁Sc",
+ "i"
+ ],
+ [
+ "ни",
+ "ка"
+ ],
+ [
+ "ник",
+ "а"
+ ],
+ [
+ "▁T",
+ "O"
+ ],
+ [
+ "▁",
+ "TO"
+ ],
+ [
+ "▁individ",
+ "uals"
+ ],
+ [
+ "▁individual",
+ "s"
+ ],
+ [
+ "сс",
+ "и"
+ ],
+ [
+ "с",
+ "си"
+ ],
+ [
+ "})",
+ ","
+ ],
+ [
+ "}",
+ "),"
+ ],
+ [
+ "F",
+ "alse"
+ ],
+ [
+ "(\"",
+ "%"
+ ],
+ [
+ "(",
+ "\"%"
+ ],
+ [
+ "▁op",
+ "tim"
+ ],
+ [
+ "▁opt",
+ "im"
+ ],
+ [
+ "▁",
+ "optim"
+ ],
+ [
+ "▁-",
+ "->"
+ ],
+ [
+ "▁--",
+ ">"
+ ],
+ [
+ "▁",
+ "-->"
+ ],
+ [
+ "▁f",
+ "actor"
+ ],
+ [
+ "▁fact",
+ "or"
+ ],
+ [
+ "▁fac",
+ "tor"
+ ],
+ [
+ "▁fa",
+ "ctor"
+ ],
+ [
+ "▁",
+ "factor"
+ ],
+ [
+ "▁sm",
+ "aller"
+ ],
+ [
+ "▁small",
+ "er"
+ ],
+ [
+ "▁con",
+ "tain"
+ ],
+ [
+ "▁cont",
+ "ain"
+ ],
+ [
+ "sp",
+ "ect"
+ ],
+ [
+ "spec",
+ "t"
+ ],
+ [
+ "spe",
+ "ct"
+ ],
+ [
+ "s",
+ "pect"
+ ],
+ [
+ "Eng",
+ "ine"
+ ],
+ [
+ "▁ann",
+ "ounced"
+ ],
+ [
+ "▁announ",
+ "ced"
+ ],
+ [
+ "▁announce",
+ "d"
+ ],
+ [
+ "▁Dem",
+ "ocr"
+ ],
+ [
+ "▁r",
+ "ob"
+ ],
+ [
+ "▁ro",
+ "b"
+ ],
+ [
+ "▁",
+ "rob"
+ ],
+ [
+ "▁f",
+ "lat"
+ ],
+ [
+ "▁fl",
+ "at"
+ ],
+ [
+ "▁",
+ "flat"
+ ],
+ [
+ "os",
+ "oph"
+ ],
+ [
+ "oso",
+ "ph"
+ ],
+ [
+ "Se",
+ "arch"
+ ],
+ [
+ "S",
+ "earch"
+ ],
+ [
+ "ah",
+ "l"
+ ],
+ [
+ "a",
+ "hl"
+ ],
+ [
+ "▁Ex",
+ "ception"
+ ],
+ [
+ "▁Except",
+ "ion"
+ ],
+ [
+ "▁",
+ "Exception"
+ ],
+ [
+ "▁O",
+ "l"
+ ],
+ [
+ "equ",
+ "als"
+ ],
+ [
+ "eq",
+ "uals"
+ ],
+ [
+ "equal",
+ "s"
+ ],
+ [
+ "▁un",
+ "ter"
+ ],
+ [
+ "▁unt",
+ "er"
+ ],
+ [
+ "▁",
+ "unter"
+ ],
+ [
+ "sh",
+ "ape"
+ ],
+ [
+ "sha",
+ "pe"
+ ],
+ [
+ "N",
+ "S"
+ ],
+ [
+ "Ob",
+ "j"
+ ],
+ [
+ "▁spec",
+ "ies"
+ ],
+ [
+ "▁spe",
+ "cies"
+ ],
+ [
+ "we",
+ "ight"
+ ],
+ [
+ "wei",
+ "ght"
+ ],
+ [
+ "w",
+ "eight"
+ ],
+ [
+ "yo",
+ "u"
+ ],
+ [
+ "y",
+ "ou"
+ ],
+ [
+ "▁e",
+ "ste"
+ ],
+ [
+ "▁est",
+ "e"
+ ],
+ [
+ "▁es",
+ "te"
+ ],
+ [
+ "▁",
+ "este"
+ ],
+ [
+ "▁V",
+ "iew"
+ ],
+ [
+ "▁Vi",
+ "ew"
+ ],
+ [
+ "▁",
+ "View"
+ ],
+ [
+ "▁m",
+ "ission"
+ ],
+ [
+ "▁miss",
+ "ion"
+ ],
+ [
+ "▁",
+ "mission"
+ ],
+ [
+ "▁j",
+ "ournal"
+ ],
+ [
+ "▁jour",
+ "nal"
+ ],
+ [
+ "▁",
+ "journal"
+ ],
+ [
+ "Value",
+ "s"
+ ],
+ [
+ "Val",
+ "ues"
+ ],
+ [
+ "▁ein",
+ "em"
+ ],
+ [
+ "▁eine",
+ "m"
+ ],
+ [
+ "is",
+ "mo"
+ ],
+ [
+ "ism",
+ "o"
+ ],
+ [
+ "▁project",
+ "s"
+ ],
+ [
+ "▁",
+ "projects"
+ ],
+ [
+ "▁D",
+ "as"
+ ],
+ [
+ "▁Da",
+ "s"
+ ],
+ [
+ "ri",
+ "ble"
+ ],
+ [
+ "rib",
+ "le"
+ ],
+ [
+ "r",
+ "ible"
+ ],
+ [
+ "▁s",
+ "erve"
+ ],
+ [
+ "▁ser",
+ "ve"
+ ],
+ [
+ "▁serv",
+ "e"
+ ],
+ [
+ "▁",
+ "serve"
+ ],
+ [
+ "▁op",
+ "ening"
+ ],
+ [
+ "▁open",
+ "ing"
+ ],
+ [
+ "▁h",
+ "ur"
+ ],
+ [
+ "▁program",
+ "s"
+ ],
+ [
+ "▁U",
+ "SA"
+ ],
+ [
+ "▁US",
+ "A"
+ ],
+ [
+ "▁",
+ "USA"
+ ],
+ [
+ "il",
+ "iar"
+ ],
+ [
+ "ili",
+ "ar"
+ ],
+ [
+ "ilia",
+ "r"
+ ],
+ [
+ "id",
+ "os"
+ ],
+ [
+ "ido",
+ "s"
+ ],
+ [
+ "B",
+ "r"
+ ],
+ [
+ "est",
+ "amp"
+ ],
+ [
+ "esta",
+ "mp"
+ ],
+ [
+ "▁t",
+ "ools"
+ ],
+ [
+ "▁to",
+ "ols"
+ ],
+ [
+ "▁too",
+ "ls"
+ ],
+ [
+ "▁tool",
+ "s"
+ ],
+ [
+ "▁",
+ "tools"
+ ],
+ [
+ "an",
+ "ner"
+ ],
+ [
+ "ann",
+ "er"
+ ],
+ [
+ "anne",
+ "r"
+ ],
+ [
+ "R",
+ "T"
+ ],
+ [
+ "▁St",
+ "art"
+ ],
+ [
+ "▁Star",
+ "t"
+ ],
+ [
+ "▁Sta",
+ "rt"
+ ],
+ [
+ "▁",
+ "Start"
+ ],
+ [
+ "▁b",
+ "ath"
+ ],
+ [
+ "▁bat",
+ "h"
+ ],
+ [
+ "▁ba",
+ "th"
+ ],
+ [
+ "▁coff",
+ "ee"
+ ],
+ [
+ "or",
+ "ter"
+ ],
+ [
+ "ort",
+ "er"
+ ],
+ [
+ "orte",
+ "r"
+ ],
+ [
+ "in",
+ "ternal"
+ ],
+ [
+ "inter",
+ "nal"
+ ],
+ [
+ "intern",
+ "al"
+ ],
+ [
+ "file",
+ "s"
+ ],
+ [
+ "fil",
+ "es"
+ ],
+ [
+ "fi",
+ "les"
+ ],
+ [
+ "f",
+ "iles"
+ ],
+ [
+ "IN",
+ "VAL"
+ ],
+ [
+ "ak",
+ "o"
+ ],
+ [
+ "a",
+ "ko"
+ ],
+ [
+ "d",
+ "t"
+ ],
+ [
+ "▁Se",
+ "cond"
+ ],
+ [
+ "▁Sec",
+ "ond"
+ ],
+ [
+ "▁",
+ "Second"
+ ],
+ [
+ "▁al",
+ "loc"
+ ],
+ [
+ "▁all",
+ "oc"
+ ],
+ [
+ "▁",
+ "alloc"
+ ],
+ [
+ "▁en",
+ "ded"
+ ],
+ [
+ "▁end",
+ "ed"
+ ],
+ [
+ "▁ende",
+ "d"
+ ],
+ [
+ "▁",
+ "ended"
+ ],
+ [
+ "ac",
+ "ional"
+ ],
+ [
+ "aci",
+ "onal"
+ ],
+ [
+ "acion",
+ "al"
+ ],
+ [
+ "acio",
+ "nal"
+ ],
+ [
+ "▁man",
+ "ager"
+ ],
+ [
+ "▁manage",
+ "r"
+ ],
+ [
+ "▁",
+ "manager"
+ ],
+ [
+ "▁S",
+ "un"
+ ],
+ [
+ "▁Su",
+ "n"
+ ],
+ [
+ "▁",
+ "Sun"
+ ],
+ [
+ "ag",
+ "g"
+ ],
+ [
+ "a",
+ "gg"
+ ],
+ [
+ "▁le",
+ "ader"
+ ],
+ [
+ "▁lead",
+ "er"
+ ],
+ [
+ "ol",
+ "ved"
+ ],
+ [
+ "olve",
+ "d"
+ ],
+ [
+ "olv",
+ "ed"
+ ],
+ [
+ "▁ч",
+ "то"
+ ],
+ [
+ "▁trad",
+ "itional"
+ ],
+ [
+ "▁tradition",
+ "al"
+ ],
+ [
+ "sh",
+ "ot"
+ ],
+ [
+ "s",
+ "hot"
+ ],
+ [
+ "ru",
+ "p"
+ ],
+ [
+ "r",
+ "up"
+ ],
+ [
+ "C",
+ "F"
+ ],
+ [
+ "▁E",
+ "ach"
+ ],
+ [
+ "▁",
+ "Each"
+ ],
+ [
+ "w",
+ "r"
+ ],
+ [
+ "▁S",
+ "om"
+ ],
+ [
+ "▁So",
+ "m"
+ ],
+ [
+ "▁",
+ "Som"
+ ],
+ [
+ "▁material",
+ "s"
+ ],
+ [
+ "▁mater",
+ "ials"
+ ],
+ [
+ "▁m",
+ "sg"
+ ],
+ [
+ "▁ms",
+ "g"
+ ],
+ [
+ "▁",
+ "msg"
+ ],
+ [
+ "▁s",
+ "yn"
+ ],
+ [
+ "▁sy",
+ "n"
+ ],
+ [
+ "▁",
+ "syn"
+ ],
+ [
+ "▁produ",
+ "ce"
+ ],
+ [
+ "▁prod",
+ "uce"
+ ],
+ [
+ "▁st",
+ "orage"
+ ],
+ [
+ "▁stor",
+ "age"
+ ],
+ [
+ "▁sto",
+ "rage"
+ ],
+ [
+ "▁",
+ "storage"
+ ],
+ [
+ "sub",
+ "section"
+ ],
+ [
+ "▁S",
+ "ie"
+ ],
+ [
+ "▁Si",
+ "e"
+ ],
+ [
+ "▁I",
+ "P"
+ ],
+ [
+ "▁",
+ "IP"
+ ],
+ [
+ "CE",
+ "SS"
+ ],
+ [
+ "▁w",
+ "a"
+ ],
+ [
+ "▁",
+ "wa"
+ ],
+ [
+ "Re",
+ "cord"
+ ],
+ [
+ "Rec",
+ "ord"
+ ],
+ [
+ "▁mark",
+ "eting"
+ ],
+ [
+ "▁market",
+ "ing"
+ ],
+ [
+ "pl",
+ "et"
+ ],
+ [
+ "ple",
+ "t"
+ ],
+ [
+ "p",
+ "let"
+ ],
+ [
+ "D",
+ "ialog"
+ ],
+ [
+ "▁mention",
+ "ed"
+ ],
+ [
+ "▁ment",
+ "ioned"
+ ],
+ [
+ "▁N",
+ "a"
+ ],
+ [
+ "▁",
+ "Na"
+ ],
+ [
+ "▁Un",
+ "ion"
+ ],
+ [
+ "▁",
+ "Union"
+ ],
+ [
+ "▁A",
+ "PI"
+ ],
+ [
+ "▁AP",
+ "I"
+ ],
+ [
+ "▁",
+ "API"
+ ],
+ [
+ "▁neg",
+ "ative"
+ ],
+ [
+ "▁",
+ "negative"
+ ],
+ [
+ "tx",
+ "t"
+ ],
+ [
+ "t",
+ "xt"
+ ],
+ [
+ "▁eas",
+ "ier"
+ ],
+ [
+ "le",
+ "gal"
+ ],
+ [
+ "leg",
+ "al"
+ ],
+ [
+ "De",
+ "p"
+ ],
+ [
+ "D",
+ "ep"
+ ],
+ [
+ "▁no",
+ "vel"
+ ],
+ [
+ "▁nov",
+ "el"
+ ],
+ [
+ "▁nove",
+ "l"
+ ],
+ [
+ "eu",
+ "r"
+ ],
+ [
+ "e",
+ "ur"
+ ],
+ [
+ "ac",
+ "ió"
+ ],
+ [
+ "aci",
+ "ó"
+ ],
+ [
+ "a",
+ "ció"
+ ],
+ [
+ "▁B",
+ "ud"
+ ],
+ [
+ "▁Bu",
+ "d"
+ ],
+ [
+ "▁c",
+ "arry"
+ ],
+ [
+ "▁car",
+ "ry"
+ ],
+ [
+ "sch",
+ "aft"
+ ],
+ [
+ "s",
+ "chaft"
+ ],
+ [
+ "▁br",
+ "oken"
+ ],
+ [
+ "▁bro",
+ "ken"
+ ],
+ [
+ "▁broke",
+ "n"
+ ],
+ [
+ "▁t",
+ "rees"
+ ],
+ [
+ "▁tr",
+ "ees"
+ ],
+ [
+ "▁tre",
+ "es"
+ ],
+ [
+ "▁tree",
+ "s"
+ ],
+ [
+ ">(",
+ ");"
+ ],
+ [
+ ">()",
+ ";"
+ ],
+ [
+ ">",
+ "();"
+ ],
+ [
+ "▁e",
+ "mb"
+ ],
+ [
+ "▁em",
+ "b"
+ ],
+ [
+ "▁",
+ "emb"
+ ],
+ [
+ "ie",
+ "der"
+ ],
+ [
+ "ied",
+ "er"
+ ],
+ [
+ "i",
+ "eder"
+ ],
+ [
+ "▁r",
+ "oute"
+ ],
+ [
+ "▁ro",
+ "ute"
+ ],
+ [
+ "▁rout",
+ "e"
+ ],
+ [
+ "▁rou",
+ "te"
+ ],
+ [
+ "▁",
+ "route"
+ ],
+ [
+ "ik",
+ "el"
+ ],
+ [
+ "ike",
+ "l"
+ ],
+ [
+ "i",
+ "kel"
+ ],
+ [
+ "▁l",
+ "isten"
+ ],
+ [
+ "▁li",
+ "sten"
+ ],
+ [
+ "▁list",
+ "en"
+ ],
+ [
+ "▁",
+ "listen"
+ ],
+ [
+ "ash",
+ "ion"
+ ],
+ [
+ "ashi",
+ "on"
+ ],
+ [
+ "▁M",
+ "rs"
+ ],
+ [
+ "▁Mr",
+ "s"
+ ],
+ [
+ "▁equip",
+ "ment"
+ ],
+ [
+ "ag",
+ "ger"
+ ],
+ [
+ "agg",
+ "er"
+ ],
+ [
+ "▁T",
+ "hus"
+ ],
+ [
+ "▁Th",
+ "us"
+ ],
+ [
+ "▁mat",
+ "rix"
+ ],
+ [
+ "▁",
+ "matrix"
+ ],
+ [
+ "al",
+ "la"
+ ],
+ [
+ "all",
+ "a"
+ ],
+ [
+ "a",
+ "lla"
+ ],
+ [
+ "▁T",
+ "our"
+ ],
+ [
+ "▁To",
+ "ur"
+ ],
+ [
+ "▁con",
+ "versation"
+ ],
+ [
+ "▁convers",
+ "ation"
+ ],
+ [
+ "Mo",
+ "n"
+ ],
+ [
+ "M",
+ "on"
+ ],
+ [
+ "our",
+ "nal"
+ ],
+ [
+ "▁min",
+ "ute"
+ ],
+ [
+ "▁minut",
+ "e"
+ ],
+ [
+ "▁",
+ "minute"
+ ],
+ [
+ "A",
+ "m"
+ ],
+ [
+ "Ap",
+ "i"
+ ],
+ [
+ "A",
+ "pi"
+ ],
+ [
+ "▁for",
+ "get"
+ ],
+ [
+ "▁forg",
+ "et"
+ ],
+ [
+ "M",
+ "e"
+ ],
+ [
+ "lev",
+ "ant"
+ ],
+ [
+ "te",
+ "mp"
+ ],
+ [
+ "tem",
+ "p"
+ ],
+ [
+ "t",
+ "emp"
+ ],
+ [
+ "▁t",
+ "elling"
+ ],
+ [
+ "▁tell",
+ "ing"
+ ],
+ [
+ "▁tel",
+ "ling"
+ ],
+ [
+ "mo",
+ "ve"
+ ],
+ [
+ "mov",
+ "e"
+ ],
+ [
+ "m",
+ "ove"
+ ],
+ [
+ "▁in",
+ "dependent"
+ ],
+ [
+ "▁independ",
+ "ent"
+ ],
+ [
+ "to",
+ "String"
+ ],
+ [
+ "ed",
+ "it"
+ ],
+ [
+ "edi",
+ "t"
+ ],
+ [
+ "e",
+ "dit"
+ ],
+ [
+ "▁J",
+ "ac"
+ ],
+ [
+ "▁Ja",
+ "c"
+ ],
+ [
+ "az",
+ "z"
+ ],
+ [
+ "a",
+ "zz"
+ ],
+ [
+ "re",
+ "act"
+ ],
+ [
+ "rea",
+ "ct"
+ ],
+ [
+ "▁c",
+ "in"
+ ],
+ [
+ "▁ci",
+ "n"
+ ],
+ [
+ "▁",
+ "cin"
+ ],
+ [
+ "▁P",
+ "rov"
+ ],
+ [
+ "▁Pro",
+ "v"
+ ],
+ [
+ "▁Pr",
+ "ov"
+ ],
+ [
+ "▁",
+ "Prov"
+ ],
+ [
+ "is",
+ "ted"
+ ],
+ [
+ "ist",
+ "ed"
+ ],
+ [
+ "iste",
+ "d"
+ ],
+ [
+ "i",
+ "sted"
+ ],
+ [
+ "▁h",
+ "ash"
+ ],
+ [
+ "▁has",
+ "h"
+ ],
+ [
+ "▁ha",
+ "sh"
+ ],
+ [
+ "▁",
+ "hash"
+ ],
+ [
+ "on",
+ "na"
+ ],
+ [
+ "ik",
+ "i"
+ ],
+ [
+ "i",
+ "ki"
+ ],
+ [
+ "▁gener",
+ "ated"
+ ],
+ [
+ "▁generate",
+ "d"
+ ],
+ [
+ "▁gene",
+ "rated"
+ ],
+ [
+ "▁",
+ "generated"
+ ],
+ [
+ "Re",
+ "nder"
+ ],
+ [
+ "Rend",
+ "er"
+ ],
+ [
+ "R",
+ "ender"
+ ],
+ [
+ "▁psy",
+ "ch"
+ ],
+ [
+ "▁ps",
+ "ych"
+ ],
+ [
+ "na",
+ "v"
+ ],
+ [
+ "n",
+ "av"
+ ],
+ [
+ "▁en",
+ "tr"
+ ],
+ [
+ "▁ent",
+ "r"
+ ],
+ [
+ "▁",
+ "entr"
+ ],
+ [
+ "п",
+ "ра"
+ ],
+ [
+ "r",
+ "x"
+ ],
+ [
+ "AT",
+ "H"
+ ],
+ [
+ "A",
+ "TH"
+ ],
+ [
+ "▁ass",
+ "ume"
+ ],
+ [
+ "▁assum",
+ "e"
+ ],
+ [
+ "Tr",
+ "ee"
+ ],
+ [
+ "T",
+ "ree"
+ ],
+ [
+ "semb",
+ "ly"
+ ],
+ [
+ "sembl",
+ "y"
+ ],
+ [
+ "▁M",
+ "att"
+ ],
+ [
+ "▁Mat",
+ "t"
+ ],
+ [
+ "▁Ma",
+ "tt"
+ ],
+ [
+ "ca",
+ "ption"
+ ],
+ [
+ "c",
+ "aption"
+ ],
+ [
+ "▁s",
+ "olutions"
+ ],
+ [
+ "▁solution",
+ "s"
+ ],
+ [
+ "▁fa",
+ "ith"
+ ],
+ [
+ "▁fait",
+ "h"
+ ],
+ [
+ "▁dig",
+ "ital"
+ ],
+ [
+ "▁digit",
+ "al"
+ ],
+ [
+ "▁ex",
+ "cell"
+ ],
+ [
+ "▁exc",
+ "ell"
+ ],
+ [
+ "▁V",
+ "ersion"
+ ],
+ [
+ "▁Vers",
+ "ion"
+ ],
+ [
+ "▁",
+ "Version"
+ ],
+ [
+ "De",
+ "bug"
+ ],
+ [
+ "D",
+ "ebug"
+ ],
+ [
+ "▁ж",
+ "и"
+ ],
+ [
+ "▁",
+ "жи"
+ ],
+ [
+ "▁car",
+ "ried"
+ ],
+ [
+ "re",
+ "set"
+ ],
+ [
+ "res",
+ "et"
+ ],
+ [
+ "▁slow",
+ "ly"
+ ],
+ [
+ "an",
+ "cing"
+ ],
+ [
+ "anc",
+ "ing"
+ ],
+ [
+ "▁own",
+ "er"
+ ],
+ [
+ "▁",
+ "owner"
+ ],
+ [
+ "▁T",
+ "er"
+ ],
+ [
+ "▁Te",
+ "r"
+ ],
+ [
+ "▁D",
+ "id"
+ ],
+ [
+ "▁Di",
+ "d"
+ ],
+ [
+ "▁",
+ "Did"
+ ],
+ [
+ "▁g",
+ "est"
+ ],
+ [
+ "▁ge",
+ "st"
+ ],
+ [
+ "▁ges",
+ "t"
+ ],
+ [
+ "▁",
+ "gest"
+ ],
+ [
+ "▁é",
+ "té"
+ ],
+ [
+ "▁ét",
+ "é"
+ ],
+ [
+ "▁",
+ "été"
+ ],
+ [
+ "▁pro",
+ "of"
+ ],
+ [
+ "▁",
+ "proof"
+ ],
+ [
+ "F",
+ "ont"
+ ],
+ [
+ "▁n",
+ "ob"
+ ],
+ [
+ "▁no",
+ "b"
+ ],
+ [
+ "▁",
+ "nob"
+ ],
+ [
+ "C",
+ "o"
+ ],
+ [
+ "▁G",
+ "NU"
+ ],
+ [
+ "▁l",
+ "iber"
+ ],
+ [
+ "▁li",
+ "ber"
+ ],
+ [
+ "▁lib",
+ "er"
+ ],
+ [
+ "it",
+ "ness"
+ ],
+ [
+ "▁h",
+ "ij"
+ ],
+ [
+ "▁hi",
+ "j"
+ ],
+ [
+ "▁v",
+ "ert"
+ ],
+ [
+ "▁ver",
+ "t"
+ ],
+ [
+ "▁ve",
+ "rt"
+ ],
+ [
+ "▁",
+ "vert"
+ ],
+ [
+ "ш",
+ "а"
+ ],
+ [
+ "FL",
+ "AG"
+ ],
+ [
+ "ME",
+ "NT"
+ ],
+ [
+ "M",
+ "ENT"
+ ],
+ [
+ "▁S",
+ "on"
+ ],
+ [
+ "▁So",
+ "n"
+ ],
+ [
+ "Mu",
+ "lt"
+ ],
+ [
+ "M",
+ "ult"
+ ],
+ [
+ "▁d",
+ "istrict"
+ ],
+ [
+ "▁di",
+ "strict"
+ ],
+ [
+ "▁dist",
+ "rict"
+ ],
+ [
+ "conne",
+ "ct"
+ ],
+ [
+ "conn",
+ "ect"
+ ],
+ [
+ "ject",
+ "ion"
+ ],
+ [
+ "je",
+ "ction"
+ ],
+ [
+ "j",
+ "ection"
+ ],
+ [
+ "ly",
+ "mp"
+ ],
+ [
+ "▁real",
+ "ized"
+ ],
+ [
+ "▁realize",
+ "d"
+ ],
+ [
+ "▁realiz",
+ "ed"
+ ],
+ [
+ "mo",
+ "s"
+ ],
+ [
+ "m",
+ "os"
+ ],
+ [
+ "y",
+ "e"
+ ],
+ [
+ "▁re",
+ "nder"
+ ],
+ [
+ "▁r",
+ "ender"
+ ],
+ [
+ "▁ren",
+ "der"
+ ],
+ [
+ "▁rend",
+ "er"
+ ],
+ [
+ "▁",
+ "render"
+ ],
+ [
+ "ri",
+ "o"
+ ],
+ [
+ "r",
+ "io"
+ ],
+ [
+ "▁inter",
+ "pret"
+ ],
+ [
+ "▁",
+ "interpret"
+ ],
+ [
+ "▁slight",
+ "ly"
+ ],
+ [
+ "fi",
+ "x"
+ ],
+ [
+ "f",
+ "ix"
+ ],
+ [
+ "▁stud",
+ "ies"
+ ],
+ [
+ "▁r",
+ "id"
+ ],
+ [
+ "▁ri",
+ "d"
+ ],
+ [
+ "▁",
+ "rid"
+ ],
+ [
+ "at",
+ "re"
+ ],
+ [
+ "atr",
+ "e"
+ ],
+ [
+ "a",
+ "tre"
+ ],
+ [
+ "▁benef",
+ "its"
+ ],
+ [
+ "▁benefit",
+ "s"
+ ],
+ [
+ "▁F",
+ "ace"
+ ],
+ [
+ "▁Fa",
+ "ce"
+ ],
+ [
+ "▁Fac",
+ "e"
+ ],
+ [
+ "▁",
+ "Face"
+ ],
+ [
+ "iv",
+ "ery"
+ ],
+ [
+ "ive",
+ "ry"
+ ],
+ [
+ "iver",
+ "y"
+ ],
+ [
+ "i",
+ "very"
+ ],
+ [
+ "ри",
+ "я"
+ ],
+ [
+ "doc",
+ "ument"
+ ],
+ [
+ "d",
+ "ocument"
+ ],
+ [
+ "▁as",
+ "king"
+ ],
+ [
+ "▁ask",
+ "ing"
+ ],
+ [
+ "La",
+ "st"
+ ],
+ [
+ "L",
+ "ast"
+ ],
+ [
+ "ar",
+ "ante"
+ ],
+ [
+ "ara",
+ "nte"
+ ],
+ [
+ "aran",
+ "te"
+ ],
+ [
+ "▁Mart",
+ "in"
+ ],
+ [
+ "▁E",
+ "ll"
+ ],
+ [
+ "▁El",
+ "l"
+ ],
+ [
+ "▁v",
+ "ector"
+ ],
+ [
+ "▁ve",
+ "ctor"
+ ],
+ [
+ "▁vec",
+ "tor"
+ ],
+ [
+ "▁",
+ "vector"
+ ],
+ [
+ "▁for",
+ "ced"
+ ],
+ [
+ "▁force",
+ "d"
+ ],
+ [
+ "▁",
+ "forced"
+ ],
+ [
+ "о",
+ "ло"
+ ],
+ [
+ "P",
+ "H"
+ ],
+ [
+ "W",
+ "R"
+ ],
+ [
+ "▁K",
+ "l"
+ ],
+ [
+ "▁s",
+ "ky"
+ ],
+ [
+ "▁sk",
+ "y"
+ ],
+ [
+ "▁",
+ "sky"
+ ],
+ [
+ "▁str",
+ "ategy"
+ ],
+ [
+ "▁strateg",
+ "y"
+ ],
+ [
+ "▁strat",
+ "egy"
+ ],
+ [
+ "oc",
+ "ked"
+ ],
+ [
+ "ock",
+ "ed"
+ ],
+ [
+ "▁ne",
+ "ck"
+ ],
+ [
+ "ś",
+ "ci"
+ ],
+ [
+ "O",
+ "UT"
+ ],
+ [
+ "))",
+ ","
+ ],
+ [
+ ")",
+ "),"
+ ],
+ [
+ "C",
+ "ustom"
+ ],
+ [
+ "▁w",
+ "ie"
+ ],
+ [
+ "▁",
+ "wie"
+ ],
+ [
+ "▁s",
+ "weet"
+ ],
+ [
+ "▁swe",
+ "et"
+ ],
+ [
+ "▁t",
+ "emp"
+ ],
+ [
+ "▁te",
+ "mp"
+ ],
+ [
+ "▁tem",
+ "p"
+ ],
+ [
+ "▁",
+ "temp"
+ ],
+ [
+ "▁fore",
+ "ign"
+ ],
+ [
+ "▁h",
+ "all"
+ ],
+ [
+ "▁ha",
+ "ll"
+ ],
+ [
+ "▁hal",
+ "l"
+ ],
+ [
+ "▁",
+ "hall"
+ ],
+ [
+ "as",
+ "tr"
+ ],
+ [
+ "ast",
+ "r"
+ ],
+ [
+ "a",
+ "str"
+ ],
+ [
+ "As",
+ "s"
+ ],
+ [
+ "A",
+ "ss"
+ ],
+ [
+ "MO",
+ "DE"
+ ],
+ [
+ "MOD",
+ "E"
+ ],
+ [
+ "▁max",
+ "imum"
+ ],
+ [
+ "▁maxim",
+ "um"
+ ],
+ [
+ "an",
+ "nels"
+ ],
+ [
+ "ann",
+ "els"
+ ],
+ [
+ "annel",
+ "s"
+ ],
+ [
+ "anne",
+ "ls"
+ ],
+ [
+ "▁t",
+ "ip"
+ ],
+ [
+ "▁ti",
+ "p"
+ ],
+ [
+ "▁",
+ "tip"
+ ],
+ [
+ "▁second",
+ "s"
+ ],
+ [
+ "▁sec",
+ "onds"
+ ],
+ [
+ "▁",
+ "seconds"
+ ],
+ [
+ "▁st",
+ "ack"
+ ],
+ [
+ "▁sta",
+ "ck"
+ ],
+ [
+ "▁",
+ "stack"
+ ],
+ [
+ "ig",
+ "a"
+ ],
+ [
+ "i",
+ "ga"
+ ],
+ [
+ "▁r",
+ "aise"
+ ],
+ [
+ "▁rais",
+ "e"
+ ],
+ [
+ "▁ra",
+ "ise"
+ ],
+ [
+ "▁",
+ "raise"
+ ],
+ [
+ "en",
+ "able"
+ ],
+ [
+ "ena",
+ "ble"
+ ],
+ [
+ "oi",
+ "r"
+ ],
+ [
+ "o",
+ "ir"
+ ],
+ [
+ "▁s",
+ "oul"
+ ],
+ [
+ "▁so",
+ "ul"
+ ],
+ [
+ "▁sou",
+ "l"
+ ],
+ [
+ "K",
+ "e"
+ ],
+ [
+ ")$",
+ "."
+ ],
+ [
+ ")",
+ "$."
+ ],
+ [
+ "▁T",
+ "im"
+ ],
+ [
+ "▁Ti",
+ "m"
+ ],
+ [
+ "▁",
+ "Tim"
+ ],
+ [
+ "AL",
+ "SE"
+ ],
+ [
+ "is",
+ "er"
+ ],
+ [
+ "ise",
+ "r"
+ ],
+ [
+ "i",
+ "ser"
+ ],
+ [
+ "cont",
+ "in"
+ ],
+ [
+ "be",
+ "l"
+ ],
+ [
+ "b",
+ "el"
+ ],
+ [
+ "▁m",
+ "ad"
+ ],
+ [
+ "▁ma",
+ "d"
+ ],
+ [
+ "▁",
+ "mad"
+ ],
+ [
+ "lic",
+ "hen"
+ ],
+ [
+ "li",
+ "chen"
+ ],
+ [
+ "lich",
+ "en"
+ ],
+ [
+ "liche",
+ "n"
+ ],
+ [
+ "l",
+ "ichen"
+ ],
+ [
+ "ab",
+ "e"
+ ],
+ [
+ "a",
+ "be"
+ ],
+ [
+ "sa",
+ "fe"
+ ],
+ [
+ "▁con",
+ "cent"
+ ],
+ [
+ "▁conc",
+ "ent"
+ ],
+ [
+ "▁conce",
+ "nt"
+ ],
+ [
+ "bo",
+ "und"
+ ],
+ [
+ "b",
+ "ound"
+ ],
+ [
+ "▁R",
+ "equ"
+ ],
+ [
+ "▁Re",
+ "qu"
+ ],
+ [
+ "▁",
+ "Requ"
+ ],
+ [
+ "sw",
+ "itch"
+ ],
+ [
+ "▁st",
+ "one"
+ ],
+ [
+ "▁sto",
+ "ne"
+ ],
+ [
+ "▁",
+ "stone"
+ ],
+ [
+ "▁trans",
+ "l"
+ ],
+ [
+ "▁",
+ "transl"
+ ],
+ [
+ "▁v",
+ "ac"
+ ],
+ [
+ "▁va",
+ "c"
+ ],
+ [
+ "an",
+ "don"
+ ],
+ [
+ "and",
+ "on"
+ ],
+ [
+ "ando",
+ "n"
+ ],
+ [
+ "▁F",
+ "ore"
+ ],
+ [
+ "▁For",
+ "e"
+ ],
+ [
+ "▁Fo",
+ "re"
+ ],
+ [
+ "▁",
+ "Fore"
+ ],
+ [
+ "▁s",
+ "ounds"
+ ],
+ [
+ "▁sound",
+ "s"
+ ],
+ [
+ "▁P",
+ "op"
+ ],
+ [
+ "▁Po",
+ "p"
+ ],
+ [
+ "▁",
+ "Pop"
+ ],
+ [
+ "▁H",
+ "T"
+ ],
+ [
+ "▁",
+ "HT"
+ ],
+ [
+ "li",
+ "a"
+ ],
+ [
+ "l",
+ "ia"
+ ],
+ [
+ "en",
+ "ter"
+ ],
+ [
+ "ent",
+ "er"
+ ],
+ [
+ "ente",
+ "r"
+ ],
+ [
+ "▁hel",
+ "ps"
+ ],
+ [
+ "▁help",
+ "s"
+ ],
+ [
+ "ed",
+ "y"
+ ],
+ [
+ "e",
+ "dy"
+ ],
+ [
+ "ст",
+ "вен"
+ ],
+ [
+ "ств",
+ "ен"
+ ],
+ [
+ "стве",
+ "н"
+ ],
+ [
+ "an",
+ "ted"
+ ],
+ [
+ "ant",
+ "ed"
+ ],
+ [
+ "ante",
+ "d"
+ ],
+ [
+ "▁I",
+ "ts"
+ ],
+ [
+ "▁It",
+ "s"
+ ],
+ [
+ "▁St",
+ "ep"
+ ],
+ [
+ "▁Ste",
+ "p"
+ ],
+ [
+ "▁",
+ "Step"
+ ],
+ [
+ "I",
+ "con"
+ ],
+ [
+ "▁EX",
+ "PECT"
+ ],
+ [
+ "▁",
+ "EXPECT"
+ ],
+ [
+ "ial",
+ "ized"
+ ],
+ [
+ "ialize",
+ "d"
+ ],
+ [
+ "Pos",
+ "t"
+ ],
+ [
+ "Po",
+ "st"
+ ],
+ [
+ "P",
+ "ost"
+ ],
+ [
+ "az",
+ "e"
+ ],
+ [
+ "a",
+ "ze"
+ ],
+ [
+ "▁Car",
+ "ol"
+ ],
+ [
+ "▁Ca",
+ "rol"
+ ],
+ [
+ "▁re",
+ "q"
+ ],
+ [
+ "▁r",
+ "eq"
+ ],
+ [
+ "▁",
+ "req"
+ ],
+ [
+ "▁crit",
+ "ical"
+ ],
+ [
+ "▁critic",
+ "al"
+ ],
+ [
+ "D",
+ "S"
+ ],
+ [
+ "▁se",
+ "at"
+ ],
+ [
+ "▁sea",
+ "t"
+ ],
+ [
+ "ap",
+ "ed"
+ ],
+ [
+ "ape",
+ "d"
+ ],
+ [
+ "a",
+ "ped"
+ ],
+ [
+ "▁up",
+ "per"
+ ],
+ [
+ "▁upp",
+ "er"
+ ],
+ [
+ "▁",
+ "upper"
+ ],
+ [
+ "▁S",
+ "y"
+ ],
+ [
+ "▁",
+ "Sy"
+ ],
+ [
+ "▁ex",
+ "plain"
+ ],
+ [
+ "▁expl",
+ "ain"
+ ],
+ [
+ "▁'",
+ "./"
+ ],
+ [
+ "▁'.",
+ "/"
+ ],
+ [
+ "ut",
+ "ils"
+ ],
+ [
+ "util",
+ "s"
+ ],
+ [
+ "uti",
+ "ls"
+ ],
+ [
+ "poss",
+ "ible"
+ ],
+ [
+ "▁d",
+ "ont"
+ ],
+ [
+ "▁do",
+ "nt"
+ ],
+ [
+ "▁don",
+ "t"
+ ],
+ [
+ "H",
+ "ost"
+ ],
+ [
+ "▁appro",
+ "xim"
+ ],
+ [
+ "▁approx",
+ "im"
+ ],
+ [
+ "As",
+ "ync"
+ ],
+ [
+ "A",
+ "sync"
+ ],
+ [
+ "▁g",
+ "rab"
+ ],
+ [
+ "▁gr",
+ "ab"
+ ],
+ [
+ "▁gra",
+ "b"
+ ],
+ [
+ "▁s",
+ "ources"
+ ],
+ [
+ "▁source",
+ "s"
+ ],
+ [
+ "▁sour",
+ "ces"
+ ],
+ [
+ "▁",
+ "sources"
+ ],
+ [
+ "▁M",
+ "os"
+ ],
+ [
+ "▁Mo",
+ "s"
+ ],
+ [
+ "▁Germ",
+ "any"
+ ],
+ [
+ "▁German",
+ "y"
+ ],
+ [
+ "▁Ger",
+ "many"
+ ],
+ [
+ "▁r",
+ "ub"
+ ],
+ [
+ "▁ru",
+ "b"
+ ],
+ [
+ "▁",
+ "rub"
+ ],
+ [
+ "CH",
+ "AN"
+ ],
+ [
+ "▁r",
+ "ain"
+ ],
+ [
+ "▁ra",
+ "in"
+ ],
+ [
+ "▁tr",
+ "uly"
+ ],
+ [
+ "▁join",
+ "ed"
+ ],
+ [
+ "▁jo",
+ "ined"
+ ],
+ [
+ "▁<",
+ "?"
+ ],
+ [
+ "▁",
+ ""
+ ],
+ [
+ "▁L",
+ "o"
+ ],
+ [
+ "▁",
+ "Lo"
+ ],
+ [
+ "Des",
+ "cription"
+ ],
+ [
+ "De",
+ "scription"
+ ],
+ [
+ "ak",
+ "t"
+ ],
+ [
+ "a",
+ "kt"
+ ],
+ [
+ "▁A",
+ "nn"
+ ],
+ [
+ "▁An",
+ "n"
+ ],
+ [
+ "▁",
+ "Ann"
+ ],
+ [
+ "^",
+ "*"
+ ],
+ [
+ "id",
+ "ae"
+ ],
+ [
+ "ida",
+ "e"
+ ],
+ [
+ "(",
+ ":"
+ ],
+ [
+ "t",
+ "w"
+ ],
+ [
+ "Ma",
+ "r"
+ ],
+ [
+ "M",
+ "ar"
+ ],
+ [
+ "pro",
+ "du"
+ ],
+ [
+ "prod",
+ "u"
+ ],
+ [
+ "p",
+ "rodu"
+ ],
+ [
+ "▁sp",
+ "oke"
+ ],
+ [
+ "▁spo",
+ "ke"
+ ],
+ [
+ "ю",
+ "т"
+ ],
+ [
+ "▁walk",
+ "ing"
+ ],
+ [
+ "▁wal",
+ "king"
+ ],
+ [
+ "▁nod",
+ "ded"
+ ],
+ [
+ "Pro",
+ "ps"
+ ],
+ [
+ "Pr",
+ "ops"
+ ],
+ [
+ "Prop",
+ "s"
+ ],
+ [
+ "En",
+ "abled"
+ ],
+ [
+ "Enable",
+ "d"
+ ],
+ [
+ "ir",
+ "k"
+ ],
+ [
+ "FI",
+ "LE"
+ ],
+ [
+ "FIL",
+ "E"
+ ],
+ [
+ "F",
+ "ILE"
+ ],
+ [
+ "equ",
+ "al"
+ ],
+ [
+ "eq",
+ "ual"
+ ],
+ [
+ "e",
+ "qual"
+ ],
+ [
+ "pp",
+ "ing"
+ ],
+ [
+ "p",
+ "ping"
+ ],
+ [
+ "ol",
+ "i"
+ ],
+ [
+ "o",
+ "li"
+ ],
+ [
+ "E",
+ "V"
+ ],
+ [
+ "en",
+ "z"
+ ],
+ [
+ "et",
+ "ing"
+ ],
+ [
+ "eti",
+ "ng"
+ ],
+ [
+ "e",
+ "ting"
+ ],
+ [
+ "▁s",
+ "ample"
+ ],
+ [
+ "▁sam",
+ "ple"
+ ],
+ [
+ "▁",
+ "sample"
+ ],
+ [
+ "▁art",
+ "ist"
+ ],
+ [
+ "[",
+ "$"
+ ],
+ [
+ "it",
+ "à"
+ ],
+ [
+ "й",
+ "о"
+ ],
+ [
+ "pro",
+ "ps"
+ ],
+ [
+ "pr",
+ "ops"
+ ],
+ [
+ "prop",
+ "s"
+ ],
+ [
+ "b",
+ "u"
+ ],
+ [
+ "е",
+ "в"
+ ],
+ [
+ "▁respons",
+ "ible"
+ ],
+ [
+ "M",
+ "T"
+ ],
+ [
+ "▁caus",
+ "ed"
+ ],
+ [
+ "▁cause",
+ "d"
+ ],
+ [
+ "▁ca",
+ "used"
+ ],
+ [
+ "▁the",
+ "me"
+ ],
+ [
+ "▁th",
+ "eme"
+ ],
+ [
+ "▁them",
+ "e"
+ ],
+ [
+ "▁",
+ "theme"
+ ],
+ [
+ "▁W",
+ "as"
+ ],
+ [
+ "▁Wa",
+ "s"
+ ],
+ [
+ "▁",
+ "Was"
+ ],
+ [
+ "▁B",
+ "efore"
+ ],
+ [
+ "▁Be",
+ "fore"
+ ],
+ [
+ "▁",
+ "Before"
+ ],
+ [
+ "ac",
+ "le"
+ ],
+ [
+ "acl",
+ "e"
+ ],
+ [
+ "a",
+ "cle"
+ ],
+ [
+ "▁ро",
+ "ку"
+ ],
+ [
+ "c",
+ "u"
+ ],
+ [
+ "DE",
+ "V"
+ ],
+ [
+ "D",
+ "EV"
+ ],
+ [
+ "▁h",
+ "ung"
+ ],
+ [
+ "▁hun",
+ "g"
+ ],
+ [
+ "▁",
+ "hung"
+ ],
+ [
+ "text",
+ "bf"
+ ],
+ [
+ "▁s",
+ "pin"
+ ],
+ [
+ "▁sp",
+ "in"
+ ],
+ [
+ "▁",
+ "spin"
+ ],
+ [
+ "▁la",
+ "test"
+ ],
+ [
+ "▁late",
+ "st"
+ ],
+ [
+ "▁lat",
+ "est"
+ ],
+ [
+ "▁",
+ "latest"
+ ],
+ [
+ "ent",
+ "ially"
+ ],
+ [
+ "ential",
+ "ly"
+ ],
+ [
+ "enti",
+ "ally"
+ ],
+ [
+ "▁Pro",
+ "gram"
+ ],
+ [
+ "▁Pr",
+ "ogram"
+ ],
+ [
+ "▁",
+ "Program"
+ ],
+ [
+ "Met",
+ "adata"
+ ],
+ [
+ "Meta",
+ "data"
+ ],
+ [
+ "pass",
+ "word"
+ ],
+ [
+ "▁h",
+ "urt"
+ ],
+ [
+ "▁hur",
+ "t"
+ ],
+ [
+ "к",
+ "с"
+ ],
+ [
+ "▁A",
+ "us"
+ ],
+ [
+ "▁Au",
+ "s"
+ ],
+ [
+ "se",
+ "y"
+ ],
+ [
+ "s",
+ "ey"
+ ],
+ [
+ "al",
+ "let"
+ ],
+ [
+ "all",
+ "et"
+ ],
+ [
+ "alle",
+ "t"
+ ],
+ [
+ "x",
+ "F"
+ ],
+ [
+ "▁R",
+ "oad"
+ ],
+ [
+ "▁Ro",
+ "ad"
+ ],
+ [
+ "ет",
+ "ся"
+ ],
+ [
+ "е",
+ "тся"
+ ],
+ [
+ "▁re",
+ "nt"
+ ],
+ [
+ "▁r",
+ "ent"
+ ],
+ [
+ "▁ren",
+ "t"
+ ],
+ [
+ "▁",
+ "rent"
+ ],
+ [
+ "ци",
+ "я"
+ ],
+ [
+ "▁As",
+ "sert"
+ ],
+ [
+ "▁Ass",
+ "ert"
+ ],
+ [
+ "▁",
+ "Assert"
+ ],
+ [
+ "і",
+ "ль"
+ ],
+ [
+ "ü",
+ "ck"
+ ],
+ [
+ "▁s",
+ "ites"
+ ],
+ [
+ "▁sit",
+ "es"
+ ],
+ [
+ "▁si",
+ "tes"
+ ],
+ [
+ "▁site",
+ "s"
+ ],
+ [
+ "Doc",
+ "ument"
+ ],
+ [
+ "D",
+ "ocument"
+ ],
+ [
+ "▁obt",
+ "ained"
+ ],
+ [
+ "▁obtain",
+ "ed"
+ ],
+ [
+ "▁c",
+ "i"
+ ],
+ [
+ "▁",
+ "ci"
+ ],
+ [
+ "▁[",
+ "\""
+ ],
+ [
+ "▁",
+ "[\""
+ ],
+ [
+ "▁com",
+ "pleted"
+ ],
+ [
+ "▁comp",
+ "leted"
+ ],
+ [
+ "▁complet",
+ "ed"
+ ],
+ [
+ "▁compl",
+ "eted"
+ ],
+ [
+ "▁complete",
+ "d"
+ ],
+ [
+ "as",
+ "et"
+ ],
+ [
+ "ase",
+ "t"
+ ],
+ [
+ "a",
+ "set"
+ ],
+ [
+ "ra",
+ "id"
+ ],
+ [
+ "rai",
+ "d"
+ ],
+ [
+ "r",
+ "aid"
+ ],
+ [
+ "▁s",
+ "orry"
+ ],
+ [
+ "▁sor",
+ "ry"
+ ],
+ [
+ "▁f",
+ "ab"
+ ],
+ [
+ "▁fa",
+ "b"
+ ],
+ [
+ "▁",
+ "fab"
+ ],
+ [
+ "▁sch",
+ "ools"
+ ],
+ [
+ "▁school",
+ "s"
+ ],
+ [
+ "хо",
+ "ди"
+ ],
+ [
+ "ход",
+ "и"
+ ],
+ [
+ "▁s",
+ "cr"
+ ],
+ [
+ "▁sc",
+ "r"
+ ],
+ [
+ "▁",
+ "scr"
+ ],
+ [
+ "▁in",
+ "cor"
+ ],
+ [
+ "▁inc",
+ "or"
+ ],
+ [
+ "▁'",
+ "/"
+ ],
+ [
+ "▁s",
+ "pr"
+ ],
+ [
+ "▁sp",
+ "r"
+ ],
+ [
+ "▁",
+ "spr"
+ ],
+ [
+ "▁T",
+ "ext"
+ ],
+ [
+ "▁Te",
+ "xt"
+ ],
+ [
+ "▁Tex",
+ "t"
+ ],
+ [
+ "▁",
+ "Text"
+ ],
+ [
+ "▁com",
+ "mercial"
+ ],
+ [
+ "▁commer",
+ "cial"
+ ],
+ [
+ "in",
+ "gly"
+ ],
+ [
+ "ing",
+ "ly"
+ ],
+ [
+ "▁opin",
+ "ion"
+ ],
+ [
+ "▁S",
+ "tar"
+ ],
+ [
+ "▁St",
+ "ar"
+ ],
+ [
+ "▁Sta",
+ "r"
+ ],
+ [
+ "▁",
+ "Star"
+ ],
+ [
+ "Si",
+ "gn"
+ ],
+ [
+ "Sig",
+ "n"
+ ],
+ [
+ "S",
+ "ign"
+ ],
+ [
+ "▁j",
+ "avax"
+ ],
+ [
+ "▁java",
+ "x"
+ ],
+ [
+ "▁",
+ "javax"
+ ],
+ [
+ "w",
+ "i"
+ ],
+ [
+ "la",
+ "t"
+ ],
+ [
+ "l",
+ "at"
+ ],
+ [
+ "▁K",
+ "ey"
+ ],
+ [
+ "▁Ke",
+ "y"
+ ],
+ [
+ "▁",
+ "Key"
+ ],
+ [
+ "var",
+ "phi"
+ ],
+ [
+ "д",
+ "ы"
+ ],
+ [
+ "▁conne",
+ "cted"
+ ],
+ [
+ "▁connect",
+ "ed"
+ ],
+ [
+ "▁",
+ "connected"
+ ],
+ [
+ "▁ad",
+ "just"
+ ],
+ [
+ "▁adj",
+ "ust"
+ ],
+ [
+ "▁",
+ "adjust"
+ ],
+ [
+ "▁A",
+ "z"
+ ],
+ [
+ "▁",
+ "Az"
+ ],
+ [
+ "▁pl",
+ "anning"
+ ],
+ [
+ "▁plan",
+ "ning"
+ ],
+ [
+ "--",
+ "-"
+ ],
+ [
+ "-",
+ "--"
+ ],
+ [
+ "In",
+ "teger"
+ ],
+ [
+ "au",
+ "f"
+ ],
+ [
+ "a",
+ "uf"
+ ],
+ [
+ "ex",
+ "pected"
+ ],
+ [
+ "expect",
+ "ed"
+ ],
+ [
+ "e",
+ "xpected"
+ ],
+ [
+ "▁f",
+ "ant"
+ ],
+ [
+ "▁fa",
+ "nt"
+ ],
+ [
+ "▁fan",
+ "t"
+ ],
+ [
+ "▁t",
+ "ou"
+ ],
+ [
+ "▁to",
+ "u"
+ ],
+ [
+ "Par",
+ "ent"
+ ],
+ [
+ "P",
+ "arent"
+ ],
+ [
+ "▁L",
+ "at"
+ ],
+ [
+ "▁La",
+ "t"
+ ],
+ [
+ "▁",
+ "Lat"
+ ],
+ [
+ "▁thought",
+ "s"
+ ],
+ [
+ "▁though",
+ "ts"
+ ],
+ [
+ "▁J",
+ "ud"
+ ],
+ [
+ "▁Ju",
+ "d"
+ ],
+ [
+ "Param",
+ "eters"
+ ],
+ [
+ "Parameter",
+ "s"
+ ],
+ [
+ "G",
+ "r"
+ ],
+ [
+ "ро",
+ "м"
+ ],
+ [
+ "I",
+ "A"
+ ],
+ [
+ "▁B",
+ "ob"
+ ],
+ [
+ "▁Bo",
+ "b"
+ ],
+ [
+ "lic",
+ "t"
+ ],
+ [
+ "li",
+ "ct"
+ ],
+ [
+ "l",
+ "ict"
+ ],
+ [
+ "la",
+ "n"
+ ],
+ [
+ "l",
+ "an"
+ ],
+ [
+ "om",
+ "ic"
+ ],
+ [
+ "omi",
+ "c"
+ ],
+ [
+ "o",
+ "mic"
+ ],
+ [
+ "▁a",
+ "part"
+ ],
+ [
+ "▁ap",
+ "art"
+ ],
+ [
+ "▁t",
+ "rou"
+ ],
+ [
+ "▁tr",
+ "ou"
+ ],
+ [
+ "▁tro",
+ "u"
+ ],
+ [
+ "▁app",
+ "reci"
+ ],
+ [
+ "▁Christ",
+ "mas"
+ ],
+ [
+ "ir",
+ "q"
+ ],
+ [
+ "i",
+ "rq"
+ ],
+ [
+ "th",
+ "on"
+ ],
+ [
+ "t",
+ "hon"
+ ],
+ [
+ "▁Er",
+ "ror"
+ ],
+ [
+ "▁Err",
+ "or"
+ ],
+ [
+ "▁",
+ "Error"
+ ],
+ [
+ "▁s",
+ "core"
+ ],
+ [
+ "▁sc",
+ "ore"
+ ],
+ [
+ "▁",
+ "score"
+ ],
+ [
+ "ro",
+ "me"
+ ],
+ [
+ "rom",
+ "e"
+ ],
+ [
+ "r",
+ "ome"
+ ],
+ [
+ "▁ne",
+ "ighbor"
+ ],
+ [
+ "▁neigh",
+ "bor"
+ ],
+ [
+ "▁neighb",
+ "or"
+ ],
+ [
+ "▁M",
+ "ur"
+ ],
+ [
+ "▁Mu",
+ "r"
+ ],
+ [
+ "ad",
+ "min"
+ ],
+ [
+ "▁Fil",
+ "m"
+ ],
+ [
+ "▁Fi",
+ "lm"
+ ],
+ [
+ "Re",
+ "ct"
+ ],
+ [
+ "Rec",
+ "t"
+ ],
+ [
+ "R",
+ "ect"
+ ],
+ [
+ "▁config",
+ "uration"
+ ],
+ [
+ "▁",
+ "configuration"
+ ],
+ [
+ "▁c",
+ "s"
+ ],
+ [
+ "▁",
+ "cs"
+ ],
+ [
+ "gu",
+ "n"
+ ],
+ [
+ "g",
+ "un"
+ ],
+ [
+ "ch",
+ "annel"
+ ],
+ [
+ "chan",
+ "nel"
+ ],
+ [
+ "▁Re",
+ "port"
+ ],
+ [
+ "▁Rep",
+ "ort"
+ ],
+ [
+ "▁",
+ "Report"
+ ],
+ [
+ "▁str",
+ "ateg"
+ ],
+ [
+ "▁strat",
+ "eg"
+ ],
+ [
+ "▁work",
+ "ers"
+ ],
+ [
+ "▁wor",
+ "kers"
+ ],
+ [
+ "▁worker",
+ "s"
+ ],
+ [
+ "▁",
+ "workers"
+ ],
+ [
+ "field",
+ "s"
+ ],
+ [
+ "Sch",
+ "ema"
+ ],
+ [
+ "Sche",
+ "ma"
+ ],
+ [
+ "S",
+ "chema"
+ ],
+ [
+ "ap",
+ "pa"
+ ],
+ [
+ "app",
+ "a"
+ ],
+ [
+ "ol",
+ "ic"
+ ],
+ [
+ "oli",
+ "c"
+ ],
+ [
+ "o",
+ "lic"
+ ],
+ [
+ "E",
+ "O"
+ ],
+ [
+ "▁Ch",
+ "arl"
+ ],
+ [
+ "▁Char",
+ "l"
+ ],
+ [
+ "▁Cha",
+ "rl"
+ ],
+ [
+ "▁C",
+ "up"
+ ],
+ [
+ "▁Cu",
+ "p"
+ ],
+ [
+ "pn",
+ "g"
+ ],
+ [
+ "p",
+ "ng"
+ ],
+ [
+ "▁H",
+ "ill"
+ ],
+ [
+ "▁Hi",
+ "ll"
+ ],
+ [
+ "▁Hil",
+ "l"
+ ],
+ [
+ "ow",
+ "e"
+ ],
+ [
+ "o",
+ "we"
+ ],
+ [
+ "▁most",
+ "ly"
+ ],
+ [
+ "”",
+ "."
+ ],
+ [
+ "▁fin",
+ "ish"
+ ],
+ [
+ "▁",
+ "finish"
+ ],
+ [
+ "▁С",
+ "о"
+ ],
+ [
+ "▁st",
+ "ars"
+ ],
+ [
+ "▁star",
+ "s"
+ ],
+ [
+ "▁sta",
+ "rs"
+ ],
+ [
+ "pl",
+ "ayer"
+ ],
+ [
+ "play",
+ "er"
+ ],
+ [
+ "p",
+ "layer"
+ ],
+ [
+ "▁in",
+ "ner"
+ ],
+ [
+ "▁inn",
+ "er"
+ ],
+ [
+ "▁",
+ "inner"
+ ],
+ [
+ "com",
+ "ponent"
+ ],
+ [
+ "ti",
+ "m"
+ ],
+ [
+ "t",
+ "im"
+ ],
+ [
+ "I",
+ "E"
+ ],
+ [
+ "▁t",
+ "her"
+ ],
+ [
+ "▁the",
+ "r"
+ ],
+ [
+ "▁th",
+ "er"
+ ],
+ [
+ "▁",
+ "ther"
+ ],
+ [
+ "▁s",
+ "mart"
+ ],
+ [
+ "▁sm",
+ "art"
+ ],
+ [
+ "▁",
+ "smart"
+ ],
+ [
+ "▁s",
+ "ad"
+ ],
+ [
+ "▁sa",
+ "d"
+ ],
+ [
+ "▁Coun",
+ "cil"
+ ],
+ [
+ "ar",
+ "ea"
+ ],
+ [
+ "are",
+ "a"
+ ],
+ [
+ "a",
+ "rea"
+ ],
+ [
+ "la",
+ "y"
+ ],
+ [
+ "l",
+ "ay"
+ ],
+ [
+ "▁б",
+ "а"
+ ],
+ [
+ "▁",
+ "ба"
+ ],
+ [
+ "▁gr",
+ "adu"
+ ],
+ [
+ "▁grad",
+ "u"
+ ],
+ [
+ "▁gra",
+ "du"
+ ],
+ [
+ "▁c",
+ "hem"
+ ],
+ [
+ "▁ch",
+ "em"
+ ],
+ [
+ "▁che",
+ "m"
+ ],
+ [
+ "▁",
+ "chem"
+ ],
+ [
+ "▁h",
+ "o"
+ ],
+ [
+ "▁",
+ "ho"
+ ],
+ [
+ "Se",
+ "lect"
+ ],
+ [
+ "S",
+ "elect"
+ ],
+ [
+ "▁in",
+ "str"
+ ],
+ [
+ "▁inst",
+ "r"
+ ],
+ [
+ "▁ins",
+ "tr"
+ ],
+ [
+ "▁",
+ "instr"
+ ],
+ [
+ "▁k",
+ "l"
+ ],
+ [
+ "▁",
+ "kl"
+ ],
+ [
+ "if",
+ "ications"
+ ],
+ [
+ "ific",
+ "ations"
+ ],
+ [
+ "ification",
+ "s"
+ ],
+ [
+ "Lo",
+ "ng"
+ ],
+ [
+ "L",
+ "ong"
+ ],
+ [
+ "▁s",
+ "obre"
+ ],
+ [
+ "▁so",
+ "bre"
+ ],
+ [
+ "▁sob",
+ "re"
+ ],
+ [
+ "▁O",
+ "ld"
+ ],
+ [
+ "▁Ol",
+ "d"
+ ],
+ [
+ "▁",
+ "Old"
+ ],
+ [
+ "we",
+ "st"
+ ],
+ [
+ "w",
+ "est"
+ ],
+ [
+ "},",
+ "\\"
+ ],
+ [
+ "}",
+ ",\\"
+ ],
+ [
+ "in",
+ "gu"
+ ],
+ [
+ "ing",
+ "u"
+ ],
+ [
+ "▁sp",
+ "ring"
+ ],
+ [
+ "▁spr",
+ "ing"
+ ],
+ [
+ "▁",
+ "spring"
+ ],
+ [
+ "▁n",
+ "ur"
+ ],
+ [
+ "▁nu",
+ "r"
+ ],
+ [
+ "ex",
+ "ample"
+ ],
+ [
+ "Wh",
+ "en"
+ ],
+ [
+ "Whe",
+ "n"
+ ],
+ [
+ "W",
+ "hen"
+ ],
+ [
+ "▁adv",
+ "ice"
+ ],
+ [
+ "▁u",
+ "lt"
+ ],
+ [
+ "▁ul",
+ "t"
+ ],
+ [
+ "▁",
+ "ult"
+ ],
+ [
+ "en",
+ "nis"
+ ],
+ [
+ "enn",
+ "is"
+ ],
+ [
+ "▁L",
+ "ove"
+ ],
+ [
+ "▁Lo",
+ "ve"
+ ],
+ [
+ "▁Lov",
+ "e"
+ ],
+ [
+ "▁",
+ "Love"
+ ],
+ [
+ "▁\"",
+ "\""
+ ],
+ [
+ "▁",
+ "\"\""
+ ],
+ [
+ "▁incre",
+ "ased"
+ ],
+ [
+ "▁increase",
+ "d"
+ ],
+ [
+ "▁f",
+ "inding"
+ ],
+ [
+ "▁fin",
+ "ding"
+ ],
+ [
+ "▁find",
+ "ing"
+ ],
+ [
+ "ir",
+ "ty"
+ ],
+ [
+ "irt",
+ "y"
+ ],
+ [
+ "ist",
+ "rict"
+ ],
+ [
+ "istr",
+ "ict"
+ ],
+ [
+ "i",
+ "strict"
+ ],
+ [
+ "▁l",
+ "ayer"
+ ],
+ [
+ "▁la",
+ "yer"
+ ],
+ [
+ "▁lay",
+ "er"
+ ],
+ [
+ "▁",
+ "layer"
+ ],
+ [
+ "temp",
+ "late"
+ ],
+ [
+ "t",
+ "emplate"
+ ],
+ [
+ "F",
+ "irst"
+ ],
+ [
+ "ны",
+ "м"
+ ],
+ [
+ "igr",
+ "ation"
+ ],
+ [
+ "ren",
+ "cy"
+ ],
+ [
+ "r",
+ "ency"
+ ],
+ [
+ "ow",
+ "ie"
+ ],
+ [
+ "owi",
+ "e"
+ ],
+ [
+ "o",
+ "wie"
+ ],
+ [
+ "▁n",
+ "p"
+ ],
+ [
+ "▁",
+ "np"
+ ],
+ [
+ "▁s",
+ "election"
+ ],
+ [
+ "▁se",
+ "lection"
+ ],
+ [
+ "▁select",
+ "ion"
+ ],
+ [
+ "▁sel",
+ "ection"
+ ],
+ [
+ "▁sele",
+ "ction"
+ ],
+ [
+ "▁",
+ "selection"
+ ],
+ [
+ "▁N",
+ "ach"
+ ],
+ [
+ "▁Na",
+ "ch"
+ ],
+ [
+ "▁P",
+ "RO"
+ ],
+ [
+ "▁PR",
+ "O"
+ ],
+ [
+ "▁",
+ "PRO"
+ ],
+ [
+ "▁p",
+ "olic"
+ ],
+ [
+ "▁pol",
+ "ic"
+ ],
+ [
+ "▁po",
+ "lic"
+ ],
+ [
+ "▁data",
+ "base"
+ ],
+ [
+ "▁dat",
+ "abase"
+ ],
+ [
+ "▁",
+ "database"
+ ],
+ [
+ "▁by",
+ "te"
+ ],
+ [
+ "▁",
+ "byte"
+ ],
+ [
+ "▁prov",
+ "iding"
+ ],
+ [
+ "ma",
+ "c"
+ ],
+ [
+ "m",
+ "ac"
+ ],
+ [
+ "▁me",
+ "tal"
+ ],
+ [
+ "▁met",
+ "al"
+ ],
+ [
+ "▁meta",
+ "l"
+ ],
+ [
+ "mod",
+ "ules"
+ ],
+ [
+ "module",
+ "s"
+ ],
+ [
+ "▁Ge",
+ "org"
+ ],
+ [
+ "▁S",
+ "a"
+ ],
+ [
+ "▁",
+ "Sa"
+ ],
+ [
+ "▁est",
+ "ablish"
+ ],
+ [
+ "▁estab",
+ "lish"
+ ],
+ [
+ "..",
+ ".\""
+ ],
+ [
+ "...",
+ "\""
+ ],
+ [
+ "i",
+ "u"
+ ],
+ [
+ "ki",
+ "n"
+ ],
+ [
+ "k",
+ "in"
+ ],
+ [
+ "▁e",
+ "th"
+ ],
+ [
+ "▁et",
+ "h"
+ ],
+ [
+ "▁",
+ "eth"
+ ],
+ [
+ "▁S",
+ "and"
+ ],
+ [
+ "▁San",
+ "d"
+ ],
+ [
+ "▁Sa",
+ "nd"
+ ],
+ [
+ "▁Ch",
+ "apter"
+ ],
+ [
+ "▁Chap",
+ "ter"
+ ],
+ [
+ "▁g",
+ "al"
+ ],
+ [
+ "▁ga",
+ "l"
+ ],
+ [
+ "▁",
+ "gal"
+ ],
+ [
+ "▁i",
+ "ce"
+ ],
+ [
+ "▁ic",
+ "e"
+ ],
+ [
+ "▁",
+ "ice"
+ ],
+ [
+ "Re",
+ "d"
+ ],
+ [
+ "R",
+ "ed"
+ ],
+ [
+ "▁d",
+ "al"
+ ],
+ [
+ "▁da",
+ "l"
+ ],
+ [
+ "▁",
+ "dal"
+ ],
+ [
+ "▁pr",
+ "incipal"
+ ],
+ [
+ "▁princip",
+ "al"
+ ],
+ [
+ "Ms",
+ "g"
+ ],
+ [
+ "M",
+ "sg"
+ ],
+ [
+ "▁rem",
+ "ains"
+ ],
+ [
+ "▁remain",
+ "s"
+ ],
+ [
+ "н",
+ "г"
+ ],
+ [
+ "T",
+ "itle"
+ ],
+ [
+ "Re",
+ "l"
+ ],
+ [
+ "R",
+ "el"
+ ],
+ [
+ "Dis",
+ "play"
+ ],
+ [
+ "No",
+ "n"
+ ],
+ [
+ "N",
+ "on"
+ ],
+ [
+ "▁def",
+ "inition"
+ ],
+ [
+ "▁definit",
+ "ion"
+ ],
+ [
+ "▁defin",
+ "ition"
+ ],
+ [
+ "▁",
+ "definition"
+ ],
+ [
+ "▁at",
+ "tr"
+ ],
+ [
+ "▁att",
+ "r"
+ ],
+ [
+ "▁",
+ "attr"
+ ],
+ [
+ "▁sign",
+ "al"
+ ],
+ [
+ "▁sig",
+ "nal"
+ ],
+ [
+ "▁",
+ "signal"
+ ],
+ [
+ "h",
+ "l"
+ ],
+ [
+ "▁s",
+ "el"
+ ],
+ [
+ "▁se",
+ "l"
+ ],
+ [
+ "▁",
+ "sel"
+ ],
+ [
+ "▁vol",
+ "ume"
+ ],
+ [
+ "▁",
+ "volume"
+ ],
+ [
+ "▁c",
+ "ache"
+ ],
+ [
+ "▁ca",
+ "che"
+ ],
+ [
+ "▁",
+ "cache"
+ ],
+ [
+ "he",
+ "ns"
+ ],
+ [
+ "hen",
+ "s"
+ ],
+ [
+ "h",
+ "ens"
+ ],
+ [
+ "▁w",
+ "ird"
+ ],
+ [
+ "▁wir",
+ "d"
+ ],
+ [
+ "[",
+ "\\"
+ ],
+ [
+ "NO",
+ "T"
+ ],
+ [
+ "N",
+ "OT"
+ ],
+ [
+ "▁e",
+ "lection"
+ ],
+ [
+ "▁el",
+ "ection"
+ ],
+ [
+ "▁elect",
+ "ion"
+ ],
+ [
+ "▁ele",
+ "ction"
+ ],
+ [
+ "▁",
+ "election"
+ ],
+ [
+ "ut",
+ "t"
+ ],
+ [
+ "u",
+ "tt"
+ ],
+ [
+ "▁W",
+ "indow"
+ ],
+ [
+ "▁Wind",
+ "ow"
+ ],
+ [
+ "▁",
+ "Window"
+ ],
+ [
+ "en",
+ "tal"
+ ],
+ [
+ "ent",
+ "al"
+ ],
+ [
+ "enta",
+ "l"
+ ],
+ [
+ "if",
+ "est"
+ ],
+ [
+ "ife",
+ "st"
+ ],
+ [
+ "x",
+ "f"
+ ],
+ [
+ "▁Р",
+ "а"
+ ],
+ [
+ "▁over",
+ "all"
+ ],
+ [
+ "bl",
+ "ic"
+ ],
+ [
+ "b",
+ "lic"
+ ],
+ [
+ "▁ed",
+ "itor"
+ ],
+ [
+ "▁edit",
+ "or"
+ ],
+ [
+ "▁",
+ "editor"
+ ],
+ [
+ "ad",
+ "en"
+ ],
+ [
+ "ade",
+ "n"
+ ],
+ [
+ "a",
+ "den"
+ ],
+ [
+ "▁c",
+ "art"
+ ],
+ [
+ "▁car",
+ "t"
+ ],
+ [
+ "▁ca",
+ "rt"
+ ],
+ [
+ "▁",
+ "cart"
+ ],
+ [
+ "Le",
+ "ft"
+ ],
+ [
+ "L",
+ "eft"
+ ],
+ [
+ "ul",
+ "s"
+ ],
+ [
+ "u",
+ "ls"
+ ],
+ [
+ "bin",
+ "g"
+ ],
+ [
+ "bi",
+ "ng"
+ ],
+ [
+ "b",
+ "ing"
+ ],
+ [
+ "R",
+ "ight"
+ ],
+ [
+ "▁s",
+ "é"
+ ],
+ [
+ "Si",
+ "m"
+ ],
+ [
+ "S",
+ "im"
+ ],
+ [
+ "▁came",
+ "ra"
+ ],
+ [
+ "▁cam",
+ "era"
+ ],
+ [
+ "▁",
+ "camera"
+ ],
+ [
+ "▁f",
+ "av"
+ ],
+ [
+ "▁fa",
+ "v"
+ ],
+ [
+ "De",
+ "cl"
+ ],
+ [
+ "Dec",
+ "l"
+ ],
+ [
+ "sp",
+ "ring"
+ ],
+ [
+ "spr",
+ "ing"
+ ],
+ [
+ "▁err",
+ "ors"
+ ],
+ [
+ "▁er",
+ "rors"
+ ],
+ [
+ "▁error",
+ "s"
+ ],
+ [
+ "▁",
+ "errors"
+ ],
+ [
+ "T",
+ "ab"
+ ],
+ [
+ "print",
+ "ln"
+ ],
+ [
+ "▁B",
+ "ern"
+ ],
+ [
+ "▁Be",
+ "rn"
+ ],
+ [
+ "▁Ber",
+ "n"
+ ],
+ [
+ "na",
+ "b"
+ ],
+ [
+ "n",
+ "ab"
+ ],
+ [
+ "▁B",
+ "ase"
+ ],
+ [
+ "▁Bas",
+ "e"
+ ],
+ [
+ "▁Ba",
+ "se"
+ ],
+ [
+ "▁",
+ "Base"
+ ],
+ [
+ "▁a",
+ "uth"
+ ],
+ [
+ "▁aut",
+ "h"
+ ],
+ [
+ "▁au",
+ "th"
+ ],
+ [
+ "▁",
+ "auth"
+ ],
+ [
+ "▁app",
+ "arent"
+ ],
+ [
+ "▁ap",
+ "parent"
+ ],
+ [
+ "▁appar",
+ "ent"
+ ],
+ [
+ "▁pres",
+ "ented"
+ ],
+ [
+ "▁present",
+ "ed"
+ ],
+ [
+ "▁rem",
+ "ained"
+ ],
+ [
+ "▁remain",
+ "ed"
+ ],
+ [
+ "▁w",
+ "et"
+ ],
+ [
+ "▁we",
+ "t"
+ ],
+ [
+ "En",
+ "c"
+ ],
+ [
+ "E",
+ "nc"
+ ],
+ [
+ "IN",
+ "FO"
+ ],
+ [
+ "▁S",
+ "ing"
+ ],
+ [
+ "▁Si",
+ "ng"
+ ],
+ [
+ "▁Sin",
+ "g"
+ ],
+ [
+ "▁",
+ "Sing"
+ ],
+ [
+ "pack",
+ "age"
+ ],
+ [
+ "))",
+ ");"
+ ],
+ [
+ ")))",
+ ";"
+ ],
+ [
+ ")",
+ "));"
+ ],
+ [
+ "▁S",
+ "ocial"
+ ],
+ [
+ "▁So",
+ "cial"
+ ],
+ [
+ "▁Soc",
+ "ial"
+ ],
+ [
+ "▁Soci",
+ "al"
+ ],
+ [
+ "▁M",
+ "ass"
+ ],
+ [
+ "▁Ma",
+ "ss"
+ ],
+ [
+ "▁Mas",
+ "s"
+ ],
+ [
+ "▁",
+ "Mass"
+ ],
+ [
+ "▁des",
+ "pite"
+ ],
+ [
+ "▁desp",
+ "ite"
+ ],
+ [
+ "▁m",
+ "obile"
+ ],
+ [
+ "▁mob",
+ "ile"
+ ],
+ [
+ "▁mobil",
+ "e"
+ ],
+ [
+ "▁",
+ "mobile"
+ ],
+ [
+ "▁l",
+ "abor"
+ ],
+ [
+ "▁la",
+ "bor"
+ ],
+ [
+ "▁lab",
+ "or"
+ ],
+ [
+ "G",
+ "o"
+ ],
+ [
+ "▁e",
+ "sp"
+ ],
+ [
+ "▁es",
+ "p"
+ ],
+ [
+ "▁",
+ "esp"
+ ],
+ [
+ "▁T",
+ "able"
+ ],
+ [
+ "▁Ta",
+ "ble"
+ ],
+ [
+ "▁Tab",
+ "le"
+ ],
+ [
+ "▁",
+ "Table"
+ ],
+ [
+ "▁ex",
+ "pert"
+ ],
+ [
+ "▁exper",
+ "t"
+ ],
+ [
+ "▁exp",
+ "ert"
+ ],
+ [
+ "▁f",
+ "lex"
+ ],
+ [
+ "▁fl",
+ "ex"
+ ],
+ [
+ "▁fle",
+ "x"
+ ],
+ [
+ "▁",
+ "flex"
+ ],
+ [
+ "▁prof",
+ "ession"
+ ],
+ [
+ "▁profess",
+ "ion"
+ ],
+ [
+ "▁p",
+ "il"
+ ],
+ [
+ "▁pi",
+ "l"
+ ],
+ [
+ "Col",
+ "lection"
+ ],
+ [
+ "Coll",
+ "ection"
+ ],
+ [
+ "Collect",
+ "ion"
+ ],
+ [
+ "LO",
+ "CK"
+ ],
+ [
+ "LOC",
+ "K"
+ ],
+ [
+ "▁ap",
+ "plied"
+ ],
+ [
+ "▁appl",
+ "ied"
+ ],
+ [
+ "al",
+ "ler"
+ ],
+ [
+ "all",
+ "er"
+ ],
+ [
+ "alle",
+ "r"
+ ],
+ [
+ "or",
+ "ph"
+ ],
+ [
+ "orp",
+ "h"
+ ],
+ [
+ "EN",
+ "SE"
+ ],
+ [
+ "ENS",
+ "E"
+ ],
+ [
+ "▁бы",
+ "л"
+ ],
+ [
+ "▁d",
+ "b"
+ ],
+ [
+ "▁",
+ "db"
+ ],
+ [
+ "over",
+ "line"
+ ],
+ [
+ "▁C",
+ "ode"
+ ],
+ [
+ "▁Co",
+ "de"
+ ],
+ [
+ "▁",
+ "Code"
+ ],
+ [
+ "▁by",
+ "tes"
+ ],
+ [
+ "▁byte",
+ "s"
+ ],
+ [
+ "▁",
+ "bytes"
+ ],
+ [
+ "▁tr",
+ "ouble"
+ ],
+ [
+ "▁trou",
+ "ble"
+ ],
+ [
+ "▁на",
+ "се"
+ ],
+ [
+ "D",
+ "D"
+ ],
+ [
+ "▁Y",
+ "ear"
+ ],
+ [
+ "▁Ye",
+ "ar"
+ ],
+ [
+ "▁",
+ "Year"
+ ],
+ [
+ "mb",
+ "ox"
+ ],
+ [
+ "m",
+ "box"
+ ],
+ [
+ "▁ke",
+ "eping"
+ ],
+ [
+ "▁keep",
+ "ing"
+ ],
+ [
+ "▁",
+ "keeping"
+ ],
+ [
+ "▁k",
+ "ick"
+ ],
+ [
+ "▁ki",
+ "ck"
+ ],
+ [
+ "än",
+ "g"
+ ],
+ [
+ "ä",
+ "ng"
+ ],
+ [
+ "▁correspon",
+ "ding"
+ ],
+ [
+ "▁correspond",
+ "ing"
+ ],
+ [
+ "▁l",
+ "ibrary"
+ ],
+ [
+ "▁",
+ "library"
+ ],
+ [
+ "▁*/",
+ "\r"
+ ],
+ [
+ "▁",
+ "*/\r"
+ ],
+ [
+ "call",
+ "back"
+ ],
+ [
+ "um",
+ "s"
+ ],
+ [
+ "u",
+ "ms"
+ ],
+ [
+ "▁j",
+ "son"
+ ],
+ [
+ "▁js",
+ "on"
+ ],
+ [
+ "▁",
+ "json"
+ ],
+ [
+ "▁M",
+ "ount"
+ ],
+ [
+ "▁Mo",
+ "unt"
+ ],
+ [
+ "▁",
+ "Mount"
+ ],
+ [
+ "▁St",
+ "and"
+ ],
+ [
+ "▁Stan",
+ "d"
+ ],
+ [
+ "▁Sta",
+ "nd"
+ ],
+ [
+ "▁",
+ "Stand"
+ ],
+ [
+ "IG",
+ "HT"
+ ],
+ [
+ "IGH",
+ "T"
+ ],
+ [
+ "▁New",
+ "s"
+ ],
+ [
+ "▁Ne",
+ "ws"
+ ],
+ [
+ "▁",
+ "News"
+ ],
+ [
+ "▁com",
+ "ments"
+ ],
+ [
+ "▁comm",
+ "ents"
+ ],
+ [
+ "▁comment",
+ "s"
+ ],
+ [
+ "▁",
+ "comments"
+ ],
+ [
+ "return",
+ "s"
+ ],
+ [
+ "C",
+ "al"
+ ],
+ [
+ "▁a",
+ "ward"
+ ],
+ [
+ "▁aw",
+ "ard"
+ ],
+ [
+ "▁b",
+ "ought"
+ ],
+ [
+ "▁bou",
+ "ght"
+ ],
+ [
+ "include",
+ "graphics"
+ ],
+ [
+ "▁",
+ "ле"
+ ],
+ [
+ "do",
+ "t"
+ ],
+ [
+ "d",
+ "ot"
+ ],
+ [
+ "ro",
+ "nic"
+ ],
+ [
+ "ron",
+ "ic"
+ ],
+ [
+ "r",
+ "onic"
+ ],
+ [
+ "▁extrem",
+ "ely"
+ ],
+ [
+ "▁extreme",
+ "ly"
+ ],
+ [
+ "▁min",
+ "or"
+ ],
+ [
+ "▁mi",
+ "nor"
+ ],
+ [
+ "if",
+ "er"
+ ],
+ [
+ "ife",
+ "r"
+ ],
+ [
+ "i",
+ "fer"
+ ],
+ [
+ "ja",
+ "va"
+ ],
+ [
+ "jav",
+ "a"
+ ],
+ [
+ "j",
+ "ava"
+ ],
+ [
+ "en",
+ "dar"
+ ],
+ [
+ "end",
+ "ar"
+ ],
+ [
+ "enda",
+ "r"
+ ],
+ [
+ "la",
+ "yout"
+ ],
+ [
+ "lay",
+ "out"
+ ],
+ [
+ "l",
+ "ayout"
+ ],
+ [
+ "pl",
+ "ies"
+ ],
+ [
+ "▁b",
+ "uf"
+ ],
+ [
+ "▁bu",
+ "f"
+ ],
+ [
+ "▁",
+ "buf"
+ ],
+ [
+ "▁Is",
+ "land"
+ ],
+ [
+ "▁Ab",
+ "out"
+ ],
+ [
+ "▁",
+ "About"
+ ],
+ [
+ "▁w",
+ "est"
+ ],
+ [
+ "▁we",
+ "st"
+ ],
+ [
+ "▁",
+ "west"
+ ],
+ [
+ "▁S",
+ "cott"
+ ],
+ [
+ "▁Sc",
+ "ott"
+ ],
+ [
+ "▁Scot",
+ "t"
+ ],
+ [
+ "AC",
+ "T"
+ ],
+ [
+ "A",
+ "CT"
+ ],
+ [
+ "Wh",
+ "y"
+ ],
+ [
+ "W",
+ "hy"
+ ],
+ [
+ "▁large",
+ "st"
+ ],
+ [
+ "▁larg",
+ "est"
+ ],
+ [
+ "▁cont",
+ "ainer"
+ ],
+ [
+ "▁contain",
+ "er"
+ ],
+ [
+ "▁",
+ "container"
+ ],
+ [
+ "▁t",
+ "emperature"
+ ],
+ [
+ "▁temper",
+ "ature"
+ ],
+ [
+ "▁",
+ "£"
+ ],
+ [
+ "▁red",
+ "uce"
+ ],
+ [
+ "▁redu",
+ "ce"
+ ],
+ [
+ "▁",
+ "reduce"
+ ],
+ [
+ "▁f",
+ "oi"
+ ],
+ [
+ "▁fo",
+ "i"
+ ],
+ [
+ "ha",
+ "n"
+ ],
+ [
+ "h",
+ "an"
+ ],
+ [
+ "▁b",
+ "od"
+ ],
+ [
+ "▁bo",
+ "d"
+ ],
+ [
+ "▁V",
+ "an"
+ ],
+ [
+ "▁Va",
+ "n"
+ ],
+ [
+ "▁null",
+ "ptr"
+ ],
+ [
+ "▁",
+ "nullptr"
+ ],
+ [
+ "▁d",
+ "ating"
+ ],
+ [
+ "▁da",
+ "ting"
+ ],
+ [
+ "▁dat",
+ "ing"
+ ],
+ [
+ "▁",
+ "dating"
+ ],
+ [
+ "▁ch",
+ "ain"
+ ],
+ [
+ "▁cha",
+ "in"
+ ],
+ [
+ "▁",
+ "chain"
+ ],
+ [
+ "Fl",
+ "ags"
+ ],
+ [
+ "Flag",
+ "s"
+ ],
+ [
+ "ient",
+ "o"
+ ],
+ [
+ "ien",
+ "to"
+ ],
+ [
+ "i",
+ "ento"
+ ],
+ [
+ "so",
+ "rt"
+ ],
+ [
+ "sor",
+ "t"
+ ],
+ [
+ "s",
+ "ort"
+ ],
+ [
+ "▁f",
+ "an"
+ ],
+ [
+ "▁fa",
+ "n"
+ ],
+ [
+ "▁",
+ "fan"
+ ],
+ [
+ "▁det",
+ "ermine"
+ ],
+ [
+ "▁determ",
+ "ine"
+ ],
+ [
+ "▁determin",
+ "e"
+ ],
+ [
+ "▁deter",
+ "mine"
+ ],
+ [
+ "▁w",
+ "ear"
+ ],
+ [
+ "▁we",
+ "ar"
+ ],
+ [
+ "▁",
+ "wear"
+ ],
+ [
+ "B",
+ "E"
+ ],
+ [
+ "▁appropri",
+ "ate"
+ ],
+ [
+ "л",
+ "ся"
+ ],
+ [
+ "то",
+ "в"
+ ],
+ [
+ "т",
+ "ов"
+ ],
+ [
+ "▁go",
+ "als"
+ ],
+ [
+ "▁goal",
+ "s"
+ ],
+ [
+ "▁M",
+ "ap"
+ ],
+ [
+ "▁Ma",
+ "p"
+ ],
+ [
+ "▁",
+ "Map"
+ ],
+ [
+ "▁S",
+ "ar"
+ ],
+ [
+ "▁Sa",
+ "r"
+ ],
+ [
+ "▁O",
+ "ption"
+ ],
+ [
+ "▁Opt",
+ "ion"
+ ],
+ [
+ "▁",
+ "Option"
+ ],
+ [
+ "▁h",
+ "ate"
+ ],
+ [
+ "▁ha",
+ "te"
+ ],
+ [
+ "▁hat",
+ "e"
+ ],
+ [
+ "▁z",
+ "ijn"
+ ],
+ [
+ ",",
+ "-"
+ ],
+ [
+ "▁im",
+ "plied"
+ ],
+ [
+ "▁impl",
+ "ied"
+ ],
+ [
+ "bit",
+ "s"
+ ],
+ [
+ "bi",
+ "ts"
+ ],
+ [
+ "b",
+ "its"
+ ],
+ [
+ "▁M",
+ "en"
+ ],
+ [
+ "▁Me",
+ "n"
+ ],
+ [
+ "▁",
+ "Men"
+ ],
+ [
+ "sk",
+ "ip"
+ ],
+ [
+ "ski",
+ "p"
+ ],
+ [
+ "▁M",
+ "ond"
+ ],
+ [
+ "▁Mon",
+ "d"
+ ],
+ [
+ "▁Mo",
+ "nd"
+ ],
+ [
+ "▁H",
+ "on"
+ ],
+ [
+ "▁Ho",
+ "n"
+ ],
+ [
+ "▁pro",
+ "ve"
+ ],
+ [
+ "▁pr",
+ "ove"
+ ],
+ [
+ "▁prov",
+ "e"
+ ],
+ [
+ "va",
+ "n"
+ ],
+ [
+ "v",
+ "an"
+ ],
+ [
+ "▁tr",
+ "aff"
+ ],
+ [
+ "▁tra",
+ "ff"
+ ],
+ [
+ "▁in",
+ "tr"
+ ],
+ [
+ "▁int",
+ "r"
+ ],
+ [
+ "▁",
+ "intr"
+ ],
+ [
+ "pi",
+ "c"
+ ],
+ [
+ "p",
+ "ic"
+ ],
+ [
+ "▁dro",
+ "pped"
+ ],
+ [
+ "▁drop",
+ "ped"
+ ],
+ [
+ "▁w",
+ "erd"
+ ],
+ [
+ "▁we",
+ "rd"
+ ],
+ [
+ "▁wer",
+ "d"
+ ],
+ [
+ "▁separ",
+ "ate"
+ ],
+ [
+ "is",
+ "a"
+ ],
+ [
+ "i",
+ "sa"
+ ],
+ [
+ "▁t",
+ "ab"
+ ],
+ [
+ "▁ta",
+ "b"
+ ],
+ [
+ "▁",
+ "tab"
+ ],
+ [
+ "tm",
+ "l"
+ ],
+ [
+ "t",
+ "ml"
+ ],
+ [
+ "▁\"",
+ "$"
+ ],
+ [
+ "mu",
+ "tex"
+ ],
+ [
+ "mut",
+ "ex"
+ ],
+ [
+ "▁P",
+ "an"
+ ],
+ [
+ "▁Pa",
+ "n"
+ ],
+ [
+ "▁",
+ "Pan"
+ ],
+ [
+ "ser",
+ "ve"
+ ],
+ [
+ "serv",
+ "e"
+ ],
+ [
+ "s",
+ "erve"
+ ],
+ [
+ "▁hot",
+ "el"
+ ],
+ [
+ "▁L",
+ "ast"
+ ],
+ [
+ "▁La",
+ "st"
+ ],
+ [
+ "▁Las",
+ "t"
+ ],
+ [
+ "▁",
+ "Last"
+ ],
+ [
+ "st",
+ "ep"
+ ],
+ [
+ "ste",
+ "p"
+ ],
+ [
+ "▁v",
+ "ir"
+ ],
+ [
+ "▁vi",
+ "r"
+ ],
+ [
+ "▁",
+ "vir"
+ ],
+ [
+ "R",
+ "ule"
+ ],
+ [
+ "is",
+ "tan"
+ ],
+ [
+ "ist",
+ "an"
+ ],
+ [
+ "ista",
+ "n"
+ ],
+ [
+ "i",
+ "stan"
+ ],
+ [
+ "ot",
+ "ing"
+ ],
+ [
+ "oti",
+ "ng"
+ ],
+ [
+ "o",
+ "ting"
+ ],
+ [
+ "ar",
+ "ks"
+ ],
+ [
+ "ark",
+ "s"
+ ],
+ [
+ "(_",
+ "_"
+ ],
+ [
+ "(",
+ "__"
+ ],
+ [
+ "▁e",
+ "ls"
+ ],
+ [
+ "▁el",
+ "s"
+ ],
+ [
+ "▁",
+ "els"
+ ],
+ [
+ "Pl",
+ "ayer"
+ ],
+ [
+ "Play",
+ "er"
+ ],
+ [
+ "P",
+ "layer"
+ ],
+ [
+ "]",
+ "]"
+ ],
+ [
+ "ви",
+ "ч"
+ ],
+ [
+ "yc",
+ "h"
+ ],
+ [
+ "y",
+ "ch"
+ ],
+ [
+ "ex",
+ "ception"
+ ],
+ [
+ "except",
+ "ion"
+ ],
+ [
+ "=\"",
+ "../"
+ ],
+ [
+ "▁im",
+ "agine"
+ ],
+ [
+ "▁imag",
+ "ine"
+ ],
+ [
+ "▁imagin",
+ "e"
+ ],
+ [
+ "\"}",
+ ","
+ ],
+ [
+ "\"",
+ "},"
+ ],
+ [
+ "ic",
+ "ago"
+ ],
+ [
+ "ica",
+ "go"
+ ],
+ [
+ "el",
+ "er"
+ ],
+ [
+ "ele",
+ "r"
+ ],
+ [
+ "e",
+ "ler"
+ ],
+ [
+ "▁v",
+ "s"
+ ],
+ [
+ "▁",
+ "vs"
+ ],
+ [
+ "▁A",
+ "frica"
+ ],
+ [
+ "▁Afr",
+ "ica"
+ ],
+ [
+ "▁Bus",
+ "iness"
+ ],
+ [
+ "oc",
+ "ks"
+ ],
+ [
+ "ock",
+ "s"
+ ],
+ [
+ "o",
+ "cks"
+ ],
+ [
+ "▁p",
+ "rz"
+ ],
+ [
+ "▁pr",
+ "z"
+ ],
+ [
+ "▁fuck",
+ "ing"
+ ],
+ [
+ "▁p",
+ "icked"
+ ],
+ [
+ "▁pick",
+ "ed"
+ ],
+ [
+ "▁pic",
+ "ked"
+ ],
+ [
+ "▁в",
+ "і"
+ ],
+ [
+ "▁",
+ "ві"
+ ],
+ [
+ "▁\"",
+ ","
+ ],
+ [
+ "▁",
+ "\","
+ ],
+ [
+ "▁b",
+ "ott"
+ ],
+ [
+ "▁bo",
+ "tt"
+ ],
+ [
+ "▁bot",
+ "t"
+ ],
+ [
+ "▁fail",
+ "ure"
+ ],
+ [
+ "▁",
+ "failure"
+ ],
+ [
+ "[",
+ ":"
+ ],
+ [
+ "▁G",
+ "ar"
+ ],
+ [
+ "▁Ga",
+ "r"
+ ],
+ [
+ "ap",
+ "es"
+ ],
+ [
+ "ape",
+ "s"
+ ],
+ [
+ "a",
+ "pes"
+ ],
+ [
+ "up",
+ "le"
+ ],
+ [
+ "u",
+ "ple"
+ ],
+ [
+ "▁f",
+ "er"
+ ],
+ [
+ "▁fe",
+ "r"
+ ],
+ [
+ "▁",
+ "fer"
+ ],
+ [
+ "▁p",
+ "urchase"
+ ],
+ [
+ "▁purch",
+ "ase"
+ ],
+ [
+ "▁п",
+ "ер"
+ ],
+ [
+ "▁пе",
+ "р"
+ ],
+ [
+ "▁",
+ "пер"
+ ],
+ [
+ "▁b",
+ "ird"
+ ],
+ [
+ "▁bi",
+ "rd"
+ ],
+ [
+ "▁",
+ "bird"
+ ],
+ [
+ "W",
+ "idget"
+ ],
+ [
+ "▁Sund",
+ "ay"
+ ],
+ [
+ "▁Sun",
+ "day"
+ ],
+ [
+ "▁A",
+ "maz"
+ ],
+ [
+ "▁Am",
+ "az"
+ ],
+ [
+ "▁",
+ "Amaz"
+ ],
+ [
+ "▁cons",
+ "ult"
+ ],
+ [
+ "ut",
+ "sch"
+ ],
+ [
+ "uts",
+ "ch"
+ ],
+ [
+ "an",
+ "to"
+ ],
+ [
+ "ant",
+ "o"
+ ],
+ [
+ "St",
+ "orage"
+ ],
+ [
+ "▁he",
+ "ader"
+ ],
+ [
+ "▁head",
+ "er"
+ ],
+ [
+ "▁",
+ "header"
+ ],
+ [
+ "üh",
+ "r"
+ ],
+ [
+ "ü",
+ "hr"
+ ],
+ [
+ "▁H",
+ "a"
+ ],
+ [
+ "▁",
+ "Ha"
+ ],
+ [
+ "▁Associ",
+ "ation"
+ ],
+ [
+ "▁s",
+ "ight"
+ ],
+ [
+ "▁si",
+ "ght"
+ ],
+ [
+ "▁sig",
+ "ht"
+ ],
+ [
+ "▁sigh",
+ "t"
+ ],
+ [
+ "C",
+ "ell"
+ ],
+ [
+ "▁pro",
+ "file"
+ ],
+ [
+ "▁prof",
+ "ile"
+ ],
+ [
+ "▁",
+ "profile"
+ ],
+ [
+ "▁fem",
+ "ale"
+ ],
+ [
+ "å",
+ "n"
+ ],
+ [
+ "▁w",
+ "id"
+ ],
+ [
+ "▁",
+ "wid"
+ ],
+ [
+ "z",
+ "n"
+ ],
+ [
+ "Dir",
+ "ect"
+ ],
+ [
+ "Di",
+ "rect"
+ ],
+ [
+ "D",
+ "irect"
+ ],
+ [
+ "▁st",
+ "ret"
+ ],
+ [
+ "▁str",
+ "et"
+ ],
+ [
+ "▁stre",
+ "t"
+ ],
+ [
+ "▁",
+ "stret"
+ ],
+ [
+ "aa",
+ "t"
+ ],
+ [
+ "a",
+ "at"
+ ],
+ [
+ "▁pat",
+ "ient"
+ ],
+ [
+ "▁",
+ "patient"
+ ],
+ [
+ "he",
+ "re"
+ ],
+ [
+ "her",
+ "e"
+ ],
+ [
+ "h",
+ "ere"
+ ],
+ [
+ "▁A",
+ "tl"
+ ],
+ [
+ "▁At",
+ "l"
+ ],
+ [
+ "in",
+ "et"
+ ],
+ [
+ "ine",
+ "t"
+ ],
+ [
+ "i",
+ "net"
+ ],
+ [
+ "Def",
+ "inition"
+ ],
+ [
+ "im",
+ "ary"
+ ],
+ [
+ "ima",
+ "ry"
+ ],
+ [
+ "i",
+ "mary"
+ ],
+ [
+ "Pol",
+ "icy"
+ ],
+ [
+ "▁d",
+ "ut"
+ ],
+ [
+ "▁du",
+ "t"
+ ],
+ [
+ "▁major",
+ "ity"
+ ],
+ [
+ "с",
+ "і"
+ ],
+ [
+ "▁Pro",
+ "ject"
+ ],
+ [
+ "▁",
+ "Project"
+ ],
+ [
+ "By",
+ "Id"
+ ],
+ [
+ "▁belie",
+ "ved"
+ ],
+ [
+ "▁believe",
+ "d"
+ ],
+ [
+ "▁Mus",
+ "ic"
+ ],
+ [
+ "▁",
+ "Music"
+ ],
+ [
+ "з",
+ "ы"
+ ],
+ [
+ "an",
+ "ti"
+ ],
+ [
+ "ant",
+ "i"
+ ],
+ [
+ "▁o",
+ "der"
+ ],
+ [
+ "▁od",
+ "er"
+ ],
+ [
+ "▁",
+ "oder"
+ ],
+ [
+ "Ch",
+ "annel"
+ ],
+ [
+ "▁s",
+ "le"
+ ],
+ [
+ "▁sl",
+ "e"
+ ],
+ [
+ "▁sequ",
+ "ence"
+ ],
+ [
+ "▁",
+ "sequence"
+ ],
+ [
+ "▁pie",
+ "ces"
+ ],
+ [
+ "▁piece",
+ "s"
+ ],
+ [
+ "▁k",
+ "ne"
+ ],
+ [
+ "▁kn",
+ "e"
+ ],
+ [
+ "▁abs",
+ "olutely"
+ ],
+ [
+ "▁absolut",
+ "ely"
+ ],
+ [
+ "▁absolute",
+ "ly"
+ ],
+ [
+ "▁Phil",
+ "ip"
+ ],
+ [
+ "ab",
+ "ilities"
+ ],
+ [
+ "abil",
+ "ities"
+ ],
+ [
+ "Qu",
+ "e"
+ ],
+ [
+ "Q",
+ "ue"
+ ],
+ [
+ "▁K",
+ "ar"
+ ],
+ [
+ "▁Ka",
+ "r"
+ ],
+ [
+ "Ex",
+ "ecut"
+ ],
+ [
+ "Exec",
+ "ut"
+ ],
+ [
+ "▁D",
+ "evel"
+ ],
+ [
+ "▁De",
+ "vel"
+ ],
+ [
+ "▁Dev",
+ "el"
+ ],
+ [
+ "▁elect",
+ "ric"
+ ],
+ [
+ "ful",
+ "l"
+ ],
+ [
+ "fu",
+ "ll"
+ ],
+ [
+ "f",
+ "ull"
+ ],
+ [
+ "rol",
+ "led"
+ ],
+ [
+ "roll",
+ "ed"
+ ],
+ [
+ "Do",
+ "m"
+ ],
+ [
+ "D",
+ "om"
+ ],
+ [
+ "▁r",
+ "iver"
+ ],
+ [
+ "▁ri",
+ "ver"
+ ],
+ [
+ "▁riv",
+ "er"
+ ],
+ [
+ "▁",
+ "river"
+ ],
+ [
+ "▁health",
+ "y"
+ ],
+ [
+ "▁heal",
+ "thy"
+ ],
+ [
+ "▁ex",
+ "tern"
+ ],
+ [
+ "▁ext",
+ "ern"
+ ],
+ [
+ "fi",
+ "t"
+ ],
+ [
+ "f",
+ "it"
+ ],
+ [
+ "▁co",
+ "ach"
+ ],
+ [
+ "▁K",
+ "r"
+ ],
+ [
+ "as",
+ "ta"
+ ],
+ [
+ "ast",
+ "a"
+ ],
+ [
+ "a",
+ "sta"
+ ],
+ [
+ "Com",
+ "pat"
+ ],
+ [
+ "Comp",
+ "at"
+ ],
+ [
+ "▁e",
+ "xit"
+ ],
+ [
+ "▁ex",
+ "it"
+ ],
+ [
+ "▁",
+ "exit"
+ ],
+ [
+ "▁Con",
+ "st"
+ ],
+ [
+ "▁Cons",
+ "t"
+ ],
+ [
+ "▁",
+ "Const"
+ ],
+ [
+ "af",
+ "ter"
+ ],
+ [
+ "aft",
+ "er"
+ ],
+ [
+ "a",
+ "fter"
+ ],
+ [
+ "▁should",
+ "er"
+ ],
+ [
+ "▁j",
+ "obs"
+ ],
+ [
+ "▁job",
+ "s"
+ ],
+ [
+ "▁jo",
+ "bs"
+ ],
+ [
+ "zo",
+ "ne"
+ ],
+ [
+ "zon",
+ "e"
+ ],
+ [
+ "z",
+ "one"
+ ],
+ [
+ "▁s",
+ "ale"
+ ],
+ [
+ "▁sa",
+ "le"
+ ],
+ [
+ "▁sal",
+ "e"
+ ],
+ [
+ "ix",
+ "el"
+ ],
+ [
+ "▁determ",
+ "ined"
+ ],
+ [
+ "▁determine",
+ "d"
+ ],
+ [
+ "▁determin",
+ "ed"
+ ],
+ [
+ "▁any",
+ "way"
+ ],
+ [
+ "or",
+ "f"
+ ],
+ [
+ "o",
+ "rf"
+ ],
+ [
+ "▁G",
+ "er"
+ ],
+ [
+ "▁Ge",
+ "r"
+ ],
+ [
+ "all",
+ "el"
+ ],
+ [
+ "alle",
+ "l"
+ ],
+ [
+ "re",
+ "es"
+ ],
+ [
+ "ree",
+ "s"
+ ],
+ [
+ "r",
+ "ees"
+ ],
+ [
+ "as",
+ "m"
+ ],
+ [
+ "a",
+ "sm"
+ ],
+ [
+ "im",
+ "s"
+ ],
+ [
+ "i",
+ "ms"
+ ],
+ [
+ "▁rec",
+ "ords"
+ ],
+ [
+ "▁record",
+ "s"
+ ],
+ [
+ "▁",
+ "records"
+ ],
+ [
+ "▁cor",
+ "por"
+ ],
+ [
+ "▁int",
+ "ellig"
+ ],
+ [
+ "▁intel",
+ "lig"
+ ],
+ [
+ "▁P",
+ "rem"
+ ],
+ [
+ "▁Pr",
+ "em"
+ ],
+ [
+ "▁Pre",
+ "m"
+ ],
+ [
+ "▁d",
+ "riving"
+ ],
+ [
+ "▁dr",
+ "iving"
+ ],
+ [
+ "▁dri",
+ "ving"
+ ],
+ [
+ "▁driv",
+ "ing"
+ ],
+ [
+ "▁mar",
+ "riage"
+ ],
+ [
+ "▁Th",
+ "ank"
+ ],
+ [
+ "▁",
+ "Thank"
+ ],
+ [
+ "▁w",
+ "illing"
+ ],
+ [
+ "▁will",
+ "ing"
+ ],
+ [
+ "M",
+ "C"
+ ],
+ [
+ "Field",
+ "s"
+ ],
+ [
+ "It",
+ "ems"
+ ],
+ [
+ "Item",
+ "s"
+ ],
+ [
+ "▁m",
+ "icro"
+ ],
+ [
+ "▁mi",
+ "cro"
+ ],
+ [
+ "▁mic",
+ "ro"
+ ],
+ [
+ "▁l",
+ "ift"
+ ],
+ [
+ "▁li",
+ "ft"
+ ],
+ [
+ "▁lif",
+ "t"
+ ],
+ [
+ "ir",
+ "ection"
+ ],
+ [
+ "ire",
+ "ction"
+ ],
+ [
+ "irect",
+ "ion"
+ ],
+ [
+ "i",
+ "rection"
+ ],
+ [
+ "Acc",
+ "ount"
+ ],
+ [
+ "Ac",
+ "count"
+ ],
+ [
+ "▁arch",
+ "itect"
+ ],
+ [
+ "tr",
+ "ack"
+ ],
+ [
+ "tra",
+ "ck"
+ ],
+ [
+ "▁p",
+ "rin"
+ ],
+ [
+ "▁pr",
+ "in"
+ ],
+ [
+ "▁pri",
+ "n"
+ ],
+ [
+ "P",
+ "A"
+ ],
+ [
+ "▁r",
+ "uns"
+ ],
+ [
+ "▁run",
+ "s"
+ ],
+ [
+ "▁ru",
+ "ns"
+ ],
+ [
+ "▁Tex",
+ "as"
+ ],
+ [
+ "is",
+ "her"
+ ],
+ [
+ "ish",
+ "er"
+ ],
+ [
+ "en",
+ "sure"
+ ],
+ [
+ "ens",
+ "ure"
+ ],
+ [
+ "▁B",
+ "oth"
+ ],
+ [
+ "▁Bo",
+ "th"
+ ],
+ [
+ "▁Bot",
+ "h"
+ ],
+ [
+ "ко",
+ "м"
+ ],
+ [
+ "▁Col",
+ "or"
+ ],
+ [
+ "▁Co",
+ "lor"
+ ],
+ [
+ "▁",
+ "Color"
+ ],
+ [
+ "Reg",
+ "ister"
+ ],
+ [
+ "▁J",
+ "oe"
+ ],
+ [
+ "▁Jo",
+ "e"
+ ],
+ [
+ "ge",
+ "q"
+ ],
+ [
+ "g",
+ "eq"
+ ],
+ [
+ "le",
+ "ts"
+ ],
+ [
+ "let",
+ "s"
+ ],
+ [
+ "l",
+ "ets"
+ ],
+ [
+ "ad",
+ "ing"
+ ],
+ [
+ "adi",
+ "ng"
+ ],
+ [
+ "a",
+ "ding"
+ ],
+ [
+ "▁ar",
+ "my"
+ ],
+ [
+ "▁arm",
+ "y"
+ ],
+ [
+ "▁B",
+ "ank"
+ ],
+ [
+ "▁Ban",
+ "k"
+ ],
+ [
+ "▁",
+ "Bank"
+ ],
+ [
+ "ot",
+ "ic"
+ ],
+ [
+ "oti",
+ "c"
+ ],
+ [
+ "Pro",
+ "duct"
+ ],
+ [
+ "Produ",
+ "ct"
+ ],
+ [
+ "im",
+ "port"
+ ],
+ [
+ "imp",
+ "ort"
+ ],
+ [
+ "▁W",
+ "ed"
+ ],
+ [
+ "▁We",
+ "d"
+ ],
+ [
+ "▁c",
+ "ry"
+ ],
+ [
+ "▁cr",
+ "y"
+ ],
+ [
+ "gr",
+ "ade"
+ ],
+ [
+ "grad",
+ "e"
+ ],
+ [
+ "gra",
+ "de"
+ ],
+ [
+ "g",
+ "rade"
+ ],
+ [
+ "di",
+ "g"
+ ],
+ [
+ "d",
+ "ig"
+ ],
+ [
+ "ga",
+ "l"
+ ],
+ [
+ "g",
+ "al"
+ ],
+ [
+ "к",
+ "ла"
+ ],
+ [
+ "es",
+ "ted"
+ ],
+ [
+ "est",
+ "ed"
+ ],
+ [
+ "este",
+ "d"
+ ],
+ [
+ "e",
+ "sted"
+ ],
+ [
+ "õ",
+ "es"
+ ],
+ [
+ "ge",
+ "rs"
+ ],
+ [
+ "ger",
+ "s"
+ ],
+ [
+ "g",
+ "ers"
+ ],
+ [
+ "olog",
+ "ie"
+ ],
+ [
+ "olo",
+ "gie"
+ ],
+ [
+ "то",
+ "м"
+ ],
+ [
+ "ra",
+ "zy"
+ ],
+ [
+ "raz",
+ "y"
+ ],
+ [
+ "r",
+ "azy"
+ ],
+ [
+ "▁d",
+ "inner"
+ ],
+ [
+ "▁din",
+ "ner"
+ ],
+ [
+ "Q",
+ "U"
+ ],
+ [
+ "▁fin",
+ "gers"
+ ],
+ [
+ "▁fing",
+ "ers"
+ ],
+ [
+ "▁finger",
+ "s"
+ ],
+ [
+ "UL",
+ "E"
+ ],
+ [
+ "U",
+ "LE"
+ ],
+ [
+ "cl",
+ "aim"
+ ],
+ [
+ "▁adv",
+ "antage"
+ ],
+ [
+ "▁advant",
+ "age"
+ ],
+ [
+ "▁var",
+ "iable"
+ ],
+ [
+ "▁vari",
+ "able"
+ ],
+ [
+ "▁",
+ "variable"
+ ],
+ [
+ "▁med",
+ "ic"
+ ],
+ [
+ "▁medi",
+ "c"
+ ],
+ [
+ "▁m",
+ "ale"
+ ],
+ [
+ "▁ma",
+ "le"
+ ],
+ [
+ "▁mal",
+ "e"
+ ],
+ [
+ "▁circ",
+ "um"
+ ],
+ [
+ "▁м",
+ "і"
+ ],
+ [
+ "▁",
+ "мі"
+ ],
+ [
+ "▁inter",
+ "net"
+ ],
+ [
+ "▁intern",
+ "et"
+ ],
+ [
+ "W",
+ "N"
+ ],
+ [
+ "▁l",
+ "ab"
+ ],
+ [
+ "▁la",
+ "b"
+ ],
+ [
+ "▁",
+ "lab"
+ ],
+ [
+ "az",
+ "ine"
+ ],
+ [
+ "azi",
+ "ne"
+ ],
+ [
+ "ч",
+ "но"
+ ],
+ [
+ "▁l",
+ "oop"
+ ],
+ [
+ "▁lo",
+ "op"
+ ],
+ [
+ "▁",
+ "loop"
+ ],
+ [
+ "▁p",
+ "red"
+ ],
+ [
+ "▁pre",
+ "d"
+ ],
+ [
+ "▁pr",
+ "ed"
+ ],
+ [
+ "▁",
+ "pred"
+ ],
+ [
+ "▁con",
+ "sequ"
+ ],
+ [
+ "▁cons",
+ "equ"
+ ],
+ [
+ "▁conse",
+ "qu"
+ ],
+ [
+ "▁bal",
+ "ance"
+ ],
+ [
+ "▁",
+ "balance"
+ ],
+ [
+ "fort",
+ "un"
+ ],
+ [
+ "▁g",
+ "ift"
+ ],
+ [
+ "▁gi",
+ "ft"
+ ],
+ [
+ "▁d",
+ "rug"
+ ],
+ [
+ "▁dr",
+ "ug"
+ ],
+ [
+ "▁dru",
+ "g"
+ ],
+ [
+ "▁c",
+ "ash"
+ ],
+ [
+ "▁cas",
+ "h"
+ ],
+ [
+ "▁ca",
+ "sh"
+ ],
+ [
+ "ски",
+ "х"
+ ],
+ [
+ "с",
+ "ких"
+ ],
+ [
+ "r",
+ "g"
+ ],
+ [
+ "ist",
+ "ribut"
+ ],
+ [
+ "▁high",
+ "est"
+ ],
+ [
+ "▁hig",
+ "hest"
+ ],
+ [
+ "êm",
+ "e"
+ ],
+ [
+ "ê",
+ "me"
+ ],
+ [
+ "em",
+ "ph"
+ ],
+ [
+ "emp",
+ "h"
+ ],
+ [
+ "em",
+ "on"
+ ],
+ [
+ "e",
+ "mon"
+ ],
+ [
+ "▁per",
+ "formed"
+ ],
+ [
+ "▁perform",
+ "ed"
+ ],
+ [
+ "cu",
+ "t"
+ ],
+ [
+ "c",
+ "ut"
+ ],
+ [
+ "▁cl",
+ "oser"
+ ],
+ [
+ "▁close",
+ "r"
+ ],
+ [
+ "▁clos",
+ "er"
+ ],
+ [
+ "▁clo",
+ "ser"
+ ],
+ [
+ "▁be",
+ "coming"
+ ],
+ [
+ "▁bec",
+ "oming"
+ ],
+ [
+ "▁\"",
+ "\","
+ ],
+ [
+ "▁\"\"",
+ ","
+ ],
+ [
+ "st",
+ "ar"
+ ],
+ [
+ "sta",
+ "r"
+ ],
+ [
+ "s",
+ "tar"
+ ],
+ [
+ "pu",
+ "b"
+ ],
+ [
+ "p",
+ "ub"
+ ],
+ [
+ "▁pre",
+ "par"
+ ],
+ [
+ "▁prep",
+ "ar"
+ ],
+ [
+ "▁v",
+ "ote"
+ ],
+ [
+ "▁vo",
+ "te"
+ ],
+ [
+ "▁vot",
+ "e"
+ ],
+ [
+ "▁",
+ "vote"
+ ],
+ [
+ "il",
+ "de"
+ ],
+ [
+ "ild",
+ "e"
+ ],
+ [
+ "▁im",
+ "press"
+ ],
+ [
+ "▁imp",
+ "ress"
+ ],
+ [
+ "▁employ",
+ "ees"
+ ],
+ [
+ "▁employee",
+ "s"
+ ],
+ [
+ "▁e",
+ "inen"
+ ],
+ [
+ "▁ein",
+ "en"
+ ],
+ [
+ "▁eine",
+ "n"
+ ],
+ [
+ "▁sm",
+ "ooth"
+ ],
+ [
+ "▁s",
+ "now"
+ ],
+ [
+ "▁sn",
+ "ow"
+ ],
+ [
+ "▁p",
+ "urs"
+ ],
+ [
+ "▁pur",
+ "s"
+ ],
+ [
+ "▁pu",
+ "rs"
+ ],
+ [
+ "▁v",
+ "oc"
+ ],
+ [
+ "▁vo",
+ "c"
+ ],
+ [
+ "▁M",
+ "icrosoft"
+ ],
+ [
+ "▁Micro",
+ "soft"
+ ],
+ [
+ "▁",
+ "Microsoft"
+ ],
+ [
+ "P",
+ "U"
+ ],
+ [
+ "▁in",
+ "come"
+ ],
+ [
+ "▁inc",
+ "ome"
+ ],
+ [
+ "in",
+ "os"
+ ],
+ [
+ "ino",
+ "s"
+ ],
+ [
+ "i",
+ "nos"
+ ],
+ [
+ "▁oper",
+ "ator"
+ ],
+ [
+ "▁opera",
+ "tor"
+ ],
+ [
+ "▁",
+ "operator"
+ ],
+ [
+ "▁equ",
+ "ival"
+ ],
+ [
+ "▁pass",
+ "word"
+ ],
+ [
+ "▁",
+ "password"
+ ],
+ [
+ "ci",
+ "ón"
+ ],
+ [
+ "ció",
+ "n"
+ ],
+ [
+ "c",
+ "ión"
+ ],
+ [
+ "su",
+ "ccess"
+ ],
+ [
+ "▁e",
+ "mp"
+ ],
+ [
+ "▁em",
+ "p"
+ ],
+ [
+ "▁",
+ "emp"
+ ],
+ [
+ "HO",
+ "UT"
+ ],
+ [
+ "H",
+ "OUT"
+ ],
+ [
+ "▁c",
+ "a"
+ ],
+ [
+ "▁",
+ "ca"
+ ],
+ [
+ "fl",
+ "ag"
+ ],
+ [
+ "f",
+ "lag"
+ ],
+ [
+ "il",
+ "ly"
+ ],
+ [
+ "ill",
+ "y"
+ ],
+ [
+ "cre",
+ "te"
+ ],
+ [
+ "cr",
+ "ete"
+ ],
+ [
+ "cret",
+ "e"
+ ],
+ [
+ "fr",
+ "ak"
+ ],
+ [
+ "▁h",
+ "idden"
+ ],
+ [
+ "▁hid",
+ "den"
+ ],
+ [
+ "▁",
+ "hidden"
+ ],
+ [
+ "▁\"",
+ "%"
+ ],
+ [
+ "▁",
+ "\"%"
+ ],
+ [
+ "ER",
+ "N"
+ ],
+ [
+ "ро",
+ "ва"
+ ],
+ [
+ "ров",
+ "а"
+ ],
+ [
+ "▁U",
+ "N"
+ ],
+ [
+ "▁",
+ "UN"
+ ],
+ [
+ "ro",
+ "ke"
+ ],
+ [
+ "rok",
+ "e"
+ ],
+ [
+ "r",
+ "oke"
+ ],
+ [
+ "mi",
+ "ss"
+ ],
+ [
+ "m",
+ "iss"
+ ],
+ [
+ "▁s",
+ "plit"
+ ],
+ [
+ "▁sp",
+ "lit"
+ ],
+ [
+ "▁spl",
+ "it"
+ ],
+ [
+ "▁",
+ "split"
+ ],
+ [
+ "Re",
+ "ference"
+ ],
+ [
+ ")$",
+ ","
+ ],
+ [
+ ")",
+ "$,"
+ ],
+ [
+ "ep",
+ "er"
+ ],
+ [
+ "e",
+ "per"
+ ],
+ [
+ "▁N",
+ "O"
+ ],
+ [
+ "▁",
+ "NO"
+ ],
+ [
+ "▁s",
+ "quare"
+ ],
+ [
+ "▁squ",
+ "are"
+ ],
+ [
+ "▁",
+ "square"
+ ],
+ [
+ "su",
+ "r"
+ ],
+ [
+ "s",
+ "ur"
+ ],
+ [
+ "че",
+ "н"
+ ],
+ [
+ "ч",
+ "ен"
+ ],
+ [
+ "es",
+ "ter"
+ ],
+ [
+ "est",
+ "er"
+ ],
+ [
+ "este",
+ "r"
+ ],
+ [
+ "e",
+ "ster"
+ ],
+ [
+ "н",
+ "ь"
+ ],
+ [
+ "}",
+ "\""
+ ],
+ [
+ "ra",
+ "wn"
+ ],
+ [
+ "raw",
+ "n"
+ ],
+ [
+ "r",
+ "awn"
+ ],
+ [
+ "ru",
+ "le"
+ ],
+ [
+ "r",
+ "ule"
+ ],
+ [
+ "▁aud",
+ "ience"
+ ],
+ [
+ "es",
+ "te"
+ ],
+ [
+ "est",
+ "e"
+ ],
+ [
+ "e",
+ "ste"
+ ],
+ [
+ "em",
+ "s"
+ ],
+ [
+ "e",
+ "ms"
+ ],
+ [
+ "IC",
+ "ENSE"
+ ],
+ [
+ "▁I",
+ "ll"
+ ],
+ [
+ "▁Il",
+ "l"
+ ],
+ [
+ "▁",
+ "Ill"
+ ],
+ [
+ "US",
+ "E"
+ ],
+ [
+ "U",
+ "SE"
+ ],
+ [
+ "▁b",
+ "on"
+ ],
+ [
+ "▁bo",
+ "n"
+ ],
+ [
+ "▁",
+ "bon"
+ ],
+ [
+ "bu",
+ "r"
+ ],
+ [
+ "b",
+ "ur"
+ ],
+ [
+ "▁s",
+ "ick"
+ ],
+ [
+ "▁si",
+ "ck"
+ ],
+ [
+ "▁h",
+ "orse"
+ ],
+ [
+ "▁hor",
+ "se"
+ ],
+ [
+ "▁hors",
+ "e"
+ ],
+ [
+ "▁E",
+ "duc"
+ ],
+ [
+ "▁Ed",
+ "uc"
+ ],
+ [
+ "▁Edu",
+ "c"
+ ],
+ [
+ "▁benef",
+ "it"
+ ],
+ [
+ "▁c",
+ "ro"
+ ],
+ [
+ "▁cr",
+ "o"
+ ],
+ [
+ "▁",
+ "cro"
+ ],
+ [
+ "Ap",
+ "plication"
+ ],
+ [
+ "▁cor",
+ "re"
+ ],
+ [
+ "▁gu",
+ "arante"
+ ],
+ [
+ "DA",
+ "TA"
+ ],
+ [
+ "DAT",
+ "A"
+ ],
+ [
+ "D",
+ "ATA"
+ ],
+ [
+ "▁expl",
+ "ained"
+ ],
+ [
+ "▁explain",
+ "ed"
+ ],
+ [
+ "T",
+ "X"
+ ],
+ [
+ "▁o",
+ "nt"
+ ],
+ [
+ "▁on",
+ "t"
+ ],
+ [
+ "▁",
+ "ont"
+ ],
+ [
+ "▁F",
+ "lor"
+ ],
+ [
+ "▁Fl",
+ "or"
+ ],
+ [
+ "▁Flo",
+ "r"
+ ],
+ [
+ "▁re",
+ "ports"
+ ],
+ [
+ "▁rep",
+ "orts"
+ ],
+ [
+ "▁report",
+ "s"
+ ],
+ [
+ "▁Re",
+ "al"
+ ],
+ [
+ "▁",
+ "Real"
+ ],
+ [
+ "ud",
+ "ed"
+ ],
+ [
+ "ude",
+ "d"
+ ],
+ [
+ "u",
+ "ded"
+ ],
+ [
+ "le",
+ "an"
+ ],
+ [
+ "▁cit",
+ "iz"
+ ],
+ [
+ "▁dec",
+ "ide"
+ ],
+ [
+ "▁decid",
+ "e"
+ ],
+ [
+ "W",
+ "S"
+ ],
+ [
+ "▁do",
+ "main"
+ ],
+ [
+ "▁dom",
+ "ain"
+ ],
+ [
+ "▁",
+ "domain"
+ ],
+ [
+ "▁ref",
+ "lect"
+ ],
+ [
+ "▁",
+ "reflect"
+ ],
+ [
+ "▁min",
+ "imum"
+ ],
+ [
+ "▁minim",
+ "um"
+ ],
+ [
+ "▁le",
+ "gs"
+ ],
+ [
+ "▁leg",
+ "s"
+ ],
+ [
+ "▁sm",
+ "iled"
+ ],
+ [
+ "▁smile",
+ "d"
+ ],
+ [
+ "f",
+ "i"
+ ],
+ [
+ "▁p",
+ "ure"
+ ],
+ [
+ "▁pur",
+ "e"
+ ],
+ [
+ "▁pu",
+ "re"
+ ],
+ [
+ "▁C",
+ "ustom"
+ ],
+ [
+ "▁",
+ "Custom"
+ ],
+ [
+ "▁ess",
+ "ential"
+ ],
+ [
+ "▁observ",
+ "ed"
+ ],
+ [
+ "▁observe",
+ "d"
+ ],
+ [
+ "▁obs",
+ "erved"
+ ],
+ [
+ "By",
+ "tes"
+ ],
+ [
+ "Byte",
+ "s"
+ ],
+ [
+ "▁c",
+ "tx"
+ ],
+ [
+ "▁",
+ "ctx"
+ ],
+ [
+ "▁r",
+ "ates"
+ ],
+ [
+ "▁rate",
+ "s"
+ ],
+ [
+ "▁rat",
+ "es"
+ ],
+ [
+ "▁ra",
+ "tes"
+ ],
+ [
+ "mb",
+ "re"
+ ],
+ [
+ "m",
+ "bre"
+ ],
+ [
+ "▁w",
+ "orry"
+ ],
+ [
+ "▁wor",
+ "ry"
+ ],
+ [
+ ")",
+ "^"
+ ],
+ [
+ "▁Re",
+ "search"
+ ],
+ [
+ "▁Res",
+ "earch"
+ ],
+ [
+ "Ro",
+ "ot"
+ ],
+ [
+ "R",
+ "oot"
+ ],
+ [
+ "Window",
+ "s"
+ ],
+ [
+ "ult",
+ "ure"
+ ],
+ [
+ "ultur",
+ "e"
+ ],
+ [
+ "▁rel",
+ "ative"
+ ],
+ [
+ "▁relativ",
+ "e"
+ ],
+ [
+ "▁",
+ "relative"
+ ],
+ [
+ "▁s",
+ "eu"
+ ],
+ [
+ "▁se",
+ "u"
+ ],
+ [
+ "▁n",
+ "ie"
+ ],
+ [
+ "▁ni",
+ "e"
+ ],
+ [
+ "▁",
+ "nie"
+ ],
+ [
+ "▁s",
+ "hook"
+ ],
+ [
+ "▁sh",
+ "ook"
+ ],
+ [
+ "ious",
+ "ly"
+ ],
+ [
+ "i",
+ "ously"
+ ],
+ [
+ "▁ad",
+ "vert"
+ ],
+ [
+ "▁adv",
+ "ert"
+ ],
+ [
+ "Se",
+ "e"
+ ],
+ [
+ "S",
+ "ee"
+ ],
+ [
+ "▁Cent",
+ "ral"
+ ],
+ [
+ "▁b",
+ "atter"
+ ],
+ [
+ "▁batt",
+ "er"
+ ],
+ [
+ "▁bat",
+ "ter"
+ ],
+ [
+ "▁s",
+ "igned"
+ ],
+ [
+ "▁sign",
+ "ed"
+ ],
+ [
+ "▁sig",
+ "ned"
+ ],
+ [
+ "▁",
+ "signed"
+ ],
+ [
+ "T",
+ "S"
+ ],
+ [
+ "on",
+ "i"
+ ],
+ [
+ "o",
+ "ni"
+ ],
+ [
+ "▁pre",
+ "pared"
+ ],
+ [
+ "▁prep",
+ "ared"
+ ],
+ [
+ "▁prepar",
+ "ed"
+ ],
+ [
+ "▁prepare",
+ "d"
+ ],
+ [
+ "ga",
+ "te"
+ ],
+ [
+ "g",
+ "ate"
+ ],
+ [
+ "▁C",
+ "are"
+ ],
+ [
+ "▁Car",
+ "e"
+ ],
+ [
+ "▁Ca",
+ "re"
+ ],
+ [
+ "ca",
+ "re"
+ ],
+ [
+ "car",
+ "e"
+ ],
+ [
+ "c",
+ "are"
+ ],
+ [
+ "▁sup",
+ "ply"
+ ],
+ [
+ "▁supp",
+ "ly"
+ ],
+ [
+ "Ex",
+ "p"
+ ],
+ [
+ "E",
+ "xp"
+ ],
+ [
+ "bol",
+ "ds"
+ ],
+ [
+ "bold",
+ "s"
+ ],
+ [
+ "b",
+ "olds"
+ ],
+ [
+ "▁tr",
+ "ail"
+ ],
+ [
+ "▁tra",
+ "il"
+ ],
+ [
+ "▁f",
+ "ish"
+ ],
+ [
+ "▁fi",
+ "sh"
+ ],
+ [
+ "▁fis",
+ "h"
+ ],
+ [
+ "▁",
+ "fish"
+ ],
+ [
+ "▁un",
+ "its"
+ ],
+ [
+ "▁unit",
+ "s"
+ ],
+ [
+ "▁",
+ "units"
+ ],
+ [
+ "ven",
+ "ue"
+ ],
+ [
+ "v",
+ "enue"
+ ],
+ [
+ "х",
+ "и"
+ ],
+ [
+ "▁W",
+ "ood"
+ ],
+ [
+ "▁Wo",
+ "od"
+ ],
+ [
+ "▁c",
+ "ategory"
+ ],
+ [
+ "▁categ",
+ "ory"
+ ],
+ [
+ "▁categor",
+ "y"
+ ],
+ [
+ "▁",
+ "category"
+ ],
+ [
+ "▁b",
+ "le"
+ ],
+ [
+ "▁bl",
+ "e"
+ ],
+ [
+ "▁",
+ "ble"
+ ],
+ [
+ "▁over",
+ "ride"
+ ],
+ [
+ "▁",
+ "override"
+ ],
+ [
+ "fo",
+ "o"
+ ],
+ [
+ "f",
+ "oo"
+ ],
+ [
+ "▁influ",
+ "ence"
+ ],
+ [
+ "en",
+ "th"
+ ],
+ [
+ "ent",
+ "h"
+ ],
+ [
+ "ri",
+ "j"
+ ],
+ [
+ "r",
+ "ij"
+ ],
+ [
+ "▁ad",
+ "apt"
+ ],
+ [
+ "ic",
+ "ians"
+ ],
+ [
+ "ici",
+ "ans"
+ ],
+ [
+ "ician",
+ "s"
+ ],
+ [
+ "icia",
+ "ns"
+ ],
+ [
+ "de",
+ "leted"
+ ],
+ [
+ "del",
+ "eted"
+ ],
+ [
+ "delete",
+ "d"
+ ],
+ [
+ "▁v",
+ "ision"
+ ],
+ [
+ "▁vis",
+ "ion"
+ ],
+ [
+ "▁",
+ "vision"
+ ],
+ [
+ "ct",
+ "rl"
+ ],
+ [
+ "ctr",
+ "l"
+ ],
+ [
+ "c",
+ "trl"
+ ],
+ [
+ "L",
+ "ambda"
+ ],
+ [
+ "t",
+ "p"
+ ],
+ [
+ "mon",
+ "d"
+ ],
+ [
+ "mo",
+ "nd"
+ ],
+ [
+ "m",
+ "ond"
+ ],
+ [
+ "atur",
+ "day"
+ ],
+ [
+ "norm",
+ "al"
+ ],
+ [
+ "nor",
+ "mal"
+ ],
+ [
+ "n",
+ "ormal"
+ ],
+ [
+ "▁thous",
+ "and"
+ ],
+ [
+ "▁Prof",
+ "ess"
+ ],
+ [
+ "▁dise",
+ "ase"
+ ],
+ [
+ "cl",
+ "ip"
+ ],
+ [
+ "cli",
+ "p"
+ ],
+ [
+ "▁г",
+ "ра"
+ ],
+ [
+ "▁",
+ "гра"
+ ],
+ [
+ "bolds",
+ "ymbol"
+ ],
+ [
+ "bold",
+ "symbol"
+ ],
+ [
+ "O",
+ "B"
+ ],
+ [
+ "▁chall",
+ "enge"
+ ],
+ [
+ "▁challeng",
+ "e"
+ ],
+ [
+ "▁m",
+ "otion"
+ ],
+ [
+ "▁mot",
+ "ion"
+ ],
+ [
+ "▁w",
+ "his"
+ ],
+ [
+ "▁wh",
+ "is"
+ ],
+ [
+ "▁le",
+ "aders"
+ ],
+ [
+ "▁lead",
+ "ers"
+ ],
+ [
+ "▁leader",
+ "s"
+ ],
+ [
+ "▁col",
+ "on"
+ ],
+ [
+ "▁co",
+ "lon"
+ ],
+ [
+ "▁",
+ "colon"
+ ],
+ [
+ "▁s",
+ "uit"
+ ],
+ [
+ "▁su",
+ "it"
+ ],
+ [
+ "▁",
+ "suit"
+ ],
+ [
+ "mi",
+ "d"
+ ],
+ [
+ "m",
+ "id"
+ ],
+ [
+ "amp",
+ "ion"
+ ],
+ [
+ "á",
+ "g"
+ ],
+ [
+ "▁view",
+ "s"
+ ],
+ [
+ "▁vie",
+ "ws"
+ ],
+ [
+ "▁",
+ "views"
+ ],
+ [
+ "▁app",
+ "ears"
+ ],
+ [
+ "▁appe",
+ "ars"
+ ],
+ [
+ "▁appear",
+ "s"
+ ],
+ [
+ "an",
+ "cel"
+ ],
+ [
+ "ance",
+ "l"
+ ],
+ [
+ "anc",
+ "el"
+ ],
+ [
+ "▁z",
+ "we"
+ ],
+ [
+ "▁zw",
+ "e"
+ ],
+ [
+ "IS",
+ "T"
+ ],
+ [
+ "I",
+ "ST"
+ ],
+ [
+ "▁le",
+ "aves"
+ ],
+ [
+ "▁leave",
+ "s"
+ ],
+ [
+ "▁e",
+ "nh"
+ ],
+ [
+ "▁en",
+ "h"
+ ],
+ [
+ "▁",
+ "enh"
+ ],
+ [
+ "Act",
+ "ive"
+ ],
+ [
+ "Activ",
+ "e"
+ ],
+ [
+ "▁d",
+ "it"
+ ],
+ [
+ "▁di",
+ "t"
+ ],
+ [
+ "▁",
+ "dit"
+ ],
+ [
+ "if",
+ "icate"
+ ],
+ [
+ "ific",
+ "ate"
+ ],
+ [
+ "ifica",
+ "te"
+ ],
+ [
+ "mat",
+ "rix"
+ ],
+ [
+ "Ex",
+ "pression"
+ ],
+ [
+ "Exp",
+ "ression"
+ ],
+ [
+ "Expr",
+ "ession"
+ ],
+ [
+ "Express",
+ "ion"
+ ],
+ [
+ "Re",
+ "ader"
+ ],
+ [
+ "Read",
+ "er"
+ ],
+ [
+ "▁m",
+ "ental"
+ ],
+ [
+ "▁men",
+ "tal"
+ ],
+ [
+ "▁ment",
+ "al"
+ ],
+ [
+ "em",
+ "bre"
+ ],
+ [
+ "emb",
+ "re"
+ ],
+ [
+ "e",
+ "mbre"
+ ],
+ [
+ "▁de",
+ "cor"
+ ],
+ [
+ "▁dec",
+ "or"
+ ],
+ [
+ "▁",
+ "decor"
+ ],
+ [
+ "ar",
+ "ts"
+ ],
+ [
+ "art",
+ "s"
+ ],
+ [
+ "▁v",
+ "ent"
+ ],
+ [
+ "▁ve",
+ "nt"
+ ],
+ [
+ "▁ven",
+ "t"
+ ],
+ [
+ "▁",
+ "vent"
+ ],
+ [
+ "ne",
+ "l"
+ ],
+ [
+ "n",
+ "el"
+ ],
+ [
+ "line",
+ "s"
+ ],
+ [
+ "li",
+ "nes"
+ ],
+ [
+ "lin",
+ "es"
+ ],
+ [
+ "l",
+ "ines"
+ ],
+ [
+ "up",
+ "id"
+ ],
+ [
+ "u",
+ "pid"
+ ],
+ [
+ "er",
+ "ved"
+ ],
+ [
+ "erv",
+ "ed"
+ ],
+ [
+ "erve",
+ "d"
+ ],
+ [
+ "▁bo",
+ "ys"
+ ],
+ [
+ "▁boy",
+ "s"
+ ],
+ [
+ "▁",
+ "boys"
+ ],
+ [
+ "ал",
+ "ь"
+ ],
+ [
+ "а",
+ "ль"
+ ],
+ [
+ "MO",
+ "D"
+ ],
+ [
+ "M",
+ "OD"
+ ],
+ [
+ "is",
+ "l"
+ ],
+ [
+ "i",
+ "sl"
+ ],
+ [
+ "▁[",
+ "["
+ ],
+ [
+ "▁",
+ "[["
+ ],
+ [
+ "ph",
+ "y"
+ ],
+ [
+ "p",
+ "hy"
+ ],
+ [
+ "▁.",
+ "."
+ ],
+ [
+ "▁",
+ ".."
+ ],
+ [
+ "▁a",
+ "gent"
+ ],
+ [
+ "▁ag",
+ "ent"
+ ],
+ [
+ "▁age",
+ "nt"
+ ],
+ [
+ "▁",
+ "agent"
+ ],
+ [
+ "▁S",
+ "ervices"
+ ],
+ [
+ "▁Service",
+ "s"
+ ],
+ [
+ "▁Serv",
+ "ices"
+ ],
+ [
+ "▁",
+ "Services"
+ ],
+ [
+ "▁i",
+ "ron"
+ ],
+ [
+ "▁ir",
+ "on"
+ ],
+ [
+ "▁",
+ "iron"
+ ],
+ [
+ "▁com",
+ "ponents"
+ ],
+ [
+ "▁compon",
+ "ents"
+ ],
+ [
+ "▁component",
+ "s"
+ ],
+ [
+ "▁",
+ "components"
+ ],
+ [
+ "▁f",
+ "re"
+ ],
+ [
+ "▁fr",
+ "e"
+ ],
+ [
+ "▁",
+ "fre"
+ ],
+ [
+ "iction",
+ "ary"
+ ],
+ [
+ "▁t",
+ "ests"
+ ],
+ [
+ "▁te",
+ "sts"
+ ],
+ [
+ "▁test",
+ "s"
+ ],
+ [
+ "▁",
+ "tests"
+ ],
+ [
+ ".~",
+ "\\"
+ ],
+ [
+ ".",
+ "~\\"
+ ],
+ [
+ "ob",
+ "s"
+ ],
+ [
+ "o",
+ "bs"
+ ],
+ [
+ "▁М",
+ "и"
+ ],
+ [
+ "▁об",
+ "ла"
+ ],
+ [
+ "▁ass",
+ "ess"
+ ],
+ [
+ "▁Fr",
+ "iday"
+ ],
+ [
+ "▁we",
+ "ather"
+ ],
+ [
+ "k",
+ "g"
+ ],
+ [
+ "ст",
+ "ра"
+ ],
+ [
+ "с",
+ "тра"
+ ],
+ [
+ ".",
+ "}"
+ ],
+ [
+ "end",
+ "ant"
+ ],
+ [
+ "enda",
+ "nt"
+ ],
+ [
+ "an",
+ "na"
+ ],
+ [
+ "ann",
+ "a"
+ ],
+ [
+ "▁Japan",
+ "ese"
+ ],
+ [
+ "cm",
+ "p"
+ ],
+ [
+ "c",
+ "mp"
+ ],
+ [
+ "▁Ar",
+ "my"
+ ],
+ [
+ "▁Arm",
+ "y"
+ ],
+ [
+ "on",
+ "ym"
+ ],
+ [
+ "ony",
+ "m"
+ ],
+ [
+ "o",
+ "nym"
+ ],
+ [
+ "▁rel",
+ "ax"
+ ],
+ [
+ "date",
+ "s"
+ ],
+ [
+ "da",
+ "tes"
+ ],
+ [
+ "dat",
+ "es"
+ ],
+ [
+ "d",
+ "ates"
+ ],
+ [
+ "▁R",
+ "ussian"
+ ],
+ [
+ "▁Russ",
+ "ian"
+ ],
+ [
+ "▁Russia",
+ "n"
+ ],
+ [
+ "▁excell",
+ "ent"
+ ],
+ [
+ "')",
+ ")"
+ ],
+ [
+ "'",
+ "))"
+ ],
+ [
+ "IL",
+ "ITY"
+ ],
+ [
+ "▁sh",
+ "owing"
+ ],
+ [
+ "▁show",
+ "ing"
+ ],
+ [
+ "▁Dan",
+ "iel"
+ ],
+ [
+ "м",
+ "я"
+ ],
+ [
+ "▁M",
+ "ain"
+ ],
+ [
+ "▁Ma",
+ "in"
+ ],
+ [
+ "▁Mai",
+ "n"
+ ],
+ [
+ "▁",
+ "Main"
+ ],
+ [
+ "Ph",
+ "i"
+ ],
+ [
+ "P",
+ "hi"
+ ],
+ [
+ "▁R",
+ "ock"
+ ],
+ [
+ "▁Ro",
+ "ck"
+ ],
+ [
+ "▁Roc",
+ "k"
+ ],
+ [
+ "▁g",
+ "rew"
+ ],
+ [
+ "▁gr",
+ "ew"
+ ],
+ [
+ "▁gre",
+ "w"
+ ],
+ [
+ "▁y",
+ "ield"
+ ],
+ [
+ "i",
+ "ère"
+ ],
+ [
+ "se",
+ "g"
+ ],
+ [
+ "s",
+ "eg"
+ ],
+ [
+ "}}",
+ "$"
+ ],
+ [
+ "}",
+ "}$"
+ ],
+ [
+ "▁st",
+ "rict"
+ ],
+ [
+ "▁str",
+ "ict"
+ ],
+ [
+ "▁stri",
+ "ct"
+ ],
+ [
+ "▁",
+ "strict"
+ ],
+ [
+ "▁v",
+ "ehicle"
+ ],
+ [
+ "▁veh",
+ "icle"
+ ],
+ [
+ "U",
+ "D"
+ ],
+ [
+ "A",
+ "F"
+ ],
+ [
+ "S",
+ "w"
+ ],
+ [
+ "▁c",
+ "hest"
+ ],
+ [
+ "▁ch",
+ "est"
+ ],
+ [
+ "▁che",
+ "st"
+ ],
+ [
+ "▁off",
+ "icer"
+ ],
+ [
+ "▁offic",
+ "er"
+ ],
+ [
+ "▁office",
+ "r"
+ ],
+ [
+ "▁e",
+ "ar"
+ ],
+ [
+ "▁",
+ "ear"
+ ],
+ [
+ "HE",
+ "R"
+ ],
+ [
+ "H",
+ "ER"
+ ],
+ [
+ "no",
+ "on"
+ ],
+ [
+ "n",
+ "oon"
+ ],
+ [
+ "▁jour",
+ "ney"
+ ],
+ [
+ "N",
+ "T"
+ ],
+ [
+ "▁d",
+ "ivers"
+ ],
+ [
+ "▁di",
+ "vers"
+ ],
+ [
+ "▁div",
+ "ers"
+ ],
+ [
+ "▁diver",
+ "s"
+ ],
+ [
+ "▁dive",
+ "rs"
+ ],
+ [
+ "▁Fin",
+ "ally"
+ ],
+ [
+ "▁Final",
+ "ly"
+ ],
+ [
+ "F",
+ "ound"
+ ],
+ [
+ "▁A",
+ "S"
+ ],
+ [
+ "▁",
+ "AS"
+ ],
+ [
+ "ri",
+ "k"
+ ],
+ [
+ "r",
+ "ik"
+ ],
+ [
+ "▁con",
+ "str"
+ ],
+ [
+ "▁const",
+ "r"
+ ],
+ [
+ "▁cons",
+ "tr"
+ ],
+ [
+ "▁s",
+ "ust"
+ ],
+ [
+ "▁su",
+ "st"
+ ],
+ [
+ "▁sus",
+ "t"
+ ],
+ [
+ "ac",
+ "count"
+ ],
+ [
+ "acc",
+ "ount"
+ ],
+ [
+ "acco",
+ "unt"
+ ],
+ [
+ "▁w",
+ "alls"
+ ],
+ [
+ "▁wall",
+ "s"
+ ],
+ [
+ "▁wal",
+ "ls"
+ ],
+ [
+ "▁entire",
+ "ly"
+ ],
+ [
+ "It",
+ "er"
+ ],
+ [
+ "I",
+ "ter"
+ ],
+ [
+ "ch",
+ "a"
+ ],
+ [
+ "c",
+ "ha"
+ ],
+ [
+ "is",
+ "hes"
+ ],
+ [
+ "ish",
+ "es"
+ ],
+ [
+ "IV",
+ "E"
+ ],
+ [
+ "I",
+ "VE"
+ ],
+ [
+ "▁pr",
+ "ime"
+ ],
+ [
+ "▁prim",
+ "e"
+ ],
+ [
+ "▁pri",
+ "me"
+ ],
+ [
+ "▁",
+ "prime"
+ ],
+ [
+ "▁",
+ "…"
+ ],
+ [
+ "x",
+ "e"
+ ],
+ [
+ "ut",
+ "en"
+ ],
+ [
+ "ute",
+ "n"
+ ],
+ [
+ "u",
+ "ten"
+ ],
+ [
+ "ar",
+ "se"
+ ],
+ [
+ "ars",
+ "e"
+ ],
+ [
+ "▁P",
+ "a"
+ ],
+ [
+ "put",
+ "e"
+ ],
+ [
+ "pu",
+ "te"
+ ],
+ [
+ "p",
+ "ute"
+ ],
+ [
+ "ä",
+ "l"
+ ],
+ [
+ "▁prote",
+ "ction"
+ ],
+ [
+ "▁protect",
+ "ion"
+ ],
+ [
+ "▁prot",
+ "ection"
+ ],
+ [
+ "▁ke",
+ "ys"
+ ],
+ [
+ "▁key",
+ "s"
+ ],
+ [
+ "▁",
+ "keys"
+ ],
+ [
+ "Ma",
+ "y"
+ ],
+ [
+ "M",
+ "ay"
+ ],
+ [
+ "By",
+ "te"
+ ],
+ [
+ "Con",
+ "st"
+ ],
+ [
+ "Cons",
+ "t"
+ ],
+ [
+ "B",
+ "L"
+ ],
+ [
+ "▁п",
+ "е"
+ ],
+ [
+ "▁",
+ "пе"
+ ],
+ [
+ "▁s",
+ "pl"
+ ],
+ [
+ "▁sp",
+ "l"
+ ],
+ [
+ "▁",
+ "spl"
+ ],
+ [
+ "▁cl",
+ "othes"
+ ],
+ [
+ "▁cloth",
+ "es"
+ ],
+ [
+ "as",
+ "hed"
+ ],
+ [
+ "ash",
+ "ed"
+ ],
+ [
+ "Mar",
+ "k"
+ ],
+ [
+ "M",
+ "ark"
+ ],
+ [
+ "è",
+ "me"
+ ],
+ [
+ "▁f",
+ "ait"
+ ],
+ [
+ "▁fa",
+ "it"
+ ],
+ [
+ "▁introdu",
+ "ced"
+ ],
+ [
+ "▁introduce",
+ "d"
+ ],
+ [
+ "un",
+ "lock"
+ ],
+ [
+ "▁In",
+ "stead"
+ ],
+ [
+ "▁Inst",
+ "ead"
+ ],
+ [
+ "ans",
+ "ion"
+ ],
+ [
+ "reg",
+ "ion"
+ ],
+ [
+ "▁Amer",
+ "icans"
+ ],
+ [
+ "▁American",
+ "s"
+ ],
+ [
+ "▁America",
+ "ns"
+ ],
+ [
+ "▁ind",
+ "eed"
+ ],
+ [
+ "▁inde",
+ "ed"
+ ],
+ [
+ "wid",
+ "get"
+ ],
+ [
+ "w",
+ "idget"
+ ],
+ [
+ "▁real",
+ "ize"
+ ],
+ [
+ "▁realiz",
+ "e"
+ ],
+ [
+ "▁f",
+ "ro"
+ ],
+ [
+ "▁fr",
+ "o"
+ ],
+ [
+ "BI",
+ "T"
+ ],
+ [
+ "B",
+ "IT"
+ ],
+ [
+ "▁Re",
+ "act"
+ ],
+ [
+ "▁",
+ "React"
+ ],
+ [
+ "RE",
+ "AD"
+ ],
+ [
+ "as",
+ "ket"
+ ],
+ [
+ "ask",
+ "et"
+ ],
+ [
+ "ne",
+ "ver"
+ ],
+ [
+ "n",
+ "ever"
+ ],
+ [
+ "▁p",
+ "oll"
+ ],
+ [
+ "▁pol",
+ "l"
+ ],
+ [
+ "▁po",
+ "ll"
+ ],
+ [
+ "▁",
+ "poll"
+ ],
+ [
+ "ic",
+ "ol"
+ ],
+ [
+ "ico",
+ "l"
+ ],
+ [
+ "i",
+ "col"
+ ],
+ [
+ "▁p",
+ "rev"
+ ],
+ [
+ "▁pre",
+ "v"
+ ],
+ [
+ "▁pr",
+ "ev"
+ ],
+ [
+ "▁",
+ "prev"
+ ],
+ [
+ "▁h",
+ "yp"
+ ],
+ [
+ "▁hy",
+ "p"
+ ],
+ [
+ "▁F",
+ "ur"
+ ],
+ [
+ "▁Fu",
+ "r"
+ ],
+ [
+ "cl",
+ "oud"
+ ],
+ [
+ "▁L",
+ "ee"
+ ],
+ [
+ "▁Le",
+ "e"
+ ],
+ [
+ "pl",
+ "ing"
+ ],
+ [
+ "p",
+ "ling"
+ ],
+ [
+ "▁Ch",
+ "ild"
+ ],
+ [
+ "▁Chi",
+ "ld"
+ ],
+ [
+ "▁",
+ "Child"
+ ],
+ [
+ "▁ide",
+ "al"
+ ],
+ [
+ "▁idea",
+ "l"
+ ],
+ [
+ "Se",
+ "lector"
+ ],
+ [
+ "Select",
+ "or"
+ ],
+ [
+ "STAT",
+ "US"
+ ],
+ [
+ "uct",
+ "ure"
+ ],
+ [
+ "▁w",
+ "ine"
+ ],
+ [
+ "▁win",
+ "e"
+ ],
+ [
+ "▁poss",
+ "ibly"
+ ],
+ [
+ "▁put",
+ "ting"
+ ],
+ [
+ "▁r",
+ "iv"
+ ],
+ [
+ "▁ri",
+ "v"
+ ],
+ [
+ "▁",
+ "riv"
+ ],
+ [
+ "▁w",
+ "earing"
+ ],
+ [
+ "▁we",
+ "aring"
+ ],
+ [
+ "▁wear",
+ "ing"
+ ],
+ [
+ "▁S",
+ "ource"
+ ],
+ [
+ "▁",
+ "Source"
+ ],
+ [
+ "▁C",
+ "as"
+ ],
+ [
+ "▁Ca",
+ "s"
+ ],
+ [
+ "Ch",
+ "anged"
+ ],
+ [
+ "Change",
+ "d"
+ ],
+ [
+ "▁th",
+ "anks"
+ ],
+ [
+ "▁than",
+ "ks"
+ ],
+ [
+ "▁thank",
+ "s"
+ ],
+ [
+ "TI",
+ "ME"
+ ],
+ [
+ "TIM",
+ "E"
+ ],
+ [
+ "T",
+ "IME"
+ ],
+ [
+ "▁s",
+ "port"
+ ],
+ [
+ "▁sp",
+ "ort"
+ ],
+ [
+ "▁spo",
+ "rt"
+ ],
+ [
+ "▁A",
+ "ward"
+ ],
+ [
+ "▁Aw",
+ "ard"
+ ],
+ [
+ "▁g",
+ "lad"
+ ],
+ [
+ "▁gl",
+ "ad"
+ ],
+ [
+ "▁P",
+ "ass"
+ ],
+ [
+ "▁Pa",
+ "ss"
+ ],
+ [
+ "▁Pas",
+ "s"
+ ],
+ [
+ "▁",
+ "Pass"
+ ],
+ [
+ "▁P",
+ "os"
+ ],
+ [
+ "▁Po",
+ "s"
+ ],
+ [
+ "▁",
+ "Pos"
+ ],
+ [
+ "sc",
+ "he"
+ ],
+ [
+ "sch",
+ "e"
+ ],
+ [
+ "s",
+ "che"
+ ],
+ [
+ "▁C",
+ "D"
+ ],
+ [
+ "▁",
+ "CD"
+ ],
+ [
+ "▁aff",
+ "ord"
+ ],
+ [
+ "▁af",
+ "ford"
+ ],
+ [
+ "▁W",
+ "omen"
+ ],
+ [
+ "▁Wo",
+ "men"
+ ],
+ [
+ "▁D",
+ "istrict"
+ ],
+ [
+ "▁Di",
+ "strict"
+ ],
+ [
+ "▁Dist",
+ "rict"
+ ],
+ [
+ "▁id",
+ "entity"
+ ],
+ [
+ "▁ident",
+ "ity"
+ ],
+ [
+ "▁",
+ "identity"
+ ],
+ [
+ "▁part",
+ "ies"
+ ],
+ [
+ "▁par",
+ "ties"
+ ],
+ [
+ "▁partie",
+ "s"
+ ],
+ [
+ "▁parti",
+ "es"
+ ],
+ [
+ ":",
+ "%"
+ ],
+ [
+ "▁d",
+ "rag"
+ ],
+ [
+ "▁dr",
+ "ag"
+ ],
+ [
+ "▁",
+ "drag"
+ ],
+ [
+ "▁m",
+ "ai"
+ ],
+ [
+ "▁ma",
+ "i"
+ ],
+ [
+ "!",
+ "("
+ ],
+ [
+ "lang",
+ "le"
+ ],
+ [
+ "lan",
+ "gle"
+ ],
+ [
+ "l",
+ "angle"
+ ],
+ [
+ "▁kn",
+ "owing"
+ ],
+ [
+ "▁know",
+ "ing"
+ ],
+ [
+ "Pro",
+ "ject"
+ ],
+ [
+ "▁reg",
+ "arding"
+ ],
+ [
+ "▁regard",
+ "ing"
+ ],
+ [
+ "▁Jose",
+ "ph"
+ ],
+ [
+ "▁Jos",
+ "eph"
+ ],
+ [
+ "г",
+ "е"
+ ],
+ [
+ "▁D",
+ "ar"
+ ],
+ [
+ "▁Da",
+ "r"
+ ],
+ [
+ "▁H",
+ "or"
+ ],
+ [
+ "▁Ho",
+ "r"
+ ],
+ [
+ "▁",
+ "Hor"
+ ],
+ [
+ "▁anim",
+ "als"
+ ],
+ [
+ "▁animal",
+ "s"
+ ],
+ [
+ "▁ext",
+ "ension"
+ ],
+ [
+ "▁extens",
+ "ion"
+ ],
+ [
+ "▁",
+ "extension"
+ ],
+ [
+ "ска",
+ "я"
+ ],
+ [
+ "▁H",
+ "an"
+ ],
+ [
+ "▁Ha",
+ "n"
+ ],
+ [
+ "bt",
+ "n"
+ ],
+ [
+ "b",
+ "tn"
+ ],
+ [
+ "ac",
+ "iones"
+ ],
+ [
+ "aci",
+ "ones"
+ ],
+ [
+ "acion",
+ "es"
+ ],
+ [
+ "acio",
+ "nes"
+ ],
+ [
+ "▁f",
+ "amiliar"
+ ],
+ [
+ "▁fam",
+ "iliar"
+ ],
+ [
+ "▁famil",
+ "iar"
+ ],
+ [
+ "▁familia",
+ "r"
+ ],
+ [
+ "hol",
+ "der"
+ ],
+ [
+ "hold",
+ "er"
+ ],
+ [
+ "h",
+ "older"
+ ],
+ [
+ ":",
+ "\r"
+ ],
+ [
+ "st",
+ "ood"
+ ],
+ [
+ "sto",
+ "od"
+ ],
+ [
+ "▁li",
+ "ked"
+ ],
+ [
+ "▁like",
+ "d"
+ ],
+ [
+ "▁lik",
+ "ed"
+ ],
+ [
+ "CO",
+ "DE"
+ ],
+ [
+ "▁en",
+ "able"
+ ],
+ [
+ "▁",
+ "enable"
+ ],
+ [
+ "▁p",
+ "ed"
+ ],
+ [
+ "▁pe",
+ "d"
+ ],
+ [
+ "▁",
+ "ped"
+ ],
+ [
+ "it",
+ "i"
+ ],
+ [
+ "i",
+ "ti"
+ ],
+ [
+ "ha",
+ "b"
+ ],
+ [
+ "h",
+ "ab"
+ ],
+ [
+ "DI",
+ "R"
+ ],
+ [
+ "D",
+ "IR"
+ ],
+ [
+ "▁be",
+ "at"
+ ],
+ [
+ "▁",
+ "beat"
+ ],
+ [
+ "т",
+ "і"
+ ],
+ [
+ "▁Min",
+ "ister"
+ ],
+ [
+ "▁Mini",
+ "ster"
+ ],
+ [
+ "▁p",
+ "y"
+ ],
+ [
+ "▁",
+ "py"
+ ],
+ [
+ "P",
+ "at"
+ ],
+ [
+ "▁ex",
+ "hib"
+ ],
+ [
+ "▁exh",
+ "ib"
+ ],
+ [
+ "▁B",
+ "uild"
+ ],
+ [
+ "▁Bu",
+ "ild"
+ ],
+ [
+ "▁",
+ "Build"
+ ],
+ [
+ "▁F",
+ "ield"
+ ],
+ [
+ "▁Fi",
+ "eld"
+ ],
+ [
+ "▁",
+ "Field"
+ ],
+ [
+ "ic",
+ "ian"
+ ],
+ [
+ "ici",
+ "an"
+ ],
+ [
+ "icia",
+ "n"
+ ],
+ [
+ "▁coll",
+ "abor"
+ ],
+ [
+ "▁qu",
+ "arter"
+ ],
+ [
+ "▁quart",
+ "er"
+ ],
+ [
+ "▁quar",
+ "ter"
+ ],
+ [
+ "▁F",
+ "alse"
+ ],
+ [
+ "▁Fal",
+ "se"
+ ],
+ [
+ "▁",
+ "False"
+ ],
+ [
+ "k",
+ "m"
+ ],
+ [
+ "▁v",
+ "irtual"
+ ],
+ [
+ "▁virt",
+ "ual"
+ ],
+ [
+ "▁",
+ "virtual"
+ ],
+ [
+ "ow",
+ "a"
+ ],
+ [
+ "o",
+ "wa"
+ ],
+ [
+ "▁J",
+ "on"
+ ],
+ [
+ "▁Jo",
+ "n"
+ ],
+ [
+ "am",
+ "in"
+ ],
+ [
+ "ami",
+ "n"
+ ],
+ [
+ "a",
+ "min"
+ ],
+ [
+ "ue",
+ "n"
+ ],
+ [
+ "u",
+ "en"
+ ],
+ [
+ "▁и",
+ "н"
+ ],
+ [
+ "▁",
+ "ин"
+ ],
+ [
+ "im",
+ "ation"
+ ],
+ [
+ "imat",
+ "ion"
+ ],
+ [
+ "ov",
+ "ing"
+ ],
+ [
+ "ovi",
+ "ng"
+ ],
+ [
+ "o",
+ "ving"
+ ],
+ [
+ "▁test",
+ "ing"
+ ],
+ [
+ "▁",
+ "testing"
+ ],
+ [
+ "se",
+ "ct"
+ ],
+ [
+ "sec",
+ "t"
+ ],
+ [
+ "s",
+ "ect"
+ ],
+ [
+ "IT",
+ "ION"
+ ],
+ [
+ "I",
+ "TION"
+ ],
+ [
+ "!",
+ "\\"
+ ],
+ [
+ "ap",
+ "y"
+ ],
+ [
+ "a",
+ "py"
+ ],
+ [
+ "▁trans",
+ "ition"
+ ],
+ [
+ "▁transit",
+ "ion"
+ ],
+ [
+ "▁",
+ "transition"
+ ],
+ [
+ "os",
+ "itory"
+ ],
+ [
+ "OD",
+ "O"
+ ],
+ [
+ "O",
+ "DO"
+ ],
+ [
+ "P",
+ "D"
+ ],
+ [
+ "n",
+ "é"
+ ],
+ [
+ "▁gener",
+ "ate"
+ ],
+ [
+ "▁gene",
+ "rate"
+ ],
+ [
+ "▁",
+ "generate"
+ ],
+ [
+ "▁n",
+ "ative"
+ ],
+ [
+ "▁nat",
+ "ive"
+ ],
+ [
+ "▁",
+ "native"
+ ],
+ [
+ "▁(",
+ "'"
+ ],
+ [
+ "▁",
+ "('"
+ ],
+ [
+ "▁e",
+ "lle"
+ ],
+ [
+ "▁el",
+ "le"
+ ],
+ [
+ "▁ell",
+ "e"
+ ],
+ [
+ "▁",
+ "elle"
+ ],
+ [
+ "R",
+ "R"
+ ],
+ [
+ "▁h",
+ "un"
+ ],
+ [
+ "_-",
+ ">"
+ ],
+ [
+ "_",
+ "->"
+ ],
+ [
+ "ag",
+ "nost"
+ ],
+ [
+ "agn",
+ "ost"
+ ],
+ [
+ "▁pro",
+ "posed"
+ ],
+ [
+ "▁prop",
+ "osed"
+ ],
+ [
+ "▁propos",
+ "ed"
+ ],
+ [
+ "▁propose",
+ "d"
+ ],
+ [
+ "▁G",
+ "ame"
+ ],
+ [
+ "▁Ga",
+ "me"
+ ],
+ [
+ "▁Gam",
+ "e"
+ ],
+ [
+ "▁",
+ "Game"
+ ],
+ [
+ "▁eff",
+ "orts"
+ ],
+ [
+ "▁effort",
+ "s"
+ ],
+ [
+ "в",
+ "я"
+ ],
+ [
+ "t",
+ "c"
+ ],
+ [
+ "с",
+ "к"
+ ],
+ [
+ "▁int",
+ "ent"
+ ],
+ [
+ "▁inte",
+ "nt"
+ ],
+ [
+ "▁",
+ "intent"
+ ],
+ [
+ "▁B",
+ "re"
+ ],
+ [
+ "▁Br",
+ "e"
+ ],
+ [
+ "is",
+ "c"
+ ],
+ [
+ "i",
+ "sc"
+ ],
+ [
+ "▁pro",
+ "test"
+ ],
+ [
+ "▁prote",
+ "st"
+ ],
+ [
+ "▁prot",
+ "est"
+ ],
+ [
+ "▁h",
+ "olds"
+ ],
+ [
+ "▁hold",
+ "s"
+ ],
+ [
+ "▁hol",
+ "ds"
+ ],
+ [
+ "▁",
+ "holds"
+ ],
+ [
+ "om",
+ "etry"
+ ],
+ [
+ "ome",
+ "try"
+ ],
+ [
+ "omet",
+ "ry"
+ ],
+ [
+ "o",
+ "metry"
+ ],
+ [
+ "▁H",
+ "ave"
+ ],
+ [
+ "▁Ha",
+ "ve"
+ ],
+ [
+ "▁Hav",
+ "e"
+ ],
+ [
+ "▁",
+ "Have"
+ ],
+ [
+ "▁de",
+ "tail"
+ ],
+ [
+ "▁det",
+ "ail"
+ ],
+ [
+ "▁",
+ "detail"
+ ],
+ [
+ "▁WIT",
+ "HOUT"
+ ],
+ [
+ "▁WITH",
+ "OUT"
+ ],
+ [
+ "ye",
+ "r"
+ ],
+ [
+ "y",
+ "er"
+ ],
+ [
+ "▁K",
+ "on"
+ ],
+ [
+ "▁Ko",
+ "n"
+ ],
+ [
+ "▁not",
+ "iced"
+ ],
+ [
+ "▁notice",
+ "d"
+ ],
+ [
+ "▁require",
+ "ments"
+ ],
+ [
+ "▁requirement",
+ "s"
+ ],
+ [
+ "DE",
+ "BUG"
+ ],
+ [
+ "ki",
+ "ns"
+ ],
+ [
+ "kin",
+ "s"
+ ],
+ [
+ "k",
+ "ins"
+ ],
+ [
+ "▁S",
+ "pan"
+ ],
+ [
+ "▁Sp",
+ "an"
+ ],
+ [
+ "▁",
+ "Span"
+ ],
+ [
+ "▁c",
+ "ars"
+ ],
+ [
+ "▁car",
+ "s"
+ ],
+ [
+ "▁ca",
+ "rs"
+ ],
+ [
+ "me",
+ "ta"
+ ],
+ [
+ "met",
+ "a"
+ ],
+ [
+ "m",
+ "eta"
+ ],
+ [
+ "▁k",
+ "il"
+ ],
+ [
+ "▁ki",
+ "l"
+ ],
+ [
+ "▁",
+ "kil"
+ ],
+ [
+ "▁B",
+ "ron"
+ ],
+ [
+ "▁Br",
+ "on"
+ ],
+ [
+ "▁Bro",
+ "n"
+ ],
+ [
+ "▁experience",
+ "d"
+ ],
+ [
+ "▁experi",
+ "enced"
+ ],
+ [
+ "▁re",
+ "mind"
+ ],
+ [
+ "▁rem",
+ "ind"
+ ],
+ [
+ "our",
+ "se"
+ ],
+ [
+ "ours",
+ "e"
+ ],
+ [
+ "▁W",
+ "estern"
+ ],
+ [
+ "▁West",
+ "ern"
+ ],
+ [
+ "▁Wes",
+ "tern"
+ ],
+ [
+ "ter",
+ "ed"
+ ],
+ [
+ "te",
+ "red"
+ ],
+ [
+ "tere",
+ "d"
+ ],
+ [
+ "t",
+ "ered"
+ ],
+ [
+ "▁dev",
+ "ices"
+ ],
+ [
+ "▁device",
+ "s"
+ ],
+ [
+ "▁",
+ "devices"
+ ],
+ [
+ "▁pict",
+ "ures"
+ ],
+ [
+ "▁picture",
+ "s"
+ ],
+ [
+ "▁t",
+ "ut"
+ ],
+ [
+ "▁tu",
+ "t"
+ ],
+ [
+ "\"",
+ "`"
+ ],
+ [
+ "▁im",
+ "possible"
+ ],
+ [
+ "▁r",
+ "ail"
+ ],
+ [
+ "▁ra",
+ "il"
+ ],
+ [
+ "▁fe",
+ "els"
+ ],
+ [
+ "▁feel",
+ "s"
+ ],
+ [
+ "▁fee",
+ "ls"
+ ],
+ [
+ "ic",
+ "as"
+ ],
+ [
+ "ica",
+ "s"
+ ],
+ [
+ "i",
+ "cas"
+ ],
+ [
+ "il",
+ "ling"
+ ],
+ [
+ "ill",
+ "ing"
+ ],
+ [
+ "▁acc",
+ "ident"
+ ],
+ [
+ "▁'",
+ "@"
+ ],
+ [
+ "____",
+ "____"
+ ],
+ [
+ "▁n",
+ "otes"
+ ],
+ [
+ "▁not",
+ "es"
+ ],
+ [
+ "▁no",
+ "tes"
+ ],
+ [
+ "▁note",
+ "s"
+ ],
+ [
+ "▁",
+ "notes"
+ ],
+ [
+ "om",
+ "an"
+ ],
+ [
+ "oma",
+ "n"
+ ],
+ [
+ "o",
+ "man"
+ ],
+ [
+ "Par",
+ "ser"
+ ],
+ [
+ "Parse",
+ "r"
+ ],
+ [
+ "Pars",
+ "er"
+ ],
+ [
+ "▁dis",
+ "covered"
+ ],
+ [
+ "▁discover",
+ "ed"
+ ],
+ [
+ "▁R",
+ "oman"
+ ],
+ [
+ "▁Rom",
+ "an"
+ ],
+ [
+ "▁Ro",
+ "man"
+ ],
+ [
+ "▁Roma",
+ "n"
+ ],
+ [
+ "▁bud",
+ "get"
+ ],
+ [
+ "▁gu",
+ "ide"
+ ],
+ [
+ "▁guid",
+ "e"
+ ],
+ [
+ "ki",
+ "ng"
+ ],
+ [
+ "kin",
+ "g"
+ ],
+ [
+ "k",
+ "ing"
+ ],
+ [
+ "▁in",
+ "cred"
+ ],
+ [
+ "▁inc",
+ "red"
+ ],
+ [
+ "▁incre",
+ "d"
+ ],
+ [
+ "ol",
+ "ar"
+ ],
+ [
+ "ola",
+ "r"
+ ],
+ [
+ "o",
+ "lar"
+ ],
+ [
+ "en",
+ "den"
+ ],
+ [
+ "end",
+ "en"
+ ],
+ [
+ "ende",
+ "n"
+ ],
+ [
+ "Des",
+ "c"
+ ],
+ [
+ "De",
+ "sc"
+ ],
+ [
+ "D",
+ "esc"
+ ],
+ [
+ "▁w",
+ "ave"
+ ],
+ [
+ "▁wa",
+ "ve"
+ ],
+ [
+ "▁",
+ "wave"
+ ],
+ [
+ "б",
+ "ли"
+ ],
+ [
+ "ig",
+ "t"
+ ],
+ [
+ "i",
+ "gt"
+ ],
+ [
+ "▁re",
+ "strict"
+ ],
+ [
+ "▁rest",
+ "rict"
+ ],
+ [
+ "▁restr",
+ "ict"
+ ],
+ [
+ "▁R",
+ "et"
+ ],
+ [
+ "▁Re",
+ "t"
+ ],
+ [
+ "▁",
+ "Ret"
+ ],
+ [
+ "▁m",
+ "ac"
+ ],
+ [
+ "▁ma",
+ "c"
+ ],
+ [
+ "▁",
+ "mac"
+ ],
+ [
+ "у",
+ "р"
+ ],
+ [
+ "B",
+ "S"
+ ],
+ [
+ "í",
+ "s"
+ ],
+ [
+ "▁gener",
+ "ation"
+ ],
+ [
+ "de",
+ "m"
+ ],
+ [
+ "d",
+ "em"
+ ],
+ [
+ "al",
+ "o"
+ ],
+ [
+ "a",
+ "lo"
+ ],
+ [
+ "б",
+ "ра"
+ ],
+ [
+ "▁order",
+ "ed"
+ ],
+ [
+ "▁ord",
+ "ered"
+ ],
+ [
+ "▁",
+ "ordered"
+ ],
+ [
+ "dr",
+ "op"
+ ],
+ [
+ "dro",
+ "p"
+ ],
+ [
+ "d",
+ "rop"
+ ],
+ [
+ "▁p",
+ "p"
+ ],
+ [
+ "▁",
+ "pp"
+ ],
+ [
+ "▁Re",
+ "view"
+ ],
+ [
+ "▁Rev",
+ "iew"
+ ],
+ [
+ "▁",
+ "Review"
+ ],
+ [
+ "▁liter",
+ "ally"
+ ],
+ [
+ "▁literal",
+ "ly"
+ ],
+ [
+ "▁S",
+ "ir"
+ ],
+ [
+ "▁Si",
+ "r"
+ ],
+ [
+ "▁",
+ "Sir"
+ ],
+ [
+ "▁Y",
+ "eah"
+ ],
+ [
+ "▁Ye",
+ "ah"
+ ],
+ [
+ "▁",
+ "Yeah"
+ ],
+ [
+ "▁d",
+ "ensity"
+ ],
+ [
+ "▁dens",
+ "ity"
+ ],
+ [
+ "▁",
+ "density"
+ ],
+ [
+ "ri",
+ "z"
+ ],
+ [
+ "r",
+ "iz"
+ ],
+ [
+ "in",
+ "de"
+ ],
+ [
+ "ind",
+ "e"
+ ],
+ [
+ "i",
+ "nde"
+ ],
+ [
+ "▁g",
+ "ain"
+ ],
+ [
+ "▁ga",
+ "in"
+ ],
+ [
+ "▁",
+ "gain"
+ ],
+ [
+ "▁p",
+ "anel"
+ ],
+ [
+ "▁pan",
+ "el"
+ ],
+ [
+ "▁pa",
+ "nel"
+ ],
+ [
+ "▁",
+ "panel"
+ ],
+ [
+ "je",
+ "t"
+ ],
+ [
+ "j",
+ "et"
+ ],
+ [
+ "▁T",
+ "imes"
+ ],
+ [
+ "▁Time",
+ "s"
+ ],
+ [
+ "▁Tim",
+ "es"
+ ],
+ [
+ "▁Ti",
+ "mes"
+ ],
+ [
+ "▁",
+ "Times"
+ ],
+ [
+ "▁n",
+ "ella"
+ ],
+ [
+ "▁ne",
+ "lla"
+ ],
+ [
+ "▁nel",
+ "la"
+ ],
+ [
+ "▁nell",
+ "a"
+ ],
+ [
+ "▁pre",
+ "viously"
+ ],
+ [
+ "▁previous",
+ "ly"
+ ],
+ [
+ "▁prev",
+ "iously"
+ ],
+ [
+ "point",
+ "s"
+ ],
+ [
+ "Se",
+ "nd"
+ ],
+ [
+ "S",
+ "end"
+ ],
+ [
+ "▁B",
+ "rown"
+ ],
+ [
+ "▁Br",
+ "own"
+ ],
+ [
+ "▁Bro",
+ "wn"
+ ],
+ [
+ "▁Brow",
+ "n"
+ ],
+ [
+ "ea",
+ "ch"
+ ],
+ [
+ "e",
+ "ach"
+ ],
+ [
+ "▁tr",
+ "igger"
+ ],
+ [
+ "▁",
+ "trigger"
+ ],
+ [
+ "ome",
+ "times"
+ ],
+ [
+ "omet",
+ "imes"
+ ],
+ [
+ "ic",
+ "os"
+ ],
+ [
+ "ico",
+ "s"
+ ],
+ [
+ "i",
+ "cos"
+ ],
+ [
+ "G",
+ "R"
+ ],
+ [
+ "Pane",
+ "l"
+ ],
+ [
+ "Pan",
+ "el"
+ ],
+ [
+ "P",
+ "anel"
+ ],
+ [
+ "og",
+ "en"
+ ],
+ [
+ "oge",
+ "n"
+ ],
+ [
+ "o",
+ "gen"
+ ],
+ [
+ "▁c",
+ "m"
+ ],
+ [
+ "▁",
+ "cm"
+ ],
+ [
+ "ru",
+ "ctions"
+ ],
+ [
+ "ruct",
+ "ions"
+ ],
+ [
+ "ruction",
+ "s"
+ ],
+ [
+ "▁k",
+ "iss"
+ ],
+ [
+ "▁ki",
+ "ss"
+ ],
+ [
+ "▁s",
+ "olo"
+ ],
+ [
+ "▁so",
+ "lo"
+ ],
+ [
+ "▁sol",
+ "o"
+ ],
+ [
+ "▁f",
+ "amous"
+ ],
+ [
+ "▁fam",
+ "ous"
+ ],
+ [
+ "ra",
+ "n"
+ ],
+ [
+ "r",
+ "an"
+ ],
+ [
+ "п",
+ "ро"
+ ],
+ [
+ "▁th",
+ "ro"
+ ],
+ [
+ "▁thr",
+ "o"
+ ],
+ [
+ "Gr",
+ "aph"
+ ],
+ [
+ "G",
+ "raph"
+ ],
+ [
+ "im",
+ "it"
+ ],
+ [
+ "imi",
+ "t"
+ ],
+ [
+ "i",
+ "mit"
+ ],
+ [
+ "▁V",
+ "alue"
+ ],
+ [
+ "▁Val",
+ "ue"
+ ],
+ [
+ "▁",
+ "Value"
+ ],
+ [
+ "▁st",
+ "arts"
+ ],
+ [
+ "▁start",
+ "s"
+ ],
+ [
+ "▁star",
+ "ts"
+ ],
+ [
+ "ip",
+ "eline"
+ ],
+ [
+ "ipe",
+ "line"
+ ],
+ [
+ "h",
+ "d"
+ ],
+ [
+ "T",
+ "C"
+ ],
+ [
+ "▁dis",
+ "cussion"
+ ],
+ [
+ "▁discuss",
+ "ion"
+ ],
+ [
+ "▁tr",
+ "uck"
+ ],
+ [
+ "ak",
+ "a"
+ ],
+ [
+ "a",
+ "ka"
+ ],
+ [
+ "On",
+ "ly"
+ ],
+ [
+ "▁E",
+ "qu"
+ ],
+ [
+ "▁Eq",
+ "u"
+ ],
+ [
+ "▁",
+ "Equ"
+ ],
+ [
+ "▁k",
+ "ö"
+ ],
+ [
+ "▁",
+ "kö"
+ ],
+ [
+ "▁B",
+ "es"
+ ],
+ [
+ "▁Be",
+ "s"
+ ],
+ [
+ "▁crit",
+ "ic"
+ ],
+ [
+ "▁pro",
+ "pos"
+ ],
+ [
+ "▁prop",
+ "os"
+ ],
+ [
+ "▁b",
+ "att"
+ ],
+ [
+ "▁bat",
+ "t"
+ ],
+ [
+ "▁ba",
+ "tt"
+ ],
+ [
+ "▁S",
+ "ection"
+ ],
+ [
+ "▁Se",
+ "ction"
+ ],
+ [
+ "▁",
+ "Section"
+ ],
+ [
+ "Sh",
+ "ow"
+ ],
+ [
+ "S",
+ "how"
+ ],
+ [
+ "g",
+ "p"
+ ],
+ [
+ "ST",
+ "ATE"
+ ],
+ [
+ "STAT",
+ "E"
+ ],
+ [
+ "PO",
+ "ST"
+ ],
+ [
+ "POS",
+ "T"
+ ],
+ [
+ "P",
+ "OST"
+ ],
+ [
+ "▁N",
+ "ord"
+ ],
+ [
+ "▁No",
+ "rd"
+ ],
+ [
+ "▁Nor",
+ "d"
+ ],
+ [
+ "▁in",
+ "nov"
+ ],
+ [
+ "▁inn",
+ "ov"
+ ],
+ [
+ "▁c",
+ "rim"
+ ],
+ [
+ "▁cr",
+ "im"
+ ],
+ [
+ "▁cri",
+ "m"
+ ],
+ [
+ "▁",
+ "crim"
+ ],
+ [
+ "ax",
+ "is"
+ ],
+ [
+ "a",
+ "xis"
+ ],
+ [
+ "▁T",
+ "urn"
+ ],
+ [
+ "▁Tur",
+ "n"
+ ],
+ [
+ "▁Tu",
+ "rn"
+ ],
+ [
+ "▁",
+ "Turn"
+ ],
+ [
+ "con",
+ "n"
+ ],
+ [
+ "co",
+ "nn"
+ ],
+ [
+ "Run",
+ "time"
+ ],
+ [
+ "▁rem",
+ "aining"
+ ],
+ [
+ "▁remain",
+ "ing"
+ ],
+ [
+ "os",
+ "ton"
+ ],
+ [
+ "ost",
+ "on"
+ ],
+ [
+ "osto",
+ "n"
+ ],
+ [
+ "o",
+ "ston"
+ ],
+ [
+ "▁",
+ "Э"
+ ],
+ [
+ "▁window",
+ "s"
+ ],
+ [
+ "▁wind",
+ "ows"
+ ],
+ [
+ "▁",
+ "windows"
+ ],
+ [
+ "▁R",
+ "oyal"
+ ],
+ [
+ "▁Ro",
+ "yal"
+ ],
+ [
+ "▁Roy",
+ "al"
+ ],
+ [
+ "▁v",
+ "ide"
+ ],
+ [
+ "▁vi",
+ "de"
+ ],
+ [
+ "▁vid",
+ "e"
+ ],
+ [
+ "P",
+ "P"
+ ],
+ [
+ "ch",
+ "ron"
+ ],
+ [
+ "chr",
+ "on"
+ ],
+ [
+ "▁s",
+ "an"
+ ],
+ [
+ "▁sa",
+ "n"
+ ],
+ [
+ "▁",
+ "san"
+ ],
+ [
+ "▁r",
+ "ise"
+ ],
+ [
+ "▁ri",
+ "se"
+ ],
+ [
+ "▁ris",
+ "e"
+ ],
+ [
+ "▁",
+ "rise"
+ ],
+ [
+ "▁d",
+ "elle"
+ ],
+ [
+ "▁de",
+ "lle"
+ ],
+ [
+ "▁del",
+ "le"
+ ],
+ [
+ "▁dell",
+ "e"
+ ],
+ [
+ "▁D",
+ "ur"
+ ],
+ [
+ "▁Du",
+ "r"
+ ],
+ [
+ "▁rap",
+ "id"
+ ],
+ [
+ "▁ra",
+ "pid"
+ ],
+ [
+ "ce",
+ "rt"
+ ],
+ [
+ "cer",
+ "t"
+ ],
+ [
+ "c",
+ "ert"
+ ],
+ [
+ "L",
+ "A"
+ ],
+ [
+ "ed",
+ "ge"
+ ],
+ [
+ "▁\\",
+ "]"
+ ],
+ [
+ "▁",
+ "\\]"
+ ],
+ [
+ "▁en",
+ "tered"
+ ],
+ [
+ "▁ent",
+ "ered"
+ ],
+ [
+ "▁enter",
+ "ed"
+ ],
+ [
+ "▁l",
+ "aws"
+ ],
+ [
+ "▁la",
+ "ws"
+ ],
+ [
+ "▁law",
+ "s"
+ ],
+ [
+ "▁ph",
+ "oto"
+ ],
+ [
+ "▁phot",
+ "o"
+ ],
+ [
+ "▁",
+ "photo"
+ ],
+ [
+ "▁ap",
+ "plications"
+ ],
+ [
+ "▁applic",
+ "ations"
+ ],
+ [
+ "▁application",
+ "s"
+ ],
+ [
+ "▁appl",
+ "ications"
+ ],
+ [
+ "▁Ber",
+ "lin"
+ ],
+ [
+ "▁ar",
+ "rest"
+ ],
+ [
+ "▁arr",
+ "est"
+ ],
+ [
+ "▁f",
+ "ederal"
+ ],
+ [
+ "▁fed",
+ "eral"
+ ],
+ [
+ "▁feder",
+ "al"
+ ],
+ [
+ "▁R",
+ "ussia"
+ ],
+ [
+ "▁Russ",
+ "ia"
+ ],
+ [
+ "▁us",
+ "ual"
+ ],
+ [
+ "▁r",
+ "aw"
+ ],
+ [
+ "▁ra",
+ "w"
+ ],
+ [
+ "▁",
+ "raw"
+ ],
+ [
+ "▁pi",
+ "ù"
+ ],
+ [
+ "êt",
+ "re"
+ ],
+ [
+ "ê",
+ "tre"
+ ],
+ [
+ "JS",
+ "ON"
+ ],
+ [
+ "J",
+ "SON"
+ ],
+ [
+ "SI",
+ "ON"
+ ],
+ [
+ "S",
+ "ION"
+ ],
+ [
+ "xt",
+ "ure"
+ ],
+ [
+ "ist",
+ "ent"
+ ],
+ [
+ "iste",
+ "nt"
+ ],
+ [
+ "isten",
+ "t"
+ ],
+ [
+ "▁P",
+ "ower"
+ ],
+ [
+ "▁Po",
+ "wer"
+ ],
+ [
+ "▁Pow",
+ "er"
+ ],
+ [
+ "▁",
+ "Power"
+ ],
+ [
+ "Bi",
+ "t"
+ ],
+ [
+ "B",
+ "it"
+ ],
+ [
+ "▁cap",
+ "acity"
+ ],
+ [
+ "▁capac",
+ "ity"
+ ],
+ [
+ "▁",
+ "capacity"
+ ],
+ [
+ "▁c",
+ "ards"
+ ],
+ [
+ "▁car",
+ "ds"
+ ],
+ [
+ "▁card",
+ "s"
+ ],
+ [
+ "▁",
+ "cards"
+ ],
+ [
+ "UI",
+ "D"
+ ],
+ [
+ "U",
+ "ID"
+ ],
+ [
+ "im",
+ "ents"
+ ],
+ [
+ "iment",
+ "s"
+ ],
+ [
+ "imen",
+ "ts"
+ ],
+ [
+ "i",
+ "ments"
+ ],
+ [
+ "▁d",
+ "ar"
+ ],
+ [
+ "▁da",
+ "r"
+ ],
+ [
+ "▁",
+ "dar"
+ ],
+ [
+ "▁Ch",
+ "icago"
+ ],
+ [
+ "▁comfort",
+ "able"
+ ],
+ [
+ "ti",
+ "p"
+ ],
+ [
+ "t",
+ "ip"
+ ],
+ [
+ "ba",
+ "s"
+ ],
+ [
+ "b",
+ "as"
+ ],
+ [
+ "▁m",
+ "u"
+ ],
+ [
+ "▁",
+ "mu"
+ ],
+ [
+ "▁en",
+ "emy"
+ ],
+ [
+ "▁enem",
+ "y"
+ ],
+ [
+ "ya",
+ "n"
+ ],
+ [
+ "y",
+ "an"
+ ],
+ [
+ "▁ф",
+ "и"
+ ],
+ [
+ "▁",
+ "фи"
+ ],
+ [
+ "▁up",
+ "dated"
+ ],
+ [
+ "▁update",
+ "d"
+ ],
+ [
+ "▁",
+ "updated"
+ ],
+ [
+ "an",
+ "go"
+ ],
+ [
+ "ang",
+ "o"
+ ],
+ [
+ "E",
+ "v"
+ ],
+ [
+ "E",
+ "ffect"
+ ],
+ [
+ "os",
+ "ing"
+ ],
+ [
+ "osi",
+ "ng"
+ ],
+ [
+ "o",
+ "sing"
+ ],
+ [
+ "ren",
+ "ce"
+ ],
+ [
+ "r",
+ "ence"
+ ],
+ [
+ "▁Con",
+ "gress"
+ ],
+ [
+ "▁Cong",
+ "ress"
+ ],
+ [
+ "▁d",
+ "efe"
+ ],
+ [
+ "▁de",
+ "fe"
+ ],
+ [
+ "▁def",
+ "e"
+ ],
+ [
+ "▁i",
+ "p"
+ ],
+ [
+ "▁",
+ "ip"
+ ],
+ [
+ "▁t",
+ "out"
+ ],
+ [
+ "▁to",
+ "ut"
+ ],
+ [
+ "▁tou",
+ "t"
+ ],
+ [
+ "▁f",
+ "reedom"
+ ],
+ [
+ "▁free",
+ "dom"
+ ],
+ [
+ "▁freed",
+ "om"
+ ],
+ [
+ "▁a",
+ "o"
+ ],
+ [
+ "▁",
+ "ao"
+ ],
+ [
+ "▁There",
+ "fore"
+ ],
+ [
+ "▁Ther",
+ "efore"
+ ],
+ [
+ "Ed",
+ "it"
+ ],
+ [
+ "E",
+ "dit"
+ ],
+ [
+ "▁Vir",
+ "gin"
+ ],
+ [
+ "RE",
+ "E"
+ ],
+ [
+ "R",
+ "EE"
+ ],
+ [
+ "ar",
+ "go"
+ ],
+ [
+ "arg",
+ "o"
+ ],
+ [
+ "▁D",
+ "am"
+ ],
+ [
+ "▁Da",
+ "m"
+ ],
+ [
+ "▁",
+ "Dam"
+ ],
+ [
+ "▁tra",
+ "ffic"
+ ],
+ [
+ "▁traff",
+ "ic"
+ ],
+ [
+ "ño",
+ "s"
+ ],
+ [
+ "ñ",
+ "os"
+ ],
+ [
+ "▁a",
+ "lle"
+ ],
+ [
+ "▁al",
+ "le"
+ ],
+ [
+ "▁all",
+ "e"
+ ],
+ [
+ "▁",
+ "alle"
+ ],
+ [
+ "▁dep",
+ "th"
+ ],
+ [
+ "▁",
+ "depth"
+ ],
+ [
+ "No",
+ "w"
+ ],
+ [
+ "N",
+ "ow"
+ ],
+ [
+ "▁s",
+ "ides"
+ ],
+ [
+ "▁side",
+ "s"
+ ],
+ [
+ "▁si",
+ "des"
+ ],
+ [
+ "▁sid",
+ "es"
+ ],
+ [
+ "▁го",
+ "ди"
+ ],
+ [
+ "▁год",
+ "и"
+ ],
+ [
+ "Des",
+ "criptor"
+ ],
+ [
+ "▁art",
+ "ikel"
+ ],
+ [
+ "▁n",
+ "arrow"
+ ],
+ [
+ "▁narr",
+ "ow"
+ ],
+ [
+ "▁nar",
+ "row"
+ ],
+ [
+ "__",
+ "_"
+ ],
+ [
+ "_",
+ "__"
+ ],
+ [
+ "k",
+ "w"
+ ],
+ [
+ "ut",
+ "o"
+ ],
+ [
+ "u",
+ "to"
+ ],
+ [
+ "▁Face",
+ "book"
+ ],
+ [
+ "▁Fac",
+ "ebook"
+ ],
+ [
+ "te",
+ "gr"
+ ],
+ [
+ "t",
+ "egr"
+ ],
+ [
+ "bo",
+ "olean"
+ ],
+ [
+ "ni",
+ "k"
+ ],
+ [
+ "n",
+ "ik"
+ ],
+ [
+ "b",
+ "d"
+ ],
+ [
+ "Tr",
+ "ack"
+ ],
+ [
+ "Tra",
+ "ck"
+ ],
+ [
+ "▁g",
+ "ran"
+ ],
+ [
+ "▁gr",
+ "an"
+ ],
+ [
+ "▁gra",
+ "n"
+ ],
+ [
+ "res",
+ "hold"
+ ],
+ [
+ "resh",
+ "old"
+ ],
+ [
+ "ве",
+ "т"
+ ],
+ [
+ "в",
+ "ет"
+ ],
+ [
+ "wr",
+ "ap"
+ ],
+ [
+ "w",
+ "rap"
+ ],
+ [
+ "▁n",
+ "oise"
+ ],
+ [
+ "▁no",
+ "ise"
+ ],
+ [
+ "ig",
+ "u"
+ ],
+ [
+ "i",
+ "gu"
+ ],
+ [
+ "▁B",
+ "on"
+ ],
+ [
+ "▁Bo",
+ "n"
+ ],
+ [
+ "▁",
+ "Bon"
+ ],
+ [
+ "▁w",
+ "y"
+ ],
+ [
+ "▁",
+ "wy"
+ ],
+ [
+ "lin",
+ "ux"
+ ],
+ [
+ "ck",
+ "s"
+ ],
+ [
+ "c",
+ "ks"
+ ],
+ [
+ "▁f",
+ "ans"
+ ],
+ [
+ "▁fa",
+ "ns"
+ ],
+ [
+ "▁fan",
+ "s"
+ ],
+ [
+ "▁m",
+ "ach"
+ ],
+ [
+ "▁ma",
+ "ch"
+ ],
+ [
+ "▁mac",
+ "h"
+ ],
+ [
+ "▁p",
+ "rices"
+ ],
+ [
+ "▁pr",
+ "ices"
+ ],
+ [
+ "▁pri",
+ "ces"
+ ],
+ [
+ "▁price",
+ "s"
+ ],
+ [
+ "é",
+ "v"
+ ],
+ [
+ "ou",
+ "ts"
+ ],
+ [
+ "out",
+ "s"
+ ],
+ [
+ "o",
+ "uts"
+ ],
+ [
+ "stand",
+ "ing"
+ ],
+ [
+ "stan",
+ "ding"
+ ],
+ [
+ "▁c",
+ "ateg"
+ ],
+ [
+ "▁cat",
+ "eg"
+ ],
+ [
+ ";",
+ "\\"
+ ],
+ [
+ "▁de",
+ "cre"
+ ],
+ [
+ "▁dec",
+ "re"
+ ],
+ [
+ "▁S",
+ "aturday"
+ ],
+ [
+ "▁m",
+ "enu"
+ ],
+ [
+ "▁me",
+ "nu"
+ ],
+ [
+ "▁men",
+ "u"
+ ],
+ [
+ "▁",
+ "menu"
+ ],
+ [
+ "▁N",
+ "ov"
+ ],
+ [
+ "▁No",
+ "v"
+ ],
+ [
+ "▁Y",
+ "et"
+ ],
+ [
+ "▁Ye",
+ "t"
+ ],
+ [
+ "▁та",
+ "к"
+ ],
+ [
+ "lic",
+ "he"
+ ],
+ [
+ "li",
+ "che"
+ ],
+ [
+ "lich",
+ "e"
+ ],
+ [
+ "l",
+ "iche"
+ ],
+ [
+ "▁Ac",
+ "adem"
+ ],
+ [
+ "▁commun",
+ "ication"
+ ],
+ [
+ "us",
+ "ing"
+ ],
+ [
+ "u",
+ "sing"
+ ],
+ [
+ "▁Soc",
+ "iety"
+ ],
+ [
+ "▁Soci",
+ "ety"
+ ],
+ [
+ "▁n",
+ "uc"
+ ],
+ [
+ "▁nu",
+ "c"
+ ],
+ [
+ "pect",
+ "ive"
+ ],
+ [
+ "or",
+ "ial"
+ ],
+ [
+ "oria",
+ "l"
+ ],
+ [
+ "ori",
+ "al"
+ ],
+ [
+ "o",
+ "rial"
+ ],
+ [
+ "▁af",
+ "raid"
+ ],
+ [
+ "▁an",
+ "imal"
+ ],
+ [
+ "▁anim",
+ "al"
+ ],
+ [
+ "▁turn",
+ "ing"
+ ],
+ [
+ "▁tur",
+ "ning"
+ ],
+ [
+ "ds",
+ "t"
+ ],
+ [
+ "d",
+ "st"
+ ],
+ [
+ "math",
+ "frak"
+ ],
+ [
+ "le",
+ "rs"
+ ],
+ [
+ "ler",
+ "s"
+ ],
+ [
+ "l",
+ "ers"
+ ],
+ [
+ "▁l",
+ "ots"
+ ],
+ [
+ "▁lo",
+ "ts"
+ ],
+ [
+ "▁lot",
+ "s"
+ ],
+ [
+ "▁",
+ "á"
+ ],
+ [
+ "▁T",
+ "ra"
+ ],
+ [
+ "▁Tr",
+ "a"
+ ],
+ [
+ "▁",
+ "Tra"
+ ],
+ [
+ "n",
+ "p"
+ ],
+ [
+ "▁r",
+ "ose"
+ ],
+ [
+ "▁ro",
+ "se"
+ ],
+ [
+ "▁",
+ "rose"
+ ],
+ [
+ "▁G",
+ "L"
+ ],
+ [
+ "▁",
+ "GL"
+ ],
+ [
+ "▁hel",
+ "ping"
+ ],
+ [
+ "▁help",
+ "ing"
+ ],
+ [
+ "▁w",
+ "inter"
+ ],
+ [
+ "▁win",
+ "ter"
+ ],
+ [
+ "▁ко",
+ "м"
+ ],
+ [
+ "▁",
+ "ком"
+ ],
+ [
+ "Mo",
+ "ck"
+ ],
+ [
+ "M",
+ "ock"
+ ],
+ [
+ "▁invest",
+ "ment"
+ ],
+ [
+ "Us",
+ "e"
+ ],
+ [
+ "U",
+ "se"
+ ],
+ [
+ "▁Can",
+ "ad"
+ ],
+ [
+ "н",
+ "д"
+ ],
+ [
+ "Co",
+ "py"
+ ],
+ [
+ "Cop",
+ "y"
+ ],
+ [
+ "C",
+ "opy"
+ ],
+ [
+ "▁f",
+ "ly"
+ ],
+ [
+ "▁fl",
+ "y"
+ ],
+ [
+ "▁",
+ "fly"
+ ],
+ [
+ "SE",
+ "R"
+ ],
+ [
+ "S",
+ "ER"
+ ],
+ [
+ "▁F",
+ "ar"
+ ],
+ [
+ "▁Fa",
+ "r"
+ ],
+ [
+ "▁R",
+ "os"
+ ],
+ [
+ "▁Ro",
+ "s"
+ ],
+ [
+ "am",
+ "il"
+ ],
+ [
+ "ami",
+ "l"
+ ],
+ [
+ "a",
+ "mil"
+ ],
+ [
+ "▁fight",
+ "ing"
+ ],
+ [
+ "▁rel",
+ "igious"
+ ],
+ [
+ "▁relig",
+ "ious"
+ ],
+ [
+ "su",
+ "per"
+ ],
+ [
+ "sup",
+ "er"
+ ],
+ [
+ "s",
+ "uper"
+ ],
+ [
+ "sc",
+ "reen"
+ ],
+ [
+ "scr",
+ "een"
+ ],
+ [
+ "s",
+ "creen"
+ ],
+ [
+ "▁f",
+ "urn"
+ ],
+ [
+ "▁fur",
+ "n"
+ ],
+ [
+ "▁fu",
+ "rn"
+ ],
+ [
+ "▁surpr",
+ "ised"
+ ],
+ [
+ "▁surprise",
+ "d"
+ ],
+ [
+ "▁re",
+ "plied"
+ ],
+ [
+ "▁repl",
+ "ied"
+ ],
+ [
+ "Act",
+ "ivity"
+ ],
+ [
+ "Activ",
+ "ity"
+ ],
+ [
+ "▁D",
+ "own"
+ ],
+ [
+ "▁Do",
+ "wn"
+ ],
+ [
+ "▁Dow",
+ "n"
+ ],
+ [
+ "▁",
+ "Down"
+ ],
+ [
+ "▁in",
+ "sert"
+ ],
+ [
+ "▁ins",
+ "ert"
+ ],
+ [
+ "▁",
+ "insert"
+ ],
+ [
+ "▁O",
+ "lymp"
+ ],
+ [
+ "▁point",
+ "ed"
+ ],
+ [
+ "▁po",
+ "inted"
+ ],
+ [
+ "▁C",
+ "ard"
+ ],
+ [
+ "▁Car",
+ "d"
+ ],
+ [
+ "▁Ca",
+ "rd"
+ ],
+ [
+ "▁",
+ "Card"
+ ],
+ [
+ "dr",
+ "iver"
+ ],
+ [
+ "drive",
+ "r"
+ ],
+ [
+ "d",
+ "river"
+ ],
+ [
+ "▁D",
+ "a"
+ ],
+ [
+ "▁",
+ "Da"
+ ],
+ [
+ "!",
+ "--"
+ ],
+ [
+ "ro",
+ "ud"
+ ],
+ [
+ "rou",
+ "d"
+ ],
+ [
+ "r",
+ "oud"
+ ],
+ [
+ "un",
+ "do"
+ ],
+ [
+ "und",
+ "o"
+ ],
+ [
+ "▁m",
+ "essages"
+ ],
+ [
+ "▁message",
+ "s"
+ ],
+ [
+ "▁mess",
+ "ages"
+ ],
+ [
+ "▁",
+ "messages"
+ ],
+ [
+ "▁P",
+ "oint"
+ ],
+ [
+ "▁Po",
+ "int"
+ ],
+ [
+ "▁",
+ "Point"
+ ],
+ [
+ "V",
+ "M"
+ ],
+ [
+ "▁p",
+ "lane"
+ ],
+ [
+ "▁pl",
+ "ane"
+ ],
+ [
+ "▁plan",
+ "e"
+ ],
+ [
+ "▁",
+ "plane"
+ ],
+ [
+ "x",
+ "c"
+ ],
+ [
+ "▁telev",
+ "ision"
+ ],
+ [
+ "▁tele",
+ "vision"
+ ],
+ [
+ "▁televis",
+ "ion"
+ ],
+ [
+ "ё",
+ "н"
+ ],
+ [
+ "▁thous",
+ "ands"
+ ],
+ [
+ "▁thousand",
+ "s"
+ ],
+ [
+ "▁c",
+ "ris"
+ ],
+ [
+ "▁cr",
+ "is"
+ ],
+ [
+ "▁cri",
+ "s"
+ ],
+ [
+ "▁de",
+ "lay"
+ ],
+ [
+ "▁del",
+ "ay"
+ ],
+ [
+ "▁",
+ "delay"
+ ],
+ [
+ "▁N",
+ "ext"
+ ],
+ [
+ "▁Ne",
+ "xt"
+ ],
+ [
+ "▁",
+ "Next"
+ ],
+ [
+ "▁no",
+ "mbre"
+ ],
+ [
+ "▁nom",
+ "bre"
+ ],
+ [
+ "▁t",
+ "u"
+ ],
+ [
+ "▁",
+ "tu"
+ ],
+ [
+ "▁sk",
+ "ip"
+ ],
+ [
+ "▁ski",
+ "p"
+ ],
+ [
+ "▁",
+ "skip"
+ ],
+ [
+ "ro",
+ "ad"
+ ],
+ [
+ "r",
+ "oad"
+ ],
+ [
+ "istr",
+ "ation"
+ ],
+ [
+ "▁t",
+ "ur"
+ ],
+ [
+ "▁tu",
+ "r"
+ ],
+ [
+ "▁De",
+ "velop"
+ ],
+ [
+ "▁Devel",
+ "op"
+ ],
+ [
+ "▁П",
+ "а"
+ ],
+ [
+ "▁д",
+ "ру"
+ ],
+ [
+ "▁др",
+ "у"
+ ],
+ [
+ "▁wonder",
+ "ful"
+ ],
+ [
+ ">",
+ "&"
+ ],
+ [
+ "▁L",
+ "iber"
+ ],
+ [
+ "▁Li",
+ "ber"
+ ],
+ [
+ "▁Lib",
+ "er"
+ ],
+ [
+ "▁s",
+ "cope"
+ ],
+ [
+ "▁sc",
+ "ope"
+ ],
+ [
+ "▁",
+ "scope"
+ ],
+ [
+ "▁man",
+ "age"
+ ],
+ [
+ "▁ma",
+ "nage"
+ ],
+ [
+ "▁d",
+ "ass"
+ ],
+ [
+ "▁da",
+ "ss"
+ ],
+ [
+ "▁das",
+ "s"
+ ],
+ [
+ "▁re",
+ "call"
+ ],
+ [
+ "▁rec",
+ "all"
+ ],
+ [
+ "P",
+ "M"
+ ],
+ [
+ "▁re",
+ "levant"
+ ],
+ [
+ "▁relev",
+ "ant"
+ ],
+ [
+ "▁E",
+ "arth"
+ ],
+ [
+ "▁ка",
+ "к"
+ ],
+ [
+ "▁a",
+ "pr"
+ ],
+ [
+ "▁ap",
+ "r"
+ ],
+ [
+ "▁A",
+ "SS"
+ ],
+ [
+ "▁AS",
+ "S"
+ ],
+ [
+ "▁",
+ "ASS"
+ ],
+ [
+ "ié",
+ "n"
+ ],
+ [
+ "i",
+ "én"
+ ],
+ [
+ "▁S",
+ "H"
+ ],
+ [
+ "▁",
+ "SH"
+ ],
+ [
+ "oo",
+ "m"
+ ],
+ [
+ "o",
+ "om"
+ ],
+ [
+ "it",
+ "et"
+ ],
+ [
+ "ite",
+ "t"
+ ],
+ [
+ "no",
+ "ne"
+ ],
+ [
+ "non",
+ "e"
+ ],
+ [
+ "n",
+ "one"
+ ],
+ [
+ "as",
+ "i"
+ ],
+ [
+ "a",
+ "si"
+ ],
+ [
+ "▁mot",
+ "or"
+ ],
+ [
+ "▁mo",
+ "tor"
+ ],
+ [
+ "▁S",
+ "how"
+ ],
+ [
+ "▁Sh",
+ "ow"
+ ],
+ [
+ "▁",
+ "Show"
+ ],
+ [
+ "n",
+ "b"
+ ],
+ [
+ "▁fact",
+ "ors"
+ ],
+ [
+ "▁fa",
+ "ctors"
+ ],
+ [
+ "▁factor",
+ "s"
+ ],
+ [
+ "▁f",
+ "orest"
+ ],
+ [
+ "▁for",
+ "est"
+ ],
+ [
+ "▁fore",
+ "st"
+ ],
+ [
+ "▁fo",
+ "rest"
+ ],
+ [
+ "▁в",
+ "ре"
+ ],
+ [
+ "th",
+ "m"
+ ],
+ [
+ "t",
+ "hm"
+ ],
+ [
+ "▁m",
+ "unicip"
+ ],
+ [
+ "▁turn",
+ "s"
+ ],
+ [
+ "▁tur",
+ "ns"
+ ],
+ [
+ "▁Div",
+ "ision"
+ ],
+ [
+ "▁Di",
+ "vision"
+ ],
+ [
+ "E",
+ "C"
+ ],
+ [
+ "▁dis",
+ "appe"
+ ],
+ [
+ "struct",
+ "or"
+ ],
+ [
+ "stru",
+ "ctor"
+ ],
+ [
+ "▁some",
+ "where"
+ ],
+ [
+ "▁Afr",
+ "ican"
+ ],
+ [
+ "▁Africa",
+ "n"
+ ],
+ [
+ "▁Inst",
+ "itute"
+ ],
+ [
+ "▁Institut",
+ "e"
+ ],
+ [
+ "Gr",
+ "id"
+ ],
+ [
+ "G",
+ "rid"
+ ],
+ [
+ "▁te",
+ "acher"
+ ],
+ [
+ "▁teach",
+ "er"
+ ],
+ [
+ "▁tea",
+ "cher"
+ ],
+ [
+ "ur",
+ "ies"
+ ],
+ [
+ "uri",
+ "es"
+ ],
+ [
+ "u",
+ "ries"
+ ],
+ [
+ "▁respect",
+ "ively"
+ ],
+ [
+ "▁respective",
+ "ly"
+ ],
+ [
+ "▁S",
+ "D"
+ ],
+ [
+ "▁",
+ "SD"
+ ],
+ [
+ "▁a",
+ "live"
+ ],
+ [
+ "▁al",
+ "ive"
+ ],
+ [
+ "▁ali",
+ "ve"
+ ],
+ [
+ "▁p",
+ "ou"
+ ],
+ [
+ "▁po",
+ "u"
+ ],
+ [
+ "▁W",
+ "ater"
+ ],
+ [
+ "▁Wat",
+ "er"
+ ],
+ [
+ "▁Wa",
+ "ter"
+ ],
+ [
+ "▁",
+ "Water"
+ ],
+ [
+ "ф",
+ "е"
+ ],
+ [
+ "▁ch",
+ "anging"
+ ],
+ [
+ "▁chang",
+ "ing"
+ ],
+ [
+ "▁",
+ "changing"
+ ],
+ [
+ "▁after",
+ "noon"
+ ],
+ [
+ "▁or",
+ "ders"
+ ],
+ [
+ "▁order",
+ "s"
+ ],
+ [
+ "▁ord",
+ "ers"
+ ],
+ [
+ "▁",
+ "orders"
+ ],
+ [
+ "Re",
+ "t"
+ ],
+ [
+ "R",
+ "et"
+ ],
+ [
+ "Point",
+ "er"
+ ],
+ [
+ "Po",
+ "inter"
+ ],
+ [
+ "▁s",
+ "av"
+ ],
+ [
+ "▁sa",
+ "v"
+ ],
+ [
+ "er",
+ "g"
+ ],
+ [
+ "e",
+ "rg"
+ ],
+ [
+ "ok",
+ "ed"
+ ],
+ [
+ "oke",
+ "d"
+ ],
+ [
+ "o",
+ "ked"
+ ],
+ [
+ "ess",
+ "ions"
+ ],
+ [
+ "ession",
+ "s"
+ ],
+ [
+ "▁F",
+ "ire"
+ ],
+ [
+ "▁Fi",
+ "re"
+ ],
+ [
+ "▁",
+ "Fire"
+ ],
+ [
+ "ar",
+ "et"
+ ],
+ [
+ "are",
+ "t"
+ ],
+ [
+ "a",
+ "ret"
+ ],
+ [
+ "im",
+ "m"
+ ],
+ [
+ "i",
+ "mm"
+ ],
+ [
+ "▁des",
+ "ire"
+ ],
+ [
+ "▁",
+ "що"
+ ],
+ [
+ "▁De",
+ "sign"
+ ],
+ [
+ "▁Des",
+ "ign"
+ ],
+ [
+ "▁",
+ "Design"
+ ],
+ [
+ "ut",
+ "ure"
+ ],
+ [
+ "▁Off",
+ "ice"
+ ],
+ [
+ "▁c",
+ "md"
+ ],
+ [
+ "▁cm",
+ "d"
+ ],
+ [
+ "▁",
+ "cmd"
+ ],
+ [
+ "▁e",
+ "ating"
+ ],
+ [
+ "▁eat",
+ "ing"
+ ],
+ [
+ "Net",
+ "work"
+ ],
+ [
+ "▁r",
+ "ough"
+ ],
+ [
+ "▁ro",
+ "ugh"
+ ],
+ [
+ "▁rou",
+ "gh"
+ ],
+ [
+ "▁",
+ "rough"
+ ],
+ [
+ "oper",
+ "ator"
+ ],
+ [
+ "IG",
+ "N"
+ ],
+ [
+ "I",
+ "GN"
+ ],
+ [
+ "▁s",
+ "ports"
+ ],
+ [
+ "▁sp",
+ "orts"
+ ],
+ [
+ "▁sport",
+ "s"
+ ],
+ [
+ "▁w",
+ "eren"
+ ],
+ [
+ "▁we",
+ "ren"
+ ],
+ [
+ "▁were",
+ "n"
+ ],
+ [
+ "▁wer",
+ "en"
+ ],
+ [
+ "▁n",
+ "oted"
+ ],
+ [
+ "▁not",
+ "ed"
+ ],
+ [
+ "▁no",
+ "ted"
+ ],
+ [
+ "▁note",
+ "d"
+ ],
+ [
+ "▁tw",
+ "ice"
+ ],
+ [
+ "II",
+ "I"
+ ],
+ [
+ "I",
+ "II"
+ ],
+ [
+ "▁a",
+ "nx"
+ ],
+ [
+ "▁an",
+ "x"
+ ],
+ [
+ "▁e",
+ "lim"
+ ],
+ [
+ "▁el",
+ "im"
+ ],
+ [
+ "▁а",
+ "в"
+ ],
+ [
+ "▁i",
+ "o"
+ ],
+ [
+ "▁",
+ "io"
+ ],
+ [
+ "▁spe",
+ "ech"
+ ],
+ [
+ "▁con",
+ "du"
+ ],
+ [
+ "▁cond",
+ "u"
+ ],
+ [
+ "el",
+ "les"
+ ],
+ [
+ "ell",
+ "es"
+ ],
+ [
+ "elle",
+ "s"
+ ],
+ [
+ "id",
+ "ade"
+ ],
+ [
+ "ida",
+ "de"
+ ],
+ [
+ "idad",
+ "e"
+ ],
+ [
+ "▁adv",
+ "ance"
+ ],
+ [
+ "R",
+ "I"
+ ],
+ [
+ "oc",
+ "a"
+ ],
+ [
+ "o",
+ "ca"
+ ],
+ [
+ "/",
+ "\\"
+ ],
+ [
+ "ap",
+ "shot"
+ ],
+ [
+ "aps",
+ "hot"
+ ],
+ [
+ "▁t",
+ "ail"
+ ],
+ [
+ "▁ta",
+ "il"
+ ],
+ [
+ "▁",
+ "tail"
+ ],
+ [
+ "mod",
+ "els"
+ ],
+ [
+ "model",
+ "s"
+ ],
+ [
+ "mode",
+ "ls"
+ ],
+ [
+ "og",
+ "y"
+ ],
+ [
+ "o",
+ "gy"
+ ],
+ [
+ "▁J",
+ "eff"
+ ],
+ [
+ "▁Je",
+ "ff"
+ ],
+ [
+ "ir",
+ "ation"
+ ],
+ [
+ "irat",
+ "ion"
+ ],
+ [
+ "▁K",
+ "ore"
+ ],
+ [
+ "▁Ko",
+ "re"
+ ],
+ [
+ "▁Kor",
+ "e"
+ ],
+ [
+ "▁le",
+ "ads"
+ ],
+ [
+ "▁lead",
+ "s"
+ ],
+ [
+ "ba",
+ "t"
+ ],
+ [
+ "b",
+ "at"
+ ],
+ [
+ "Ad",
+ "apter"
+ ],
+ [
+ "c",
+ "ategory"
+ ],
+ [
+ "ang",
+ "ular"
+ ],
+ [
+ "angu",
+ "lar"
+ ],
+ [
+ "▁s",
+ "aved"
+ ],
+ [
+ "▁sa",
+ "ved"
+ ],
+ [
+ "▁save",
+ "d"
+ ],
+ [
+ "▁sav",
+ "ed"
+ ],
+ [
+ "▁",
+ "saved"
+ ],
+ [
+ "▁un",
+ "iform"
+ ],
+ [
+ "▁",
+ "uniform"
+ ],
+ [
+ "▁n",
+ "é"
+ ],
+ [
+ "▁",
+ "né"
+ ],
+ [
+ "▁business",
+ "es"
+ ],
+ [
+ "His",
+ "t"
+ ],
+ [
+ "Hi",
+ "st"
+ ],
+ [
+ "H",
+ "ist"
+ ],
+ [
+ "▁а",
+ "р"
+ ],
+ [
+ "▁",
+ "ар"
+ ],
+ [
+ "do",
+ "main"
+ ],
+ [
+ "dom",
+ "ain"
+ ],
+ [
+ "▁S",
+ "i"
+ ],
+ [
+ "▁",
+ "Si"
+ ],
+ [
+ "ra",
+ "ise"
+ ],
+ [
+ "rais",
+ "e"
+ ],
+ [
+ "rai",
+ "se"
+ ],
+ [
+ "r",
+ "aise"
+ ],
+ [
+ "▁w",
+ "arn"
+ ],
+ [
+ "▁war",
+ "n"
+ ],
+ [
+ "▁wa",
+ "rn"
+ ],
+ [
+ "▁",
+ "warn"
+ ],
+ [
+ "het",
+ "ic"
+ ],
+ [
+ "h",
+ "etic"
+ ],
+ [
+ "▁G",
+ "ro"
+ ],
+ [
+ "▁Gr",
+ "o"
+ ],
+ [
+ "))",
+ "."
+ ],
+ [
+ ")",
+ ")."
+ ],
+ [
+ "}",
+ ">"
+ ],
+ [
+ "з",
+ "е"
+ ],
+ [
+ "▁Amaz",
+ "on"
+ ],
+ [
+ "▁Or",
+ "gan"
+ ],
+ [
+ "▁",
+ "Organ"
+ ],
+ [
+ "▁L",
+ "ake"
+ ],
+ [
+ "▁La",
+ "ke"
+ ],
+ [
+ "▁ag",
+ "reement"
+ ],
+ [
+ "▁agree",
+ "ment"
+ ],
+ [
+ "▁agre",
+ "ement"
+ ],
+ [
+ "x",
+ "a"
+ ],
+ [
+ "▁p",
+ "erman"
+ ],
+ [
+ "▁per",
+ "man"
+ ],
+ [
+ "▁perm",
+ "an"
+ ],
+ [
+ "▁cont",
+ "aining"
+ ],
+ [
+ "▁contain",
+ "ing"
+ ],
+ [
+ "▁st",
+ "range"
+ ],
+ [
+ "▁str",
+ "ange"
+ ],
+ [
+ "▁strang",
+ "e"
+ ],
+ [
+ "ст",
+ "і"
+ ],
+ [
+ "с",
+ "ті"
+ ],
+ [
+ "▁st",
+ "upid"
+ ],
+ [
+ "▁spe",
+ "aking"
+ ],
+ [
+ "▁speak",
+ "ing"
+ ],
+ [
+ "▁Intern",
+ "et"
+ ],
+ [
+ "▁Inter",
+ "net"
+ ],
+ [
+ "pre",
+ "fix"
+ ],
+ [
+ "pref",
+ "ix"
+ ],
+ [
+ "p",
+ "refix"
+ ],
+ [
+ "es",
+ "c"
+ ],
+ [
+ "e",
+ "sc"
+ ],
+ [
+ "As",
+ "sert"
+ ],
+ [
+ "Ass",
+ "ert"
+ ],
+ [
+ "pro",
+ "te"
+ ],
+ [
+ "pr",
+ "ote"
+ ],
+ [
+ "prot",
+ "e"
+ ],
+ [
+ "p",
+ "rote"
+ ],
+ [
+ "▁m",
+ "anner"
+ ],
+ [
+ "▁man",
+ "ner"
+ ],
+ [
+ "▁S",
+ "z"
+ ],
+ [
+ "un",
+ "te"
+ ],
+ [
+ "unt",
+ "e"
+ ],
+ [
+ "u",
+ "nte"
+ ],
+ [
+ "io",
+ "t"
+ ],
+ [
+ "i",
+ "ot"
+ ],
+ [
+ "Pro",
+ "file"
+ ],
+ [
+ "ov",
+ "en"
+ ],
+ [
+ "ove",
+ "n"
+ ],
+ [
+ "o",
+ "ven"
+ ],
+ [
+ "▁for",
+ "med"
+ ],
+ [
+ "▁form",
+ "ed"
+ ],
+ [
+ "▁forme",
+ "d"
+ ],
+ [
+ "▁",
+ "formed"
+ ],
+ [
+ "▁l",
+ "it"
+ ],
+ [
+ "▁li",
+ "t"
+ ],
+ [
+ "▁",
+ "lit"
+ ],
+ [
+ "▁econom",
+ "y"
+ ],
+ [
+ "▁ec",
+ "onomy"
+ ],
+ [
+ "▁c",
+ "z"
+ ],
+ [
+ "▁",
+ "cz"
+ ],
+ [
+ "wi",
+ "d"
+ ],
+ [
+ "w",
+ "id"
+ ],
+ [
+ "RE",
+ "Q"
+ ],
+ [
+ "R",
+ "EQ"
+ ],
+ [
+ "▁ch",
+ "osen"
+ ],
+ [
+ "▁cho",
+ "sen"
+ ],
+ [
+ "▁chose",
+ "n"
+ ],
+ [
+ "▁P",
+ "rodu"
+ ],
+ [
+ "▁Pro",
+ "du"
+ ],
+ [
+ "▁",
+ "Produ"
+ ],
+ [
+ "os",
+ "ter"
+ ],
+ [
+ "ost",
+ "er"
+ ],
+ [
+ "o",
+ "ster"
+ ],
+ [
+ "st",
+ "ances"
+ ],
+ [
+ "stance",
+ "s"
+ ],
+ [
+ "stan",
+ "ces"
+ ],
+ [
+ "aw",
+ "a"
+ ],
+ [
+ "a",
+ "wa"
+ ],
+ [
+ "▁R",
+ "en"
+ ],
+ [
+ "▁Re",
+ "n"
+ ],
+ [
+ "▁conf",
+ "irm"
+ ],
+ [
+ "▁",
+ "confirm"
+ ],
+ [
+ "▁Б",
+ "о"
+ ],
+ [
+ "▁b",
+ "illion"
+ ],
+ [
+ "▁bill",
+ "ion"
+ ],
+ [
+ "▁d",
+ "éc"
+ ],
+ [
+ "▁dé",
+ "c"
+ ],
+ [
+ "ý",
+ "ch"
+ ],
+ [
+ "▁ill",
+ "ustr"
+ ],
+ [
+ "TI",
+ "ES"
+ ],
+ [
+ "T",
+ "IES"
+ ],
+ [
+ "▁P",
+ "ub"
+ ],
+ [
+ "▁Pu",
+ "b"
+ ],
+ [
+ "▁",
+ "Pub"
+ ],
+ [
+ "▁b",
+ "an"
+ ],
+ [
+ "▁ba",
+ "n"
+ ],
+ [
+ "▁",
+ "ban"
+ ],
+ [
+ "ad",
+ "ed"
+ ],
+ [
+ "ade",
+ "d"
+ ],
+ [
+ "a",
+ "ded"
+ ],
+ [
+ "ah",
+ "n"
+ ],
+ [
+ "a",
+ "hn"
+ ],
+ [
+ "▁C",
+ "ath"
+ ],
+ [
+ "▁Cat",
+ "h"
+ ],
+ [
+ "▁Ca",
+ "th"
+ ],
+ [
+ "no",
+ "number"
+ ],
+ [
+ "non",
+ "umber"
+ ],
+ [
+ "▁wor",
+ "st"
+ ],
+ [
+ "▁М",
+ "е"
+ ],
+ [
+ "▁sugg",
+ "ested"
+ ],
+ [
+ "▁suggest",
+ "ed"
+ ],
+ [
+ "st",
+ "ats"
+ ],
+ [
+ "stat",
+ "s"
+ ],
+ [
+ "sta",
+ "ts"
+ ],
+ [
+ "▁c",
+ "ant"
+ ],
+ [
+ "▁can",
+ "t"
+ ],
+ [
+ "▁ca",
+ "nt"
+ ],
+ [
+ "▁al",
+ "ign"
+ ],
+ [
+ "▁ali",
+ "gn"
+ ],
+ [
+ "▁",
+ "align"
+ ],
+ [
+ "kap",
+ "pa"
+ ],
+ [
+ "k",
+ "appa"
+ ],
+ [
+ "▁h",
+ "en"
+ ],
+ [
+ "▁he",
+ "n"
+ ],
+ [
+ "▁",
+ "hen"
+ ],
+ [
+ "▁in",
+ "iti"
+ ],
+ [
+ "▁init",
+ "i"
+ ],
+ [
+ "']",
+ ")"
+ ],
+ [
+ "'",
+ "])"
+ ],
+ [
+ "B",
+ "I"
+ ],
+ [
+ "▁g",
+ "arden"
+ ],
+ [
+ "▁gar",
+ "den"
+ ],
+ [
+ "▁gard",
+ "en"
+ ],
+ [
+ "▁sec",
+ "ure"
+ ],
+ [
+ "▁secur",
+ "e"
+ ],
+ [
+ "▁",
+ "secure"
+ ],
+ [
+ "▁\\",
+ "["
+ ],
+ [
+ "▁",
+ "\\["
+ ],
+ [
+ "hand",
+ "ler"
+ ],
+ [
+ "handle",
+ "r"
+ ],
+ [
+ "el",
+ "li"
+ ],
+ [
+ "ell",
+ "i"
+ ],
+ [
+ "e",
+ "lli"
+ ],
+ [
+ "ld",
+ "ots"
+ ],
+ [
+ "l",
+ "dots"
+ ],
+ [
+ "se",
+ "cut"
+ ],
+ [
+ "sec",
+ "ut"
+ ],
+ [
+ "s",
+ "ecut"
+ ],
+ [
+ "▁ext",
+ "ended"
+ ],
+ [
+ "▁extend",
+ "ed"
+ ],
+ [
+ "}",
+ "-"
+ ],
+ [
+ "an",
+ "ie"
+ ],
+ [
+ "ani",
+ "e"
+ ],
+ [
+ "a",
+ "nie"
+ ],
+ [
+ "▁F",
+ "ind"
+ ],
+ [
+ "▁Fin",
+ "d"
+ ],
+ [
+ "▁Fi",
+ "nd"
+ ],
+ [
+ "▁",
+ "Find"
+ ],
+ [
+ "▁M",
+ "useum"
+ ],
+ [
+ "▁Muse",
+ "um"
+ ],
+ [
+ "▁C",
+ "onne"
+ ],
+ [
+ "▁Con",
+ "ne"
+ ],
+ [
+ "▁",
+ "Conne"
+ ],
+ [
+ "y",
+ "y"
+ ],
+ [
+ "▁pass",
+ "ion"
+ ],
+ [
+ "ak",
+ "ers"
+ ],
+ [
+ "ake",
+ "rs"
+ ],
+ [
+ "aker",
+ "s"
+ ],
+ [
+ "a",
+ "kers"
+ ],
+ [
+ "ah",
+ "r"
+ ],
+ [
+ "a",
+ "hr"
+ ],
+ [
+ "olog",
+ "ies"
+ ],
+ [
+ "ologie",
+ "s"
+ ],
+ [
+ "▁equ",
+ "ation"
+ ],
+ [
+ "▁eq",
+ "uation"
+ ],
+ [
+ "▁",
+ "equation"
+ ],
+ [
+ "▁occ",
+ "asion"
+ ],
+ [
+ "▁occas",
+ "ion"
+ ],
+ [
+ "Le",
+ "t"
+ ],
+ [
+ "L",
+ "et"
+ ],
+ [
+ "']",
+ "['"
+ ],
+ [
+ "'][",
+ "'"
+ ],
+ [
+ "'",
+ "]['"
+ ],
+ [
+ "Pr",
+ "int"
+ ],
+ [
+ "an",
+ "es"
+ ],
+ [
+ "ane",
+ "s"
+ ],
+ [
+ "a",
+ "nes"
+ ],
+ [
+ "ie",
+ "nte"
+ ],
+ [
+ "ient",
+ "e"
+ ],
+ [
+ "ien",
+ "te"
+ ],
+ [
+ "i",
+ "ente"
+ ],
+ [
+ "▁T",
+ "oday"
+ ],
+ [
+ "▁To",
+ "day"
+ ],
+ [
+ "▁Tod",
+ "ay"
+ ],
+ [
+ "LE",
+ "CT"
+ ],
+ [
+ "L",
+ "ECT"
+ ],
+ [
+ "▁A",
+ "f"
+ ],
+ [
+ "▁",
+ "Af"
+ ],
+ [
+ ",",
+ ","
+ ],
+ [
+ "▁Т",
+ "а"
+ ],
+ [
+ "▁`",
+ "``"
+ ],
+ [
+ "▁``",
+ "`"
+ ],
+ [
+ "ev",
+ "en"
+ ],
+ [
+ "eve",
+ "n"
+ ],
+ [
+ "e",
+ "ven"
+ ],
+ [
+ "si",
+ "n"
+ ],
+ [
+ "s",
+ "in"
+ ],
+ [
+ "ur",
+ "er"
+ ],
+ [
+ "ure",
+ "r"
+ ],
+ [
+ "u",
+ "rer"
+ ],
+ [
+ "▁",
+ "°"
+ ],
+ [
+ "ot",
+ "imes"
+ ],
+ [
+ "oti",
+ "mes"
+ ],
+ [
+ "o",
+ "times"
+ ],
+ [
+ "▁I",
+ "O"
+ ],
+ [
+ "▁",
+ "IO"
+ ],
+ [
+ "▁po",
+ "et"
+ ],
+ [
+ "()",
+ "));"
+ ],
+ [
+ "())",
+ ");"
+ ],
+ [
+ "()))",
+ ";"
+ ],
+ [
+ "(",
+ ")));"
+ ],
+ [
+ "▁",
+ "−"
+ ],
+ [
+ "▁ad",
+ "opt"
+ ],
+ [
+ "ph",
+ "ere"
+ ],
+ [
+ "pher",
+ "e"
+ ],
+ [
+ "p",
+ "here"
+ ],
+ [
+ "#",
+ "["
+ ],
+ [
+ "▁c",
+ "entre"
+ ],
+ [
+ "▁cent",
+ "re"
+ ],
+ [
+ "ov",
+ "es"
+ ],
+ [
+ "ove",
+ "s"
+ ],
+ [
+ "o",
+ "ves"
+ ],
+ [
+ "▁a",
+ "ns"
+ ],
+ [
+ "▁an",
+ "s"
+ ],
+ [
+ "▁",
+ "ans"
+ ],
+ [
+ "d",
+ "p"
+ ],
+ [
+ "▁K",
+ "ir"
+ ],
+ [
+ "▁Ki",
+ "r"
+ ],
+ [
+ "▁applic",
+ "able"
+ ],
+ [
+ "f",
+ "p"
+ ],
+ [
+ "▁vis",
+ "ual"
+ ],
+ [
+ "▁ok",
+ "ay"
+ ],
+ [
+ "or",
+ "o"
+ ],
+ [
+ "o",
+ "ro"
+ ],
+ [
+ "▁opportun",
+ "ities"
+ ],
+ [
+ "Re",
+ "pository"
+ ],
+ [
+ "Rep",
+ "ository"
+ ],
+ [
+ "▁l",
+ "l"
+ ],
+ [
+ "▁",
+ "ll"
+ ],
+ [
+ "▁R",
+ "od"
+ ],
+ [
+ "▁Ro",
+ "d"
+ ],
+ [
+ "▁s",
+ "hel"
+ ],
+ [
+ "▁sh",
+ "el"
+ ],
+ [
+ "▁she",
+ "l"
+ ],
+ [
+ "▁la",
+ "unch"
+ ],
+ [
+ "▁con",
+ "ven"
+ ],
+ [
+ "▁conv",
+ "en"
+ ],
+ [
+ "▁conve",
+ "n"
+ ],
+ [
+ "▁S",
+ "pe"
+ ],
+ [
+ "▁Sp",
+ "e"
+ ],
+ [
+ "▁",
+ "Spe"
+ ],
+ [
+ "Am",
+ "er"
+ ],
+ [
+ "A",
+ "mer"
+ ],
+ [
+ "▁c",
+ "ette"
+ ],
+ [
+ "▁cet",
+ "te"
+ ],
+ [
+ "Con",
+ "d"
+ ],
+ [
+ "Co",
+ "nd"
+ ],
+ [
+ "C",
+ "ond"
+ ],
+ [
+ "de",
+ "p"
+ ],
+ [
+ "d",
+ "ep"
+ ],
+ [
+ "O",
+ "wn"
+ ],
+ [
+ "▁h",
+ "ook"
+ ],
+ [
+ "▁ho",
+ "ok"
+ ],
+ [
+ "▁",
+ "hook"
+ ],
+ [
+ "▁d",
+ "ict"
+ ],
+ [
+ "▁di",
+ "ct"
+ ],
+ [
+ "▁dic",
+ "t"
+ ],
+ [
+ "▁",
+ "dict"
+ ],
+ [
+ "▁Th",
+ "ose"
+ ],
+ [
+ "▁f",
+ "ellow"
+ ],
+ [
+ "▁fell",
+ "ow"
+ ],
+ [
+ "▁fel",
+ "low"
+ ],
+ [
+ "▁phil",
+ "osoph"
+ ],
+ [
+ "▁philos",
+ "oph"
+ ],
+ [
+ "vi",
+ "n"
+ ],
+ [
+ "v",
+ "in"
+ ],
+ [
+ "fer",
+ "ences"
+ ],
+ [
+ "ference",
+ "s"
+ ],
+ [
+ "ha",
+ "v"
+ ],
+ [
+ "h",
+ "av"
+ ],
+ [
+ "▁ad",
+ "ding"
+ ],
+ [
+ "▁add",
+ "ing"
+ ],
+ [
+ "▁",
+ "adding"
+ ],
+ [
+ "ivers",
+ "e"
+ ],
+ [
+ "iver",
+ "se"
+ ],
+ [
+ "i",
+ "verse"
+ ],
+ [
+ "ga",
+ "me"
+ ],
+ [
+ "g",
+ "ame"
+ ],
+ [
+ "▁Bl",
+ "ue"
+ ],
+ [
+ "▁",
+ "Blue"
+ ],
+ [
+ "▁c",
+ "lin"
+ ],
+ [
+ "▁cl",
+ "in"
+ ],
+ [
+ "not",
+ "e"
+ ],
+ [
+ "no",
+ "te"
+ ],
+ [
+ "n",
+ "ote"
+ ],
+ [
+ "▁R",
+ "am"
+ ],
+ [
+ "▁Ra",
+ "m"
+ ],
+ [
+ "ме",
+ "р"
+ ],
+ [
+ "м",
+ "ер"
+ ],
+ [
+ "co",
+ "very"
+ ],
+ [
+ "cover",
+ "y"
+ ],
+ [
+ "cov",
+ "ery"
+ ],
+ [
+ "c",
+ "overy"
+ ],
+ [
+ "ñ",
+ "a"
+ ],
+ [
+ "▁б",
+ "и"
+ ],
+ [
+ "▁",
+ "би"
+ ],
+ [
+ "▁f",
+ "ashion"
+ ],
+ [
+ "▁b",
+ "roke"
+ ],
+ [
+ "▁br",
+ "oke"
+ ],
+ [
+ "▁bro",
+ "ke"
+ ],
+ [
+ "▁'",
+ "\\"
+ ],
+ [
+ "▁",
+ "'\\"
+ ],
+ [
+ "▁re",
+ "ader"
+ ],
+ [
+ "▁read",
+ "er"
+ ],
+ [
+ "▁",
+ "reader"
+ ],
+ [
+ "но",
+ "е"
+ ],
+ [
+ "но",
+ "сти"
+ ],
+ [
+ "ност",
+ "и"
+ ],
+ [
+ "▁pay",
+ "ment"
+ ],
+ [
+ "▁",
+ "payment"
+ ],
+ [
+ "▁L",
+ "ic"
+ ],
+ [
+ "▁Li",
+ "c"
+ ],
+ [
+ "▁l",
+ "ips"
+ ],
+ [
+ "▁li",
+ "ps"
+ ],
+ [
+ "▁lip",
+ "s"
+ ],
+ [
+ "▁ac",
+ "adem"
+ ],
+ [
+ "▁M",
+ "ot"
+ ],
+ [
+ "▁Mo",
+ "t"
+ ],
+ [
+ "el",
+ "ls"
+ ],
+ [
+ "ell",
+ "s"
+ ],
+ [
+ "C",
+ "HECK"
+ ],
+ [
+ "▁р",
+ "у"
+ ],
+ [
+ "▁",
+ "ру"
+ ],
+ [
+ "▁M",
+ "S"
+ ],
+ [
+ "▁",
+ "MS"
+ ],
+ [
+ "Ed",
+ "itor"
+ ],
+ [
+ "Edit",
+ "or"
+ ],
+ [
+ "▁z",
+ "one"
+ ],
+ [
+ "▁zo",
+ "ne"
+ ],
+ [
+ "▁",
+ "zone"
+ ],
+ [
+ "it",
+ "ure"
+ ],
+ [
+ "itu",
+ "re"
+ ],
+ [
+ "▁I",
+ "T"
+ ],
+ [
+ "▁",
+ "IT"
+ ],
+ [
+ "run",
+ "time"
+ ],
+ [
+ "▁pro",
+ "ceed"
+ ],
+ [
+ "▁proc",
+ "eed"
+ ],
+ [
+ "ло",
+ "в"
+ ],
+ [
+ "л",
+ "ов"
+ ],
+ [
+ "▁M",
+ "aria"
+ ],
+ [
+ "▁Mar",
+ "ia"
+ ],
+ [
+ "▁Ma",
+ "ria"
+ ],
+ [
+ "ol",
+ "ver"
+ ],
+ [
+ "olve",
+ "r"
+ ],
+ [
+ "olv",
+ "er"
+ ],
+ [
+ "▁Th",
+ "anks"
+ ],
+ [
+ "▁Thank",
+ "s"
+ ],
+ [
+ "▁",
+ "Thanks"
+ ],
+ [
+ "▁should",
+ "n"
+ ],
+ [
+ "▁J",
+ "oh"
+ ],
+ [
+ "▁Jo",
+ "h"
+ ],
+ [
+ "▁Mod",
+ "el"
+ ],
+ [
+ "▁Mo",
+ "del"
+ ],
+ [
+ "▁Mode",
+ "l"
+ ],
+ [
+ "▁",
+ "Model"
+ ],
+ [
+ "▁S",
+ "ov"
+ ],
+ [
+ "▁So",
+ "v"
+ ],
+ [
+ "!",
+ "'"
+ ],
+ [
+ "D",
+ "i"
+ ],
+ [
+ "▁c",
+ "ancer"
+ ],
+ [
+ "▁can",
+ "cer"
+ ],
+ [
+ "Id",
+ "ent"
+ ],
+ [
+ "▁ex",
+ "change"
+ ],
+ [
+ "il",
+ "ler"
+ ],
+ [
+ "ill",
+ "er"
+ ],
+ [
+ "ille",
+ "r"
+ ],
+ [
+ "in",
+ "f"
+ ],
+ [
+ "i",
+ "nf"
+ ],
+ [
+ "LE",
+ "N"
+ ],
+ [
+ "L",
+ "EN"
+ ],
+ [
+ "()",
+ "{"
+ ],
+ [
+ "(",
+ "){"
+ ],
+ [
+ "ag",
+ "a"
+ ],
+ [
+ "a",
+ "ga"
+ ],
+ [
+ "\"]",
+ ","
+ ],
+ [
+ "\"",
+ "],"
+ ],
+ [
+ "u",
+ "h"
+ ],
+ [
+ "▁K",
+ "en"
+ ],
+ [
+ "▁Ke",
+ "n"
+ ],
+ [
+ "▁ph",
+ "otos"
+ ],
+ [
+ "▁phot",
+ "os"
+ ],
+ [
+ "▁photo",
+ "s"
+ ],
+ [
+ "▁t",
+ "iny"
+ ],
+ [
+ "▁ti",
+ "ny"
+ ],
+ [
+ "▁tin",
+ "y"
+ ],
+ [
+ "▁",
+ "tiny"
+ ],
+ [
+ "▁g",
+ "ent"
+ ],
+ [
+ "▁gen",
+ "t"
+ ],
+ [
+ "▁ge",
+ "nt"
+ ],
+ [
+ "▁",
+ "gent"
+ ],
+ [
+ "ü",
+ "l"
+ ],
+ [
+ "▁T",
+ "ake"
+ ],
+ [
+ "▁Ta",
+ "ke"
+ ],
+ [
+ "▁Tak",
+ "e"
+ ],
+ [
+ "▁",
+ "Take"
+ ],
+ [
+ "id",
+ "el"
+ ],
+ [
+ "ide",
+ "l"
+ ],
+ [
+ "i",
+ "del"
+ ],
+ [
+ "ou",
+ "ting"
+ ],
+ [
+ "out",
+ "ing"
+ ],
+ [
+ "In",
+ "ternal"
+ ],
+ [
+ "Inter",
+ "nal"
+ ],
+ [
+ "Intern",
+ "al"
+ ],
+ [
+ "▁c",
+ "ells"
+ ],
+ [
+ "▁cell",
+ "s"
+ ],
+ [
+ "▁cel",
+ "ls"
+ ],
+ [
+ "ни",
+ "м"
+ ],
+ [
+ "н",
+ "им"
+ ],
+ [
+ "ha",
+ "rd"
+ ],
+ [
+ "har",
+ "d"
+ ],
+ [
+ "h",
+ "ard"
+ ],
+ [
+ "▁T",
+ "own"
+ ],
+ [
+ "▁To",
+ "wn"
+ ],
+ [
+ "▁Tow",
+ "n"
+ ],
+ [
+ "ob",
+ "e"
+ ],
+ [
+ "o",
+ "be"
+ ],
+ [
+ "pl",
+ "ex"
+ ],
+ [
+ "ple",
+ "x"
+ ],
+ [
+ "p",
+ "lex"
+ ],
+ [
+ "те",
+ "р"
+ ],
+ [
+ "т",
+ "ер"
+ ],
+ [
+ "to",
+ "ns"
+ ],
+ [
+ "ton",
+ "s"
+ ],
+ [
+ "t",
+ "ons"
+ ],
+ [
+ "▁conc",
+ "entr"
+ ],
+ [
+ "▁concent",
+ "r"
+ ],
+ [
+ "mo",
+ "ck"
+ ],
+ [
+ "m",
+ "ock"
+ ],
+ [
+ "v",
+ "c"
+ ],
+ [
+ "á",
+ "z"
+ ],
+ [
+ "▁Ch",
+ "ampionship"
+ ],
+ [
+ "▁Champion",
+ "ship"
+ ],
+ [
+ "▁Champions",
+ "hip"
+ ],
+ [
+ "▁б",
+ "е"
+ ],
+ [
+ "▁",
+ "бе"
+ ],
+ [
+ "?",
+ "?"
+ ],
+ [
+ "ér",
+ "i"
+ ],
+ [
+ "é",
+ "ri"
+ ],
+ [
+ "al",
+ "y"
+ ],
+ [
+ "a",
+ "ly"
+ ],
+ [
+ "▁",
+ "Ц"
+ ],
+ [
+ "ier",
+ "te"
+ ],
+ [
+ "iert",
+ "e"
+ ],
+ [
+ "▁tot",
+ "ally"
+ ],
+ [
+ "▁total",
+ "ly"
+ ],
+ [
+ "▁A",
+ "uf"
+ ],
+ [
+ "▁Au",
+ "f"
+ ],
+ [
+ "▁our",
+ "selves"
+ ],
+ [
+ "▁S",
+ "elf"
+ ],
+ [
+ "▁Sel",
+ "f"
+ ],
+ [
+ "▁",
+ "Self"
+ ],
+ [
+ "Form",
+ "s"
+ ],
+ [
+ "For",
+ "ms"
+ ],
+ [
+ "ight",
+ "er"
+ ],
+ [
+ "igh",
+ "ter"
+ ],
+ [
+ "▁is",
+ "land"
+ ],
+ [
+ "fm",
+ "t"
+ ],
+ [
+ "f",
+ "mt"
+ ],
+ [
+ "▁r",
+ "c"
+ ],
+ [
+ "▁",
+ "rc"
+ ],
+ [
+ "▁t",
+ "ells"
+ ],
+ [
+ "▁tell",
+ "s"
+ ],
+ [
+ "▁tel",
+ "ls"
+ ],
+ [
+ "B",
+ "B"
+ ],
+ [
+ "di",
+ "t"
+ ],
+ [
+ "d",
+ "it"
+ ],
+ [
+ "▁vari",
+ "ables"
+ ],
+ [
+ "▁variable",
+ "s"
+ ],
+ [
+ "▁",
+ "variables"
+ ],
+ [
+ "▁int",
+ "ended"
+ ],
+ [
+ "▁intend",
+ "ed"
+ ],
+ [
+ "iz",
+ "ont"
+ ],
+ [
+ "izon",
+ "t"
+ ],
+ [
+ "izo",
+ "nt"
+ ],
+ [
+ "▁pl",
+ "ays"
+ ],
+ [
+ "▁play",
+ "s"
+ ],
+ [
+ "da",
+ "m"
+ ],
+ [
+ "d",
+ "am"
+ ],
+ [
+ "se",
+ "q"
+ ],
+ [
+ "s",
+ "eq"
+ ],
+ [
+ "▁S",
+ "up"
+ ],
+ [
+ "▁Su",
+ "p"
+ ],
+ [
+ "▁",
+ "Sup"
+ ],
+ [
+ "▁c",
+ "ultural"
+ ],
+ [
+ "▁cult",
+ "ural"
+ ],
+ [
+ "▁sc",
+ "ream"
+ ],
+ [
+ "__",
+ ","
+ ],
+ [
+ "_",
+ "_,"
+ ],
+ [
+ "ci",
+ "pl"
+ ],
+ [
+ "cip",
+ "l"
+ ],
+ [
+ "Time",
+ "out"
+ ],
+ [
+ "▁",
+ "ж"
+ ],
+ [
+ "or",
+ "te"
+ ],
+ [
+ "ort",
+ "e"
+ ],
+ [
+ "▁repl",
+ "aced"
+ ],
+ [
+ "▁replace",
+ "d"
+ ],
+ [
+ "E",
+ "M"
+ ],
+ [
+ "▁ab",
+ "andon"
+ ],
+ [
+ "▁Spec",
+ "ial"
+ ],
+ [
+ "▁Spe",
+ "cial"
+ ],
+ [
+ "▁",
+ "Special"
+ ],
+ [
+ "el",
+ "len"
+ ],
+ [
+ "ell",
+ "en"
+ ],
+ [
+ "elle",
+ "n"
+ ],
+ [
+ "▁B",
+ "ru"
+ ],
+ [
+ "▁Br",
+ "u"
+ ],
+ [
+ "ir",
+ "med"
+ ],
+ [
+ "irm",
+ "ed"
+ ],
+ [
+ "T",
+ "e"
+ ],
+ [
+ "ol",
+ "t"
+ ],
+ [
+ "o",
+ "lt"
+ ],
+ [
+ "j",
+ "u"
+ ],
+ [
+ "Arg",
+ "ument"
+ ],
+ [
+ "▁ne",
+ "ut"
+ ],
+ [
+ "▁neu",
+ "t"
+ ],
+ [
+ "▁",
+ "neut"
+ ],
+ [
+ "sc",
+ "ape"
+ ],
+ [
+ "▁R",
+ "ay"
+ ],
+ [
+ "▁Ra",
+ "y"
+ ],
+ [
+ "▁",
+ "Ray"
+ ],
+ [
+ "▁Pol",
+ "it"
+ ],
+ [
+ "▁Po",
+ "lit"
+ ],
+ [
+ "▁crow",
+ "d"
+ ],
+ [
+ "▁cro",
+ "wd"
+ ],
+ [
+ "▁Window",
+ "s"
+ ],
+ [
+ "▁Wind",
+ "ows"
+ ],
+ [
+ "▁",
+ "Windows"
+ ],
+ [
+ "ie",
+ "go"
+ ],
+ [
+ "ieg",
+ "o"
+ ],
+ [
+ "i",
+ "ego"
+ ],
+ [
+ "▁e",
+ "scape"
+ ],
+ [
+ "▁esc",
+ "ape"
+ ],
+ [
+ "▁",
+ "escape"
+ ],
+ [
+ "▁Ap",
+ "ache"
+ ],
+ [
+ "sy",
+ "nc"
+ ],
+ [
+ "syn",
+ "c"
+ ],
+ [
+ "s",
+ "ync"
+ ],
+ [
+ "eb",
+ "en"
+ ],
+ [
+ "e",
+ "ben"
+ ],
+ [
+ "if",
+ "ies"
+ ],
+ [
+ "ifi",
+ "es"
+ ],
+ [
+ "et",
+ "her"
+ ],
+ [
+ "eth",
+ "er"
+ ],
+ [
+ "ethe",
+ "r"
+ ],
+ [
+ "e",
+ "ther"
+ ],
+ [
+ "Met",
+ "a"
+ ],
+ [
+ "Me",
+ "ta"
+ ],
+ [
+ "M",
+ "eta"
+ ],
+ [
+ "▁big",
+ "gest"
+ ],
+ [
+ "Ga",
+ "me"
+ ],
+ [
+ "G",
+ "ame"
+ ],
+ [
+ "▁trans",
+ "action"
+ ],
+ [
+ "▁",
+ "transaction"
+ ],
+ [
+ "En",
+ "v"
+ ],
+ [
+ "E",
+ "nv"
+ ],
+ [
+ "▁М",
+ "о"
+ ],
+ [
+ "▁pl",
+ "enty"
+ ],
+ [
+ "▁m",
+ "el"
+ ],
+ [
+ "▁me",
+ "l"
+ ],
+ [
+ "▁",
+ "mel"
+ ],
+ [
+ "п",
+ "ре"
+ ],
+ [
+ "▁mot",
+ "iv"
+ ],
+ [
+ "▁о",
+ "р"
+ ],
+ [
+ "▁",
+ "ор"
+ ],
+ [
+ "or",
+ "gan"
+ ],
+ [
+ "org",
+ "an"
+ ],
+ [
+ "▁m",
+ "ock"
+ ],
+ [
+ "▁mo",
+ "ck"
+ ],
+ [
+ "▁",
+ "mock"
+ ],
+ [
+ "▁$",
+ "_"
+ ],
+ [
+ "▁",
+ "$_"
+ ],
+ [
+ "ен",
+ "е"
+ ],
+ [
+ "е",
+ "не"
+ ],
+ [
+ "▁N",
+ "umber"
+ ],
+ [
+ "▁Num",
+ "ber"
+ ],
+ [
+ "▁Nu",
+ "mber"
+ ],
+ [
+ "▁",
+ "Number"
+ ],
+ [
+ "ck",
+ "now"
+ ],
+ [
+ "c",
+ "know"
+ ],
+ [
+ "▁Up",
+ "date"
+ ],
+ [
+ "▁",
+ "Update"
+ ],
+ [
+ "ze",
+ "ro"
+ ],
+ [
+ "zer",
+ "o"
+ ],
+ [
+ "z",
+ "ero"
+ ],
+ [
+ "▁sur",
+ "prise"
+ ],
+ [
+ "▁surpr",
+ "ise"
+ ],
+ [
+ "ce",
+ "an"
+ ],
+ [
+ "pd",
+ "f"
+ ],
+ [
+ "p",
+ "df"
+ ],
+ [
+ "Gl",
+ "obal"
+ ],
+ [
+ "▁att",
+ "end"
+ ],
+ [
+ "▁f",
+ "ond"
+ ],
+ [
+ "▁fo",
+ "nd"
+ ],
+ [
+ "▁fon",
+ "d"
+ ],
+ [
+ "▁under",
+ "stood"
+ ],
+ [
+ "Na",
+ "v"
+ ],
+ [
+ "N",
+ "av"
+ ],
+ [
+ "▁M",
+ "ic"
+ ],
+ [
+ "▁Mi",
+ "c"
+ ],
+ [
+ "▁",
+ "Mic"
+ ],
+ [
+ "=",
+ "$"
+ ],
+ [
+ "ok",
+ "ing"
+ ],
+ [
+ "oki",
+ "ng"
+ ],
+ [
+ "o",
+ "king"
+ ],
+ [
+ "▁Stad",
+ "ium"
+ ],
+ [
+ "Cl",
+ "ose"
+ ],
+ [
+ "▁compet",
+ "ition"
+ ],
+ [
+ "▁sold",
+ "iers"
+ ],
+ [
+ "▁soldier",
+ "s"
+ ],
+ [
+ "▁O",
+ "P"
+ ],
+ [
+ "▁",
+ "OP"
+ ],
+ [
+ "ag",
+ "ne"
+ ],
+ [
+ "agn",
+ "e"
+ ],
+ [
+ "▁An",
+ "ton"
+ ],
+ [
+ "▁Ant",
+ "on"
+ ],
+ [
+ "Ma",
+ "in"
+ ],
+ [
+ "M",
+ "ain"
+ ],
+ [
+ "á",
+ "k"
+ ],
+ [
+ "▁#",
+ "["
+ ],
+ [
+ "▁",
+ "#["
+ ],
+ [
+ "▁Com",
+ "mit"
+ ],
+ [
+ "▁Comm",
+ "it"
+ ],
+ [
+ "▁",
+ "Commit"
+ ],
+ [
+ "py",
+ "x"
+ ],
+ [
+ "▁e",
+ "ast"
+ ],
+ [
+ "▁eas",
+ "t"
+ ],
+ [
+ "▁",
+ "east"
+ ],
+ [
+ "▁Or",
+ "der"
+ ],
+ [
+ "▁Ord",
+ "er"
+ ],
+ [
+ "▁",
+ "Order"
+ ],
+ [
+ "F",
+ "loat"
+ ],
+ [
+ "▁accept",
+ "ed"
+ ],
+ [
+ "▁mon",
+ "itor"
+ ],
+ [
+ "▁",
+ "monitor"
+ ],
+ [
+ "▁p",
+ "ad"
+ ],
+ [
+ "▁pa",
+ "d"
+ ],
+ [
+ "▁",
+ "pad"
+ ],
+ [
+ "on",
+ "ic"
+ ],
+ [
+ "oni",
+ "c"
+ ],
+ [
+ "o",
+ "nic"
+ ],
+ [
+ "▁p",
+ "ushed"
+ ],
+ [
+ "▁push",
+ "ed"
+ ],
+ [
+ "▁re",
+ "place"
+ ],
+ [
+ "▁rep",
+ "lace"
+ ],
+ [
+ "▁repl",
+ "ace"
+ ],
+ [
+ "▁",
+ "replace"
+ ],
+ [
+ "CR",
+ "E"
+ ],
+ [
+ "C",
+ "RE"
+ ],
+ [
+ "▁r",
+ "ide"
+ ],
+ [
+ "▁ri",
+ "de"
+ ],
+ [
+ "▁rid",
+ "e"
+ ],
+ [
+ "▁",
+ "ride"
+ ],
+ [
+ "fo",
+ "und"
+ ],
+ [
+ "f",
+ "ound"
+ ],
+ [
+ "=",
+ "%"
+ ],
+ [
+ "во",
+ "й"
+ ],
+ [
+ "▁mat",
+ "ches"
+ ],
+ [
+ "▁match",
+ "es"
+ ],
+ [
+ "▁",
+ "matches"
+ ],
+ [
+ "▁L",
+ "ie"
+ ],
+ [
+ "▁Li",
+ "e"
+ ],
+ [
+ "▁exper",
+ "iences"
+ ],
+ [
+ "▁experience",
+ "s"
+ ],
+ [
+ "▁experi",
+ "ences"
+ ],
+ [
+ "Po",
+ "ol"
+ ],
+ [
+ "P",
+ "ool"
+ ],
+ [
+ "up",
+ "s"
+ ],
+ [
+ "u",
+ "ps"
+ ],
+ [
+ "A",
+ "V"
+ ],
+ [
+ "▁ex",
+ "istence"
+ ],
+ [
+ "▁exist",
+ "ence"
+ ],
+ [
+ "▁t",
+ "hin"
+ ],
+ [
+ "▁th",
+ "in"
+ ],
+ [
+ "▁m",
+ "agn"
+ ],
+ [
+ "▁mag",
+ "n"
+ ],
+ [
+ "▁ma",
+ "gn"
+ ],
+ [
+ "CO",
+ "MP"
+ ],
+ [
+ "COM",
+ "P"
+ ],
+ [
+ "ho",
+ "me"
+ ],
+ [
+ "hom",
+ "e"
+ ],
+ [
+ "h",
+ "ome"
+ ],
+ [
+ "▁n",
+ "i"
+ ],
+ [
+ "▁",
+ "ni"
+ ],
+ [
+ "▁wur",
+ "den"
+ ],
+ [
+ "▁wurde",
+ "n"
+ ],
+ [
+ "ла",
+ "в"
+ ],
+ [
+ "▁te",
+ "eth"
+ ],
+ [
+ "▁S",
+ "tan"
+ ],
+ [
+ "▁St",
+ "an"
+ ],
+ [
+ "▁Sta",
+ "n"
+ ],
+ [
+ "ap",
+ "pro"
+ ],
+ [
+ "app",
+ "ro"
+ ],
+ [
+ "an",
+ "ny"
+ ],
+ [
+ "ann",
+ "y"
+ ],
+ [
+ "if",
+ "ts"
+ ],
+ [
+ "ift",
+ "s"
+ ],
+ [
+ "▁un",
+ "known"
+ ],
+ [
+ "▁",
+ "unknown"
+ ],
+ [
+ "▁h",
+ "omes"
+ ],
+ [
+ "▁home",
+ "s"
+ ],
+ [
+ "▁hom",
+ "es"
+ ],
+ [
+ "▁ho",
+ "mes"
+ ],
+ [
+ "▁ent",
+ "ity"
+ ],
+ [
+ "▁",
+ "entity"
+ ],
+ [
+ "ci",
+ "e"
+ ],
+ [
+ "c",
+ "ie"
+ ],
+ [
+ "ле",
+ "ние"
+ ],
+ [
+ "ia",
+ "r"
+ ],
+ [
+ "i",
+ "ar"
+ ],
+ [
+ "▁compl",
+ "iance"
+ ],
+ [
+ "▁focus",
+ "ed"
+ ],
+ [
+ "uz",
+ "z"
+ ],
+ [
+ "u",
+ "zz"
+ ],
+ [
+ "=\\",
+ "\""
+ ],
+ [
+ "=",
+ "\\\""
+ ],
+ [
+ "com",
+ "ponents"
+ ],
+ [
+ "component",
+ "s"
+ ],
+ [
+ "Att",
+ "r"
+ ],
+ [
+ "At",
+ "tr"
+ ],
+ [
+ "all",
+ "ery"
+ ],
+ [
+ "alle",
+ "ry"
+ ],
+ [
+ "aller",
+ "y"
+ ],
+ [
+ "▁ident",
+ "ify"
+ ],
+ [
+ "O",
+ "k"
+ ],
+ [
+ "pi",
+ "e"
+ ],
+ [
+ "p",
+ "ie"
+ ],
+ [
+ "▁St",
+ "ill"
+ ],
+ [
+ "▁off",
+ "ering"
+ ],
+ [
+ "▁offer",
+ "ing"
+ ],
+ [
+ "▁bu",
+ "sy"
+ ],
+ [
+ "▁bus",
+ "y"
+ ],
+ [
+ "ct",
+ "l"
+ ],
+ [
+ "c",
+ "tl"
+ ],
+ [
+ "it",
+ "ors"
+ ],
+ [
+ "itor",
+ "s"
+ ],
+ [
+ "ito",
+ "rs"
+ ],
+ [
+ "▁concern",
+ "ed"
+ ],
+ [
+ "▁concer",
+ "ned"
+ ],
+ [
+ "▁b",
+ "rown"
+ ],
+ [
+ "▁br",
+ "own"
+ ],
+ [
+ "▁bro",
+ "wn"
+ ],
+ [
+ "▁brow",
+ "n"
+ ],
+ [
+ "cl",
+ "k"
+ ],
+ [
+ "Se",
+ "lected"
+ ],
+ [
+ "Select",
+ "ed"
+ ],
+ [
+ "▁B",
+ "lock"
+ ],
+ [
+ "▁Bl",
+ "ock"
+ ],
+ [
+ "▁Blo",
+ "ck"
+ ],
+ [
+ "▁",
+ "Block"
+ ],
+ [
+ "▁e",
+ "gy"
+ ],
+ [
+ "▁eg",
+ "y"
+ ],
+ [
+ "▁",
+ "egy"
+ ],
+ [
+ "ic",
+ "ing"
+ ],
+ [
+ "ici",
+ "ng"
+ ],
+ [
+ "i",
+ "cing"
+ ],
+ [
+ "▁U",
+ "RL"
+ ],
+ [
+ "▁",
+ "URL"
+ ],
+ [
+ "▁t",
+ "opic"
+ ],
+ [
+ "▁to",
+ "pic"
+ ],
+ [
+ "▁top",
+ "ic"
+ ],
+ [
+ "▁",
+ "topic"
+ ],
+ [
+ "▁Pro",
+ "duct"
+ ],
+ [
+ "▁Produ",
+ "ct"
+ ],
+ [
+ "▁",
+ "Product"
+ ],
+ [
+ "▁ч",
+ "и"
+ ],
+ [
+ "▁",
+ "чи"
+ ],
+ [
+ "▁t",
+ "rial"
+ ],
+ [
+ "▁tr",
+ "ial"
+ ],
+ [
+ "▁tri",
+ "al"
+ ],
+ [
+ "▁week",
+ "end"
+ ],
+ [
+ "l",
+ "u"
+ ],
+ [
+ "▁I",
+ "V"
+ ],
+ [
+ "▁",
+ "IV"
+ ],
+ [
+ "▁E",
+ "gy"
+ ],
+ [
+ "▁Eg",
+ "y"
+ ],
+ [
+ "x",
+ "C"
+ ],
+ [
+ "▁n",
+ "ove"
+ ],
+ [
+ "▁no",
+ "ve"
+ ],
+ [
+ "▁nov",
+ "e"
+ ],
+ [
+ "▁l",
+ "ett"
+ ],
+ [
+ "▁le",
+ "tt"
+ ],
+ [
+ "▁let",
+ "t"
+ ],
+ [
+ "▁",
+ "lett"
+ ],
+ [
+ "en",
+ "ne"
+ ],
+ [
+ "enn",
+ "e"
+ ],
+ [
+ "()",
+ ")."
+ ],
+ [
+ "())",
+ "."
+ ],
+ [
+ "(",
+ "))."
+ ],
+ [
+ ".*",
+ "*"
+ ],
+ [
+ ".",
+ "**"
+ ],
+ [
+ "▁p",
+ "romise"
+ ],
+ [
+ "▁prom",
+ "ise"
+ ],
+ [
+ "el",
+ "ection"
+ ],
+ [
+ "ele",
+ "ction"
+ ],
+ [
+ "elect",
+ "ion"
+ ],
+ [
+ "e",
+ "lection"
+ ],
+ [
+ "Aut",
+ "h"
+ ],
+ [
+ "A",
+ "uth"
+ ],
+ [
+ "r",
+ "v"
+ ],
+ [
+ "ri",
+ "l"
+ ],
+ [
+ "r",
+ "il"
+ ],
+ [
+ "▁con",
+ "duct"
+ ],
+ [
+ "▁cond",
+ "uct"
+ ],
+ [
+ "▁condu",
+ "ct"
+ ],
+ [
+ "▁",
+ "conduct"
+ ],
+ [
+ "▁main",
+ "tain"
+ ],
+ [
+ "▁maint",
+ "ain"
+ ],
+ [
+ "▁bo",
+ "at"
+ ],
+ [
+ "▁",
+ "boat"
+ ],
+ [
+ "▁op",
+ "posite"
+ ],
+ [
+ "▁oppos",
+ "ite"
+ ],
+ [
+ "sp",
+ "in"
+ ],
+ [
+ "spi",
+ "n"
+ ],
+ [
+ "s",
+ "pin"
+ ],
+ [
+ "web",
+ "pack"
+ ],
+ [
+ "an",
+ "ta"
+ ],
+ [
+ "ant",
+ "a"
+ ],
+ [
+ "▁o",
+ "rient"
+ ],
+ [
+ "▁or",
+ "ient"
+ ],
+ [
+ "▁",
+ "orient"
+ ],
+ [
+ "▁s",
+ "uc"
+ ],
+ [
+ "▁su",
+ "c"
+ ],
+ [
+ "▁ex",
+ "ercise"
+ ],
+ [
+ "▁exerc",
+ "ise"
+ ],
+ [
+ "▁eff",
+ "icient"
+ ],
+ [
+ "▁",
+ "efficient"
+ ],
+ [
+ "▁trad",
+ "ition"
+ ],
+ [
+ "▁z",
+ "w"
+ ],
+ [
+ "▁",
+ "zw"
+ ],
+ [
+ "▁S",
+ "ud"
+ ],
+ [
+ "▁Su",
+ "d"
+ ],
+ [
+ "go",
+ "ing"
+ ],
+ [
+ "▁P",
+ "ier"
+ ],
+ [
+ "▁Pi",
+ "er"
+ ],
+ [
+ "in",
+ "v"
+ ],
+ [
+ "i",
+ "nv"
+ ],
+ [
+ "ip",
+ "es"
+ ],
+ [
+ "ipe",
+ "s"
+ ],
+ [
+ "i",
+ "pes"
+ ],
+ [
+ "ensure",
+ "math"
+ ],
+ [
+ "▁con",
+ "ver"
+ ],
+ [
+ "▁conv",
+ "er"
+ ],
+ [
+ "▁conve",
+ "r"
+ ],
+ [
+ "cre",
+ "en"
+ ],
+ [
+ "cr",
+ "een"
+ ],
+ [
+ "c",
+ "reen"
+ ],
+ [
+ "▁t",
+ "error"
+ ],
+ [
+ "▁ter",
+ "ror"
+ ],
+ [
+ "▁terr",
+ "or"
+ ],
+ [
+ "▁D",
+ "ou"
+ ],
+ [
+ "▁Do",
+ "u"
+ ],
+ [
+ "▁in",
+ "valid"
+ ],
+ [
+ "▁",
+ "invalid"
+ ],
+ [
+ "ce",
+ "ived"
+ ],
+ [
+ "ceive",
+ "d"
+ ],
+ [
+ "▁A",
+ "rab"
+ ],
+ [
+ "▁Ar",
+ "ab"
+ ],
+ [
+ "▁w",
+ "ire"
+ ],
+ [
+ "▁wir",
+ "e"
+ ],
+ [
+ "▁",
+ "wire"
+ ],
+ [
+ "ap",
+ "plication"
+ ],
+ [
+ "sh",
+ "ift"
+ ],
+ [
+ "Gener",
+ "ic"
+ ],
+ [
+ "▁P",
+ "lan"
+ ],
+ [
+ "▁Pl",
+ "an"
+ ],
+ [
+ "▁",
+ "Plan"
+ ],
+ [
+ "▁W",
+ "all"
+ ],
+ [
+ "▁Wal",
+ "l"
+ ],
+ [
+ "▁Wa",
+ "ll"
+ ],
+ [
+ "▁",
+ "Wall"
+ ],
+ [
+ "▁direct",
+ "ory"
+ ],
+ [
+ "▁director",
+ "y"
+ ],
+ [
+ "▁",
+ "directory"
+ ],
+ [
+ "▁e",
+ "gg"
+ ],
+ [
+ "▁eg",
+ "g"
+ ],
+ [
+ "▁we",
+ "alth"
+ ],
+ [
+ "▁",
+ "wealth"
+ ],
+ [
+ "ran",
+ "dom"
+ ],
+ [
+ "rand",
+ "om"
+ ],
+ [
+ "r",
+ "andom"
+ ],
+ [
+ "att",
+ "ribute"
+ ],
+ [
+ "▁h",
+ "ide"
+ ],
+ [
+ "▁hi",
+ "de"
+ ],
+ [
+ "▁hid",
+ "e"
+ ],
+ [
+ "▁",
+ "hide"
+ ],
+ [
+ "Se",
+ "rial"
+ ],
+ [
+ "Ser",
+ "ial"
+ ],
+ [
+ "S",
+ "erial"
+ ],
+ [
+ "ca",
+ "m"
+ ],
+ [
+ "c",
+ "am"
+ ],
+ [
+ "▁it",
+ "al"
+ ],
+ [
+ "▁i",
+ "tal"
+ ],
+ [
+ "▁",
+ "ital"
+ ],
+ [
+ "▁L",
+ "ine"
+ ],
+ [
+ "▁Lin",
+ "e"
+ ],
+ [
+ "▁Li",
+ "ne"
+ ],
+ [
+ "▁",
+ "Line"
+ ],
+ [
+ "▁C",
+ "HECK"
+ ],
+ [
+ "▁",
+ "CHECK"
+ ],
+ [
+ "ploy",
+ "ment"
+ ],
+ [
+ "▁mass",
+ "ive"
+ ],
+ [
+ "▁ex",
+ "tract"
+ ],
+ [
+ "▁ext",
+ "ract"
+ ],
+ [
+ "▁extra",
+ "ct"
+ ],
+ [
+ "▁extr",
+ "act"
+ ],
+ [
+ "▁",
+ "extract"
+ ],
+ [
+ "ch",
+ "ain"
+ ],
+ [
+ "cha",
+ "in"
+ ],
+ [
+ "Res",
+ "t"
+ ],
+ [
+ "Re",
+ "st"
+ ],
+ [
+ "R",
+ "est"
+ ],
+ [
+ "▁L",
+ "as"
+ ],
+ [
+ "▁La",
+ "s"
+ ],
+ [
+ "▁b",
+ "ear"
+ ],
+ [
+ "▁be",
+ "ar"
+ ],
+ [
+ "▁",
+ "bear"
+ ],
+ [
+ "▁l",
+ "inks"
+ ],
+ [
+ "▁link",
+ "s"
+ ],
+ [
+ "▁lin",
+ "ks"
+ ],
+ [
+ "▁",
+ "links"
+ ],
+ [
+ "▁new",
+ "sp"
+ ],
+ [
+ "▁news",
+ "p"
+ ],
+ [
+ "▁F",
+ "C"
+ ],
+ [
+ "▁",
+ "FC"
+ ],
+ [
+ "Car",
+ "d"
+ ],
+ [
+ "C",
+ "ard"
+ ],
+ [
+ "ak",
+ "s"
+ ],
+ [
+ "a",
+ "ks"
+ ],
+ [
+ "▁v",
+ "isible"
+ ],
+ [
+ "▁vis",
+ "ible"
+ ],
+ [
+ "▁",
+ "visible"
+ ],
+ [
+ "▁M",
+ "arc"
+ ],
+ [
+ "▁Mar",
+ "c"
+ ],
+ [
+ "▁Ma",
+ "rc"
+ ],
+ [
+ "▁B",
+ "oston"
+ ],
+ [
+ "▁Bo",
+ "ston"
+ ],
+ [
+ "▁Bos",
+ "ton"
+ ],
+ [
+ "▁res",
+ "erved"
+ ],
+ [
+ "▁reserv",
+ "ed"
+ ],
+ [
+ "▁reserve",
+ "d"
+ ],
+ [
+ "▁ro",
+ "of"
+ ],
+ [
+ "lic",
+ "enses"
+ ],
+ [
+ "license",
+ "s"
+ ],
+ [
+ "d",
+ "c"
+ ],
+ [
+ "▁In",
+ "formation"
+ ],
+ [
+ "▁",
+ "Information"
+ ],
+ [
+ "▁w",
+ "itness"
+ ],
+ [
+ "S",
+ "k"
+ ],
+ [
+ "*)",
+ ","
+ ],
+ [
+ "*",
+ "),"
+ ],
+ [
+ "Sc",
+ "ope"
+ ],
+ [
+ "S",
+ "cope"
+ ],
+ [
+ "']",
+ ";"
+ ],
+ [
+ "'",
+ "];"
+ ],
+ [
+ "▁M",
+ "ir"
+ ],
+ [
+ "▁Mi",
+ "r"
+ ],
+ [
+ "▁",
+ "Mir"
+ ],
+ [
+ "ud",
+ "ing"
+ ],
+ [
+ "udi",
+ "ng"
+ ],
+ [
+ "u",
+ "ding"
+ ],
+ [
+ "▁t",
+ "rend"
+ ],
+ [
+ "▁tr",
+ "end"
+ ],
+ [
+ "▁tre",
+ "nd"
+ ],
+ [
+ "▁tren",
+ "d"
+ ],
+ [
+ "re",
+ "p"
+ ],
+ [
+ "r",
+ "ep"
+ ],
+ [
+ "▁mus",
+ "ical"
+ ],
+ [
+ "▁music",
+ "al"
+ ],
+ [
+ "▁ne",
+ "ither"
+ ],
+ [
+ "▁nei",
+ "ther"
+ ],
+ [
+ "▁C",
+ "reat"
+ ],
+ [
+ "▁Cre",
+ "at"
+ ],
+ [
+ "▁",
+ "Creat"
+ ],
+ [
+ "▁pos",
+ "itions"
+ ],
+ [
+ "▁position",
+ "s"
+ ],
+ [
+ "▁posit",
+ "ions"
+ ],
+ [
+ "L",
+ "C"
+ ],
+ [
+ "rid",
+ "ge"
+ ],
+ [
+ "r",
+ "idge"
+ ],
+ [
+ "▁offic",
+ "ers"
+ ],
+ [
+ "▁office",
+ "rs"
+ ],
+ [
+ "▁officer",
+ "s"
+ ],
+ [
+ "▁vi",
+ "olence"
+ ],
+ [
+ "▁viol",
+ "ence"
+ ],
+ [
+ "▁T",
+ "em"
+ ],
+ [
+ "▁Te",
+ "m"
+ ],
+ [
+ "▁S",
+ "us"
+ ],
+ [
+ "▁Su",
+ "s"
+ ],
+ [
+ "▁W",
+ "ay"
+ ],
+ [
+ "▁Wa",
+ "y"
+ ],
+ [
+ "Af",
+ "ter"
+ ],
+ [
+ "A",
+ "fter"
+ ],
+ [
+ "ac",
+ "ket"
+ ],
+ [
+ "ack",
+ "et"
+ ],
+ [
+ "▁S",
+ "ou"
+ ],
+ [
+ "▁So",
+ "u"
+ ],
+ [
+ "ac",
+ "er"
+ ],
+ [
+ "ace",
+ "r"
+ ],
+ [
+ "a",
+ "cer"
+ ],
+ [
+ "|",
+ "|"
+ ],
+ [
+ "▁re",
+ "mark"
+ ],
+ [
+ "▁r",
+ "emark"
+ ],
+ [
+ "▁rem",
+ "ark"
+ ],
+ [
+ "▁",
+ "remark"
+ ],
+ [
+ "wa",
+ "ter"
+ ],
+ [
+ "w",
+ "ater"
+ ],
+ [
+ "n",
+ "ě"
+ ],
+ [
+ "▁С",
+ "а"
+ ],
+ [
+ "▁s",
+ "ed"
+ ],
+ [
+ "▁se",
+ "d"
+ ],
+ [
+ "▁",
+ "sed"
+ ],
+ [
+ "E",
+ "ach"
+ ],
+ [
+ "▁phot",
+ "ograph"
+ ],
+ [
+ "▁photo",
+ "graph"
+ ],
+ [
+ "▁let",
+ "ters"
+ ],
+ [
+ "▁letter",
+ "s"
+ ],
+ [
+ "▁lett",
+ "ers"
+ ],
+ [
+ "▁in",
+ "vent"
+ ],
+ [
+ "▁inv",
+ "ent"
+ ],
+ [
+ "▁M",
+ "as"
+ ],
+ [
+ "▁Ma",
+ "s"
+ ],
+ [
+ "▁s",
+ "ongs"
+ ],
+ [
+ "▁son",
+ "gs"
+ ],
+ [
+ "▁song",
+ "s"
+ ],
+ [
+ "ó",
+ "l"
+ ],
+ [
+ "ki",
+ "nd"
+ ],
+ [
+ "kin",
+ "d"
+ ],
+ [
+ "k",
+ "ind"
+ ],
+ [
+ "▁N",
+ "on"
+ ],
+ [
+ "▁No",
+ "n"
+ ],
+ [
+ "▁",
+ "Non"
+ ],
+ [
+ "▁d",
+ "ust"
+ ],
+ [
+ "▁du",
+ "st"
+ ],
+ [
+ "**",
+ ":"
+ ],
+ [
+ "*",
+ "*:"
+ ],
+ [
+ "nab",
+ "la"
+ ],
+ [
+ ".\"",
+ ","
+ ],
+ [
+ ".",
+ "\","
+ ],
+ [
+ "Loc",
+ "k"
+ ],
+ [
+ "Lo",
+ "ck"
+ ],
+ [
+ "L",
+ "ock"
+ ],
+ [
+ "▁Д",
+ "о"
+ ],
+ [
+ "▁cl",
+ "uster"
+ ],
+ [
+ "▁",
+ "cluster"
+ ],
+ [
+ "lo",
+ "ss"
+ ],
+ [
+ "los",
+ "s"
+ ],
+ [
+ "l",
+ "oss"
+ ],
+ [
+ "▁ASS",
+ "ERT"
+ ],
+ [
+ "▁",
+ "ASSERT"
+ ],
+ [
+ "fa",
+ "ll"
+ ],
+ [
+ "f",
+ "all"
+ ],
+ [
+ "▁re",
+ "ject"
+ ],
+ [
+ "▁",
+ "reject"
+ ],
+ [
+ "▁Sp",
+ "ring"
+ ],
+ [
+ "▁Spr",
+ "ing"
+ ],
+ [
+ "▁",
+ "Spring"
+ ],
+ [
+ "▁wed",
+ "ding"
+ ],
+ [
+ "▁g",
+ "rav"
+ ],
+ [
+ "▁gr",
+ "av"
+ ],
+ [
+ "▁gra",
+ "v"
+ ],
+ [
+ "▁",
+ "grav"
+ ],
+ [
+ "ress",
+ "ion"
+ ],
+ [
+ "r",
+ "ession"
+ ],
+ [
+ "li",
+ "mit"
+ ],
+ [
+ "lim",
+ "it"
+ ],
+ [
+ "l",
+ "imit"
+ ],
+ [
+ "RE",
+ "S"
+ ],
+ [
+ "R",
+ "ES"
+ ],
+ [
+ "]",
+ "}"
+ ],
+ [
+ "▁l",
+ "isted"
+ ],
+ [
+ "▁li",
+ "sted"
+ ],
+ [
+ "▁list",
+ "ed"
+ ],
+ [
+ "▁",
+ "listed"
+ ],
+ [
+ "▁T",
+ "ele"
+ ],
+ [
+ "▁Te",
+ "le"
+ ],
+ [
+ "▁Tel",
+ "e"
+ ],
+ [
+ "▁",
+ "Tele"
+ ],
+ [
+ "hl",
+ "ine"
+ ],
+ [
+ "h",
+ "line"
+ ],
+ [
+ "▁ch",
+ "ief"
+ ],
+ [
+ "▁chi",
+ "ef"
+ ],
+ [
+ "ME",
+ "M"
+ ],
+ [
+ "M",
+ "EM"
+ ],
+ [
+ "да",
+ "р"
+ ],
+ [
+ "д",
+ "ар"
+ ],
+ [
+ "▁exp",
+ "ensive"
+ ],
+ [
+ "tr",
+ "ace"
+ ],
+ [
+ "tra",
+ "ce"
+ ],
+ [
+ "▁R",
+ "og"
+ ],
+ [
+ "▁Ro",
+ "g"
+ ],
+ [
+ "▁C",
+ "oll"
+ ],
+ [
+ "▁Col",
+ "l"
+ ],
+ [
+ "▁Co",
+ "ll"
+ ],
+ [
+ "▁",
+ "Coll"
+ ],
+ [
+ "▁Aut",
+ "hor"
+ ],
+ [
+ "▁Auth",
+ "or"
+ ],
+ [
+ "▁",
+ "Author"
+ ],
+ [
+ "▁B",
+ "oard"
+ ],
+ [
+ "▁Bo",
+ "ard"
+ ],
+ [
+ "▁",
+ "Board"
+ ],
+ [
+ "▁C",
+ "apt"
+ ],
+ [
+ "▁Cap",
+ "t"
+ ],
+ [
+ "▁Ca",
+ "pt"
+ ],
+ [
+ "▁",
+ "Capt"
+ ],
+ [
+ "TE",
+ "XT"
+ ],
+ [
+ "T",
+ "EXT"
+ ],
+ [
+ "▁re",
+ "con"
+ ],
+ [
+ "▁rec",
+ "on"
+ ],
+ [
+ "es",
+ "ta"
+ ],
+ [
+ "est",
+ "a"
+ ],
+ [
+ "e",
+ "sta"
+ ],
+ [
+ "▁proper",
+ "ly"
+ ],
+ [
+ "▁&",
+ "\\"
+ ],
+ [
+ "▁",
+ "&\\"
+ ],
+ [
+ "le",
+ "ton"
+ ],
+ [
+ "let",
+ "on"
+ ],
+ [
+ "l",
+ "eton"
+ ],
+ [
+ "ik",
+ "er"
+ ],
+ [
+ "ike",
+ "r"
+ ],
+ [
+ "i",
+ "ker"
+ ],
+ [
+ "G",
+ "u"
+ ],
+ [
+ "▁K",
+ "om"
+ ],
+ [
+ "▁Ko",
+ "m"
+ ],
+ [
+ "oc",
+ "o"
+ ],
+ [
+ "o",
+ "co"
+ ],
+ [
+ "▁any",
+ "more"
+ ],
+ [
+ "▁t",
+ "aste"
+ ],
+ [
+ "▁ta",
+ "ste"
+ ],
+ [
+ "▁tast",
+ "e"
+ ],
+ [
+ "▁S",
+ "anta"
+ ],
+ [
+ "▁San",
+ "ta"
+ ],
+ [
+ "▁Sant",
+ "a"
+ ],
+ [
+ "ge",
+ "x"
+ ],
+ [
+ "g",
+ "ex"
+ ],
+ [
+ "▁Se",
+ "cret"
+ ],
+ [
+ "▁Sec",
+ "ret"
+ ],
+ [
+ "▁",
+ "Secret"
+ ],
+ [
+ "▁tal",
+ "ent"
+ ],
+ [
+ "▁tale",
+ "nt"
+ ],
+ [
+ "▁mom",
+ "ents"
+ ],
+ [
+ "▁moment",
+ "s"
+ ],
+ [
+ "▁mo",
+ "ments"
+ ],
+ [
+ "▁B",
+ "a"
+ ],
+ [
+ "▁ex",
+ "tr"
+ ],
+ [
+ "▁ext",
+ "r"
+ ],
+ [
+ "▁",
+ "extr"
+ ],
+ [
+ "▁Com",
+ "mission"
+ ],
+ [
+ "▁Comm",
+ "ission"
+ ],
+ [
+ "▁mod",
+ "ify"
+ ],
+ [
+ "▁Fig",
+ "ure"
+ ],
+ [
+ "▁",
+ "Figure"
+ ],
+ [
+ "▁d",
+ "omin"
+ ],
+ [
+ "▁do",
+ "min"
+ ],
+ [
+ "▁dom",
+ "in"
+ ],
+ [
+ "▁",
+ "domin"
+ ],
+ [
+ "▁p",
+ "lot"
+ ],
+ [
+ "▁pl",
+ "ot"
+ ],
+ [
+ "▁",
+ "plot"
+ ],
+ [
+ "en",
+ "ger"
+ ],
+ [
+ "eng",
+ "er"
+ ],
+ [
+ "enge",
+ "r"
+ ],
+ [
+ "ut",
+ "ch"
+ ],
+ [
+ "▁c",
+ "ities"
+ ],
+ [
+ "▁cit",
+ "ies"
+ ],
+ [
+ "▁ci",
+ "ties"
+ ],
+ [
+ "▁n",
+ "ut"
+ ],
+ [
+ "▁nu",
+ "t"
+ ],
+ [
+ "▁",
+ "nut"
+ ],
+ [
+ "pro",
+ "file"
+ ],
+ [
+ "prof",
+ "ile"
+ ],
+ [
+ "▁S",
+ "tat"
+ ],
+ [
+ "▁St",
+ "at"
+ ],
+ [
+ "▁Sta",
+ "t"
+ ],
+ [
+ "▁",
+ "Stat"
+ ],
+ [
+ "▁n",
+ "odes"
+ ],
+ [
+ "▁no",
+ "des"
+ ],
+ [
+ "▁node",
+ "s"
+ ],
+ [
+ "▁nod",
+ "es"
+ ],
+ [
+ "▁",
+ "nodes"
+ ],
+ [
+ "▁n",
+ "s"
+ ],
+ [
+ "▁",
+ "ns"
+ ],
+ [
+ "ess",
+ "ages"
+ ],
+ [
+ "essage",
+ "s"
+ ],
+ [
+ "essa",
+ "ges"
+ ],
+ [
+ "im",
+ "pl"
+ ],
+ [
+ "imp",
+ "l"
+ ],
+ [
+ "ic",
+ "ker"
+ ],
+ [
+ "ick",
+ "er"
+ ],
+ [
+ "i",
+ "cker"
+ ],
+ [
+ "▁ex",
+ "amples"
+ ],
+ [
+ "▁example",
+ "s"
+ ],
+ [
+ "▁exam",
+ "ples"
+ ],
+ [
+ "ab",
+ "eth"
+ ],
+ [
+ "abe",
+ "th"
+ ],
+ [
+ "abet",
+ "h"
+ ],
+ [
+ "▁st",
+ "ated"
+ ],
+ [
+ "▁stat",
+ "ed"
+ ],
+ [
+ "▁state",
+ "d"
+ ],
+ [
+ "▁sta",
+ "ted"
+ ],
+ [
+ "fi",
+ "re"
+ ],
+ [
+ "f",
+ "ire"
+ ],
+ [
+ "bu",
+ "l"
+ ],
+ [
+ "b",
+ "ul"
+ ],
+ [
+ "▁danger",
+ "ous"
+ ],
+ [
+ "▁P",
+ "ay"
+ ],
+ [
+ "▁Pa",
+ "y"
+ ],
+ [
+ "▁",
+ "Pay"
+ ],
+ [
+ "▁G",
+ "re"
+ ],
+ [
+ "▁Gr",
+ "e"
+ ],
+ [
+ "▁",
+ "Gre"
+ ],
+ [
+ "▁Mon",
+ "day"
+ ],
+ [
+ "▁Mond",
+ "ay"
+ ],
+ [
+ "es",
+ "ome"
+ ],
+ [
+ "eso",
+ "me"
+ ],
+ [
+ "e",
+ "some"
+ ],
+ [
+ "ig",
+ "an"
+ ],
+ [
+ "iga",
+ "n"
+ ],
+ [
+ "i",
+ "gan"
+ ],
+ [
+ "ru",
+ "nd"
+ ],
+ [
+ "run",
+ "d"
+ ],
+ [
+ "r",
+ "und"
+ ],
+ [
+ "pr",
+ "ise"
+ ],
+ [
+ "p",
+ "rise"
+ ],
+ [
+ "fa",
+ "il"
+ ],
+ [
+ "f",
+ "ail"
+ ],
+ [
+ "▁N",
+ "ever"
+ ],
+ [
+ "▁Ne",
+ "ver"
+ ],
+ [
+ "▁Nev",
+ "er"
+ ],
+ [
+ "▁",
+ "Never"
+ ],
+ [
+ "A",
+ "v"
+ ],
+ [
+ "▁line",
+ "ar"
+ ],
+ [
+ "▁lin",
+ "ear"
+ ],
+ [
+ "▁",
+ "linear"
+ ],
+ [
+ "▁u",
+ "l"
+ ],
+ [
+ "▁",
+ "ul"
+ ],
+ [
+ "WA",
+ "R"
+ ],
+ [
+ "W",
+ "AR"
+ ],
+ [
+ "ре",
+ "н"
+ ],
+ [
+ "р",
+ "ен"
+ ],
+ [
+ "▁A",
+ "T"
+ ],
+ [
+ "▁",
+ "AT"
+ ],
+ [
+ "▁d",
+ "op"
+ ],
+ [
+ "▁do",
+ "p"
+ ],
+ [
+ "▁n",
+ "ou"
+ ],
+ [
+ "▁no",
+ "u"
+ ],
+ [
+ "Des",
+ "t"
+ ],
+ [
+ "De",
+ "st"
+ ],
+ [
+ "D",
+ "est"
+ ],
+ [
+ "▁claim",
+ "s"
+ ],
+ [
+ "en",
+ "da"
+ ],
+ [
+ "end",
+ "a"
+ ],
+ [
+ "▁c",
+ "razy"
+ ],
+ [
+ "▁cr",
+ "azy"
+ ],
+ [
+ "ge",
+ "l"
+ ],
+ [
+ "g",
+ "el"
+ ],
+ [
+ "og",
+ "gle"
+ ],
+ [
+ "ogg",
+ "le"
+ ],
+ [
+ "▁rep",
+ "resentation"
+ ],
+ [
+ "▁represent",
+ "ation"
+ ],
+ [
+ "in",
+ "en"
+ ],
+ [
+ "ine",
+ "n"
+ ],
+ [
+ "i",
+ "nen"
+ ],
+ [
+ "▁altern",
+ "ative"
+ ],
+ [
+ "▁alter",
+ "native"
+ ],
+ [
+ "D",
+ "M"
+ ],
+ [
+ "AB",
+ "ILITY"
+ ],
+ [
+ "face",
+ "s"
+ ],
+ [
+ "fa",
+ "ces"
+ ],
+ [
+ "fac",
+ "es"
+ ],
+ [
+ "f",
+ "aces"
+ ],
+ [
+ "▁do",
+ "ors"
+ ],
+ [
+ "▁door",
+ "s"
+ ],
+ [
+ "▁",
+ "doors"
+ ],
+ [
+ "at",
+ "iv"
+ ],
+ [
+ "ati",
+ "v"
+ ],
+ [
+ "Lo",
+ "ok"
+ ],
+ [
+ "L",
+ "ook"
+ ],
+ [
+ "▁J",
+ "SON"
+ ],
+ [
+ "▁JS",
+ "ON"
+ ],
+ [
+ "▁",
+ "JSON"
+ ],
+ [
+ "▁appe",
+ "arance"
+ ],
+ [
+ "▁appear",
+ "ance"
+ ],
+ [
+ "б",
+ "ря"
+ ],
+ [
+ "S",
+ "QL"
+ ],
+ [
+ "▁sil",
+ "ence"
+ ],
+ [
+ "ud",
+ "o"
+ ],
+ [
+ "u",
+ "do"
+ ],
+ [
+ "▁Direct",
+ "or"
+ ],
+ [
+ "▁Dire",
+ "ctor"
+ ],
+ [
+ "▁Dir",
+ "ector"
+ ],
+ [
+ "State",
+ "ment"
+ ],
+ [
+ "Stat",
+ "ement"
+ ],
+ [
+ "se",
+ "lected"
+ ],
+ [
+ "select",
+ "ed"
+ ],
+ [
+ "hi",
+ "gh"
+ ],
+ [
+ "h",
+ "igh"
+ ],
+ [
+ "pr",
+ "ime"
+ ],
+ [
+ "prim",
+ "e"
+ ],
+ [
+ "▁ign",
+ "ore"
+ ],
+ [
+ "▁ignor",
+ "e"
+ ],
+ [
+ "▁",
+ "ignore"
+ ],
+ [
+ "▁col",
+ "ors"
+ ],
+ [
+ "▁color",
+ "s"
+ ],
+ [
+ "▁",
+ "colors"
+ ],
+ [
+ "us",
+ "hing"
+ ],
+ [
+ "ush",
+ "ing"
+ ],
+ [
+ "▁v",
+ "irt"
+ ],
+ [
+ "▁vi",
+ "rt"
+ ],
+ [
+ "▁vir",
+ "t"
+ ],
+ [
+ "▁",
+ "virt"
+ ],
+ [
+ "man",
+ "ager"
+ ],
+ [
+ "▁rem",
+ "ote"
+ ],
+ [
+ "▁remot",
+ "e"
+ ],
+ [
+ "▁",
+ "remote"
+ ],
+ [
+ "ł",
+ "o"
+ ],
+ [
+ "sm",
+ "all"
+ ],
+ [
+ "▁cr",
+ "ime"
+ ],
+ [
+ "▁crim",
+ "e"
+ ],
+ [
+ "▁cri",
+ "me"
+ ],
+ [
+ "r",
+ "b"
+ ],
+ [
+ "▁c",
+ "reation"
+ ],
+ [
+ "▁cre",
+ "ation"
+ ],
+ [
+ "▁creat",
+ "ion"
+ ],
+ [
+ "▁f",
+ "light"
+ ],
+ [
+ "▁fl",
+ "ight"
+ ],
+ [
+ "▁S",
+ "ign"
+ ],
+ [
+ "▁Si",
+ "gn"
+ ],
+ [
+ "▁Sig",
+ "n"
+ ],
+ [
+ "▁",
+ "Sign"
+ ],
+ [
+ "IL",
+ "E"
+ ],
+ [
+ "I",
+ "LE"
+ ],
+ [
+ "▁D",
+ "O"
+ ],
+ [
+ "▁",
+ "DO"
+ ],
+ [
+ "com",
+ "ment"
+ ],
+ [
+ "comm",
+ "ent"
+ ],
+ [
+ "▁C",
+ "ost"
+ ],
+ [
+ "▁Co",
+ "st"
+ ],
+ [
+ "▁Cos",
+ "t"
+ ],
+ [
+ "▁",
+ "Cost"
+ ],
+ [
+ "._",
+ "_"
+ ],
+ [
+ ".",
+ "__"
+ ],
+ [
+ "▁C",
+ "op"
+ ],
+ [
+ "▁Co",
+ "p"
+ ],
+ [
+ "▁",
+ "Cop"
+ ],
+ [
+ "▁v",
+ "om"
+ ],
+ [
+ "▁vo",
+ "m"
+ ],
+ [
+ "▁Sc",
+ "ience"
+ ],
+ [
+ "▁Sci",
+ "ence"
+ ],
+ [
+ "ле",
+ "ния"
+ ],
+ [
+ "oo",
+ "p"
+ ],
+ [
+ "o",
+ "op"
+ ],
+ [
+ "inter",
+ "face"
+ ],
+ [
+ "▁WARRAN",
+ "TIES"
+ ],
+ [
+ "▁P",
+ "age"
+ ],
+ [
+ "▁Pa",
+ "ge"
+ ],
+ [
+ "▁",
+ "Page"
+ ],
+ [
+ "**",
+ "****"
+ ],
+ [
+ "****",
+ "**"
+ ],
+ [
+ "***",
+ "***"
+ ],
+ [
+ "ско",
+ "м"
+ ],
+ [
+ "с",
+ "ком"
+ ],
+ [
+ "TR",
+ "UE"
+ ],
+ [
+ "▁re",
+ "peated"
+ ],
+ [
+ "▁repe",
+ "ated"
+ ],
+ [
+ "▁repeat",
+ "ed"
+ ],
+ [
+ "▁е",
+ "го"
+ ],
+ [
+ "ш",
+ "о"
+ ],
+ [
+ "▁r",
+ "oz"
+ ],
+ [
+ "▁ro",
+ "z"
+ ],
+ [
+ "▁",
+ "roz"
+ ],
+ [
+ "P",
+ "e"
+ ],
+ [
+ "▁IS",
+ "BN"
+ ],
+ [
+ "ir",
+ "ts"
+ ],
+ [
+ "irt",
+ "s"
+ ],
+ [
+ "pos",
+ "es"
+ ],
+ [
+ "po",
+ "ses"
+ ],
+ [
+ "pose",
+ "s"
+ ],
+ [
+ "p",
+ "oses"
+ ],
+ [
+ "})",
+ "$"
+ ],
+ [
+ "}",
+ ")$"
+ ],
+ [
+ "▁",
+ "І"
+ ],
+ [
+ "child",
+ "ren"
+ ],
+ [
+ "ble",
+ "s"
+ ],
+ [
+ "bl",
+ "es"
+ ],
+ [
+ "b",
+ "les"
+ ],
+ [
+ "EC",
+ "T"
+ ],
+ [
+ "E",
+ "CT"
+ ],
+ [
+ "▁i",
+ "z"
+ ],
+ [
+ "▁",
+ "iz"
+ ],
+ [
+ "▁b",
+ "uilder"
+ ],
+ [
+ "▁build",
+ "er"
+ ],
+ [
+ "▁",
+ "builder"
+ ],
+ [
+ "▁M",
+ "edia"
+ ],
+ [
+ "▁Med",
+ "ia"
+ ],
+ [
+ "▁",
+ "Media"
+ ],
+ [
+ "ia",
+ "t"
+ ],
+ [
+ "i",
+ "at"
+ ],
+ [
+ "▁contr",
+ "ast"
+ ],
+ [
+ "▁contra",
+ "st"
+ ],
+ [
+ "”",
+ ","
+ ],
+ [
+ "▁L",
+ "ink"
+ ],
+ [
+ "▁Lin",
+ "k"
+ ],
+ [
+ "▁",
+ "Link"
+ ],
+ [
+ "▁Educ",
+ "ation"
+ ],
+ [
+ "▁j",
+ "oint"
+ ],
+ [
+ "▁join",
+ "t"
+ ],
+ [
+ "▁jo",
+ "int"
+ ],
+ [
+ "▁",
+ "joint"
+ ],
+ [
+ "▁ex",
+ "ternal"
+ ],
+ [
+ "▁extern",
+ "al"
+ ],
+ [
+ "▁",
+ "external"
+ ],
+ [
+ "▁ро",
+ "з"
+ ],
+ [
+ "▁b",
+ "its"
+ ],
+ [
+ "▁bit",
+ "s"
+ ],
+ [
+ "▁bi",
+ "ts"
+ ],
+ [
+ "▁",
+ "bits"
+ ],
+ [
+ "FO",
+ "RM"
+ ],
+ [
+ "FOR",
+ "M"
+ ],
+ [
+ "F",
+ "ORM"
+ ],
+ [
+ "er",
+ "man"
+ ],
+ [
+ "erm",
+ "an"
+ ],
+ [
+ "w",
+ "p"
+ ],
+ [
+ "▁M",
+ "ike"
+ ],
+ [
+ "▁Mi",
+ "ke"
+ ],
+ [
+ "▁Mik",
+ "e"
+ ],
+ [
+ "▁M",
+ "aster"
+ ],
+ [
+ "▁Ma",
+ "ster"
+ ],
+ [
+ "▁Mas",
+ "ter"
+ ],
+ [
+ "▁",
+ "Master"
+ ],
+ [
+ "▁sen",
+ "ior"
+ ],
+ [
+ "▁N",
+ "av"
+ ],
+ [
+ "▁Na",
+ "v"
+ ],
+ [
+ "▁",
+ "Nav"
+ ],
+ [
+ "▁record",
+ "ed"
+ ],
+ [
+ "el",
+ "ing"
+ ],
+ [
+ "eli",
+ "ng"
+ ],
+ [
+ "elin",
+ "g"
+ ],
+ [
+ "e",
+ "ling"
+ ],
+ [
+ "es",
+ "h"
+ ],
+ [
+ "e",
+ "sh"
+ ],
+ [
+ "f",
+ "x"
+ ],
+ [
+ "ка",
+ "н"
+ ],
+ [
+ "к",
+ "ан"
+ ],
+ [
+ "▁t",
+ "all"
+ ],
+ [
+ "▁tal",
+ "l"
+ ],
+ [
+ "▁ta",
+ "ll"
+ ],
+ [
+ "▁John",
+ "son"
+ ],
+ [
+ "▁s",
+ "ono"
+ ],
+ [
+ "▁so",
+ "no"
+ ],
+ [
+ "▁son",
+ "o"
+ ],
+ [
+ "▁an",
+ "che"
+ ],
+ [
+ "▁anc",
+ "he"
+ ],
+ [
+ "▁anch",
+ "e"
+ ],
+ [
+ "▁",
+ "anche"
+ ],
+ [
+ "ic",
+ "ken"
+ ],
+ [
+ "ick",
+ "en"
+ ],
+ [
+ "i",
+ "cken"
+ ],
+ [
+ "lo",
+ "op"
+ ],
+ [
+ "l",
+ "oop"
+ ],
+ [
+ "ici",
+ "ency"
+ ],
+ [
+ "empor",
+ "ary"
+ ],
+ [
+ "▁D",
+ "oes"
+ ],
+ [
+ "▁Do",
+ "es"
+ ],
+ [
+ "▁",
+ "Does"
+ ],
+ [
+ "▁re",
+ "lation"
+ ],
+ [
+ "▁rel",
+ "ation"
+ ],
+ [
+ "▁",
+ "relation"
+ ],
+ [
+ "м",
+ "ы"
+ ],
+ [
+ "wa",
+ "s"
+ ],
+ [
+ "w",
+ "as"
+ ],
+ [
+ "lo",
+ "w"
+ ],
+ [
+ "l",
+ "ow"
+ ],
+ [
+ "ich",
+ "te"
+ ],
+ [
+ "icht",
+ "e"
+ ],
+ [
+ "i",
+ "chte"
+ ],
+ [
+ "▁J",
+ "ones"
+ ],
+ [
+ "▁Jo",
+ "nes"
+ ],
+ [
+ "▁Jon",
+ "es"
+ ],
+ [
+ "▁bed",
+ "room"
+ ],
+ [
+ "DI",
+ "S"
+ ],
+ [
+ "D",
+ "IS"
+ ],
+ [
+ "▁mag",
+ "net"
+ ],
+ [
+ "▁magn",
+ "et"
+ ],
+ [
+ "▁Eng",
+ "ine"
+ ],
+ [
+ "▁",
+ "Engine"
+ ],
+ [
+ "▁feel",
+ "ings"
+ ],
+ [
+ "▁feeling",
+ "s"
+ ],
+ [
+ "▁fee",
+ "lings"
+ ],
+ [
+ "G",
+ "C"
+ ],
+ [
+ "▁t",
+ "orn"
+ ],
+ [
+ "▁to",
+ "rn"
+ ],
+ [
+ "▁tor",
+ "n"
+ ],
+ [
+ "▁relationship",
+ "s"
+ ],
+ [
+ "▁relation",
+ "ships"
+ ],
+ [
+ "▁Р",
+ "е"
+ ],
+ [
+ "▁p",
+ "roud"
+ ],
+ [
+ "▁pro",
+ "ud"
+ ],
+ [
+ "▁pr",
+ "oud"
+ ],
+ [
+ "▁t",
+ "we"
+ ],
+ [
+ "▁tw",
+ "e"
+ ],
+ [
+ "ov",
+ "al"
+ ],
+ [
+ "ova",
+ "l"
+ ],
+ [
+ "o",
+ "val"
+ ],
+ [
+ "▁w",
+ "aste"
+ ],
+ [
+ "▁was",
+ "te"
+ ],
+ [
+ "▁wa",
+ "ste"
+ ],
+ [
+ "▁red",
+ "uced"
+ ],
+ [
+ "▁redu",
+ "ced"
+ ],
+ [
+ "▁reduce",
+ "d"
+ ],
+ [
+ "il",
+ "ton"
+ ],
+ [
+ "ilt",
+ "on"
+ ],
+ [
+ "B",
+ "P"
+ ],
+ [
+ "▁for",
+ "got"
+ ],
+ [
+ "▁forg",
+ "ot"
+ ],
+ [
+ "▁bod",
+ "ies"
+ ],
+ [
+ "▁H",
+ "aw"
+ ],
+ [
+ "▁Ha",
+ "w"
+ ],
+ [
+ "la",
+ "g"
+ ],
+ [
+ "l",
+ "ag"
+ ],
+ [
+ "▁w",
+ "ww"
+ ],
+ [
+ "▁",
+ "www"
+ ],
+ [
+ "do",
+ "or"
+ ],
+ [
+ "d",
+ "oor"
+ ],
+ [
+ "▁s",
+ "ufficient"
+ ],
+ [
+ "▁suff",
+ "icient"
+ ],
+ [
+ "▁doll",
+ "ars"
+ ],
+ [
+ "▁dollar",
+ "s"
+ ],
+ [
+ "Le",
+ "n"
+ ],
+ [
+ "L",
+ "en"
+ ],
+ [
+ "▁talk",
+ "ed"
+ ],
+ [
+ "▁tal",
+ "ked"
+ ],
+ [
+ "▁b",
+ "ond"
+ ],
+ [
+ "▁bo",
+ "nd"
+ ],
+ [
+ "▁bon",
+ "d"
+ ],
+ [
+ "▁B",
+ "or"
+ ],
+ [
+ "▁Bo",
+ "r"
+ ],
+ [
+ "}}",
+ "{"
+ ],
+ [
+ "}",
+ "}{"
+ ],
+ [
+ "ro",
+ "d"
+ ],
+ [
+ "r",
+ "od"
+ ],
+ [
+ "Pass",
+ "word"
+ ],
+ [
+ "qu",
+ "are"
+ ],
+ [
+ "▁l",
+ "ights"
+ ],
+ [
+ "▁light",
+ "s"
+ ],
+ [
+ "▁",
+ "lights"
+ ],
+ [
+ "er",
+ "en"
+ ],
+ [
+ "ere",
+ "n"
+ ],
+ [
+ "e",
+ "ren"
+ ],
+ [
+ "▁th",
+ "irty"
+ ],
+ [
+ "N",
+ "C"
+ ],
+ [
+ "▁T",
+ "ODO"
+ ],
+ [
+ "▁TO",
+ "DO"
+ ],
+ [
+ "▁res",
+ "pond"
+ ],
+ [
+ "▁respon",
+ "d"
+ ],
+ [
+ "▁resp",
+ "ond"
+ ],
+ [
+ "▁",
+ "respond"
+ ],
+ [
+ "ки",
+ "х"
+ ],
+ [
+ "dir",
+ "ect"
+ ],
+ [
+ "di",
+ "rect"
+ ],
+ [
+ "dire",
+ "ct"
+ ],
+ [
+ "d",
+ "irect"
+ ],
+ [
+ "a",
+ "ção"
+ ],
+ [
+ "▁he",
+ "av"
+ ],
+ [
+ "Med",
+ "ia"
+ ],
+ [
+ "M",
+ "edia"
+ ],
+ [
+ "ex",
+ "it"
+ ],
+ [
+ "e",
+ "xit"
+ ],
+ [
+ "L",
+ "icense"
+ ],
+ [
+ "`",
+ "."
+ ],
+ [
+ "▁m",
+ "ixed"
+ ],
+ [
+ "▁mix",
+ "ed"
+ ],
+ [
+ "▁d",
+ "esk"
+ ],
+ [
+ "▁de",
+ "sk"
+ ],
+ [
+ "▁des",
+ "k"
+ ],
+ [
+ "▁te",
+ "aching"
+ ],
+ [
+ "▁teach",
+ "ing"
+ ],
+ [
+ "▁tea",
+ "ching"
+ ],
+ [
+ "▁m",
+ "aj"
+ ],
+ [
+ "▁ma",
+ "j"
+ ],
+ [
+ "▁n",
+ "erv"
+ ],
+ [
+ "▁ne",
+ "rv"
+ ],
+ [
+ "▁ner",
+ "v"
+ ],
+ [
+ "in",
+ "ations"
+ ],
+ [
+ "ination",
+ "s"
+ ],
+ [
+ "type",
+ "of"
+ ],
+ [
+ "▁co",
+ "ast"
+ ],
+ [
+ "▁ж",
+ "е"
+ ],
+ [
+ "▁",
+ "же"
+ ],
+ [
+ "▁be",
+ "side"
+ ],
+ [
+ "▁bes",
+ "ide"
+ ],
+ [
+ "um",
+ "my"
+ ],
+ [
+ "umm",
+ "y"
+ ],
+ [
+ "Do",
+ "c"
+ ],
+ [
+ "D",
+ "oc"
+ ],
+ [
+ "▁sche",
+ "dule"
+ ],
+ [
+ "▁schedul",
+ "e"
+ ],
+ [
+ "▁sched",
+ "ule"
+ ],
+ [
+ "▁",
+ "schedule"
+ ],
+ [
+ "▁re",
+ "cover"
+ ],
+ [
+ "▁rec",
+ "over"
+ ],
+ [
+ "▁Fur",
+ "ther"
+ ],
+ [
+ "▁ste",
+ "el"
+ ],
+ [
+ "bo",
+ "ot"
+ ],
+ [
+ "b",
+ "oot"
+ ],
+ [
+ "▁Per",
+ "haps"
+ ],
+ [
+ "▁с",
+ "ъ"
+ ],
+ [
+ "▁O",
+ "s"
+ ],
+ [
+ "▁",
+ "Os"
+ ],
+ [
+ "ri",
+ "ck"
+ ],
+ [
+ "ric",
+ "k"
+ ],
+ [
+ "r",
+ "ick"
+ ],
+ [
+ "▁В",
+ "и"
+ ],
+ [
+ "Supp",
+ "ort"
+ ],
+ [
+ "Sup",
+ "port"
+ ],
+ [
+ "S",
+ "upport"
+ ],
+ [
+ "▁(",
+ "_"
+ ],
+ [
+ "▁",
+ "(_"
+ ],
+ [
+ "ni",
+ "l"
+ ],
+ [
+ "n",
+ "il"
+ ],
+ [
+ "pi",
+ "s"
+ ],
+ [
+ "p",
+ "is"
+ ],
+ [
+ "x",
+ "pected"
+ ],
+ [
+ "▁process",
+ "ing"
+ ],
+ [
+ "▁proces",
+ "sing"
+ ],
+ [
+ "▁",
+ "processing"
+ ],
+ [
+ "Bu",
+ "ild"
+ ],
+ [
+ "B",
+ "uild"
+ ],
+ [
+ "ar",
+ "ian"
+ ],
+ [
+ "ari",
+ "an"
+ ],
+ [
+ "aria",
+ "n"
+ ],
+ [
+ "a",
+ "rian"
+ ],
+ [
+ "▁i",
+ "con"
+ ],
+ [
+ "▁ic",
+ "on"
+ ],
+ [
+ "▁",
+ "icon"
+ ],
+ [
+ "▁C",
+ "A"
+ ],
+ [
+ "▁",
+ "CA"
+ ],
+ [
+ "wi",
+ "ck"
+ ],
+ [
+ "w",
+ "ick"
+ ],
+ [
+ "=",
+ "("
+ ],
+ [
+ "▁al",
+ "gorithm"
+ ],
+ [
+ "▁",
+ "algorithm"
+ ],
+ [
+ "▁You",
+ "ng"
+ ],
+ [
+ "▁Man",
+ "agement"
+ ],
+ [
+ "▁",
+ "Management"
+ ],
+ [
+ "▁anc",
+ "ient"
+ ],
+ [
+ "▁anci",
+ "ent"
+ ],
+ [
+ "но",
+ "сть"
+ ],
+ [
+ "ност",
+ "ь"
+ ],
+ [
+ "ot",
+ "i"
+ ],
+ [
+ "o",
+ "ti"
+ ],
+ [
+ "▁comb",
+ "ination"
+ ],
+ [
+ "wor",
+ "ld"
+ ],
+ [
+ "w",
+ "orld"
+ ],
+ [
+ "n",
+ "n"
+ ],
+ [
+ "▁d",
+ "ram"
+ ],
+ [
+ "▁dr",
+ "am"
+ ],
+ [
+ "en",
+ "abled"
+ ],
+ [
+ "ena",
+ "bled"
+ ],
+ [
+ "enable",
+ "d"
+ ],
+ [
+ "A",
+ "c"
+ ],
+ [
+ "C",
+ "CESS"
+ ],
+ [
+ "ar",
+ "ation"
+ ],
+ [
+ "▁bl",
+ "ocks"
+ ],
+ [
+ "▁block",
+ "s"
+ ],
+ [
+ "▁blo",
+ "cks"
+ ],
+ [
+ "▁",
+ "blocks"
+ ],
+ [
+ "▁Ang",
+ "eles"
+ ],
+ [
+ "▁Angel",
+ "es"
+ ],
+ [
+ "▁Q",
+ "ual"
+ ],
+ [
+ "▁Qu",
+ "al"
+ ],
+ [
+ "▁",
+ "Qual"
+ ],
+ [
+ "▁suc",
+ "ceed"
+ ],
+ [
+ "▁succ",
+ "eed"
+ ],
+ [
+ "net",
+ "work"
+ ],
+ [
+ "▁ob",
+ "lig"
+ ],
+ [
+ "spring",
+ "framework"
+ ],
+ [
+ "▁T",
+ "re"
+ ],
+ [
+ "▁Tr",
+ "e"
+ ],
+ [
+ "ok",
+ "es"
+ ],
+ [
+ "oke",
+ "s"
+ ],
+ [
+ "o",
+ "kes"
+ ],
+ [
+ "mu",
+ "n"
+ ],
+ [
+ "m",
+ "un"
+ ],
+ [
+ "▁Net",
+ "work"
+ ],
+ [
+ "▁",
+ "Network"
+ ],
+ [
+ "De",
+ "l"
+ ],
+ [
+ "D",
+ "el"
+ ],
+ [
+ "▁e",
+ "state"
+ ],
+ [
+ "▁est",
+ "ate"
+ ],
+ [
+ "▁esta",
+ "te"
+ ],
+ [
+ "▁l",
+ "iqu"
+ ],
+ [
+ "▁li",
+ "qu"
+ ],
+ [
+ "▁p",
+ "ob"
+ ],
+ [
+ "▁po",
+ "b"
+ ],
+ [
+ "▁d",
+ "ad"
+ ],
+ [
+ "▁da",
+ "d"
+ ],
+ [
+ "▁dist",
+ "inct"
+ ],
+ [
+ "▁T",
+ "it"
+ ],
+ [
+ "▁Ti",
+ "t"
+ ],
+ [
+ "▁L",
+ "ear"
+ ],
+ [
+ "▁Le",
+ "ar"
+ ],
+ [
+ "fer",
+ "red"
+ ],
+ [
+ "and",
+ "roid"
+ ],
+ [
+ "andro",
+ "id"
+ ],
+ [
+ "▁sub",
+ "sequ"
+ ],
+ [
+ "▁subs",
+ "equ"
+ ],
+ [
+ "▁Flor",
+ "ida"
+ ],
+ [
+ "sub",
+ "set"
+ ],
+ [
+ "▁whis",
+ "per"
+ ],
+ [
+ "Vo",
+ "l"
+ ],
+ [
+ "V",
+ "ol"
+ ],
+ [
+ "ul",
+ "ous"
+ ],
+ [
+ "ulo",
+ "us"
+ ],
+ [
+ "▁c",
+ "rew"
+ ],
+ [
+ "▁cre",
+ "w"
+ ],
+ [
+ "▁cr",
+ "ew"
+ ],
+ [
+ "▁l",
+ "ug"
+ ],
+ [
+ "▁lu",
+ "g"
+ ],
+ [
+ "pi",
+ "d"
+ ],
+ [
+ "p",
+ "id"
+ ],
+ [
+ "oc",
+ "ity"
+ ],
+ [
+ "oci",
+ "ty"
+ ],
+ [
+ "o",
+ "city"
+ ],
+ [
+ "sk",
+ "b"
+ ],
+ [
+ "s",
+ "kb"
+ ],
+ [
+ "▁t",
+ "ea"
+ ],
+ [
+ "▁te",
+ "a"
+ ],
+ [
+ "у",
+ "н"
+ ],
+ [
+ "▁hon",
+ "or"
+ ],
+ [
+ "▁ho",
+ "nor"
+ ],
+ [
+ "▁I",
+ "ns"
+ ],
+ [
+ "▁In",
+ "s"
+ ],
+ [
+ "▁",
+ "Ins"
+ ],
+ [
+ "▁g",
+ "ew"
+ ],
+ [
+ "▁ge",
+ "w"
+ ],
+ [
+ "▁",
+ "gew"
+ ],
+ [
+ "Det",
+ "ails"
+ ],
+ [
+ "Detail",
+ "s"
+ ],
+ [
+ "ene",
+ "ath"
+ ],
+ [
+ "e",
+ "neath"
+ ],
+ [
+ "at",
+ "ar"
+ ],
+ [
+ "ata",
+ "r"
+ ],
+ [
+ "a",
+ "tar"
+ ],
+ [
+ "▁_",
+ "{"
+ ],
+ [
+ "▁",
+ "_{"
+ ],
+ [
+ "am",
+ "en"
+ ],
+ [
+ "ame",
+ "n"
+ ],
+ [
+ "a",
+ "men"
+ ],
+ [
+ "▁set",
+ "up"
+ ],
+ [
+ "▁",
+ "setup"
+ ],
+ [
+ "Trans",
+ "action"
+ ],
+ [
+ "▁bl",
+ "ank"
+ ],
+ [
+ "▁",
+ "blank"
+ ],
+ [
+ "Fail",
+ "ed"
+ ],
+ [
+ "F",
+ "ailed"
+ ],
+ [
+ "jo",
+ "b"
+ ],
+ [
+ "j",
+ "ob"
+ ],
+ [
+ "▁p",
+ "ret"
+ ],
+ [
+ "▁pre",
+ "t"
+ ],
+ [
+ "▁pr",
+ "et"
+ ],
+ [
+ "▁",
+ "pret"
+ ],
+ [
+ "ß",
+ "e"
+ ],
+ [
+ "lo",
+ "or"
+ ],
+ [
+ "l",
+ "oor"
+ ],
+ [
+ "ř",
+ "í"
+ ],
+ [
+ "nc",
+ "ia"
+ ],
+ [
+ "n",
+ "cia"
+ ],
+ [
+ "▁any",
+ "where"
+ ],
+ [
+ "▁L",
+ "ight"
+ ],
+ [
+ "▁Li",
+ "ght"
+ ],
+ [
+ "▁",
+ "Light"
+ ],
+ [
+ "▁A",
+ "k"
+ ],
+ [
+ "B",
+ "D"
+ ],
+ [
+ "▁exc",
+ "ited"
+ ],
+ [
+ "▁excit",
+ "ed"
+ ],
+ [
+ "ag",
+ "ers"
+ ],
+ [
+ "age",
+ "rs"
+ ],
+ [
+ "ager",
+ "s"
+ ],
+ [
+ "a",
+ "gers"
+ ],
+ [
+ "▁w",
+ "arning"
+ ],
+ [
+ "▁war",
+ "ning"
+ ],
+ [
+ "▁warn",
+ "ing"
+ ],
+ [
+ "▁",
+ "warning"
+ ],
+ [
+ "▁process",
+ "es"
+ ],
+ [
+ "▁proces",
+ "ses"
+ ],
+ [
+ "h",
+ "u"
+ ],
+ [
+ "▁y",
+ "outh"
+ ],
+ [
+ "▁you",
+ "th"
+ ],
+ [
+ "▁yo",
+ "uth"
+ ],
+ [
+ "▁d",
+ "ogs"
+ ],
+ [
+ "▁do",
+ "gs"
+ ],
+ [
+ "▁dog",
+ "s"
+ ],
+ [
+ "▁o",
+ "ct"
+ ],
+ [
+ "▁oc",
+ "t"
+ ],
+ [
+ "▁",
+ "oct"
+ ],
+ [
+ "▁n",
+ "ine"
+ ],
+ [
+ "▁ni",
+ "ne"
+ ],
+ [
+ "▁nin",
+ "e"
+ ],
+ [
+ "Write",
+ "r"
+ ],
+ [
+ "Wr",
+ "iter"
+ ],
+ [
+ "Writ",
+ "er"
+ ],
+ [
+ "W",
+ "riter"
+ ],
+ [
+ "gr",
+ "id"
+ ],
+ [
+ "g",
+ "rid"
+ ],
+ [
+ "▁import",
+ "ance"
+ ],
+ [
+ "est",
+ "ic"
+ ],
+ [
+ "▁care",
+ "fully"
+ ],
+ [
+ "▁careful",
+ "ly"
+ ],
+ [
+ "ma",
+ "ster"
+ ],
+ [
+ "mas",
+ "ter"
+ ],
+ [
+ "m",
+ "aster"
+ ],
+ [
+ "▁dec",
+ "isions"
+ ],
+ [
+ "▁decision",
+ "s"
+ ],
+ [
+ "▁decis",
+ "ions"
+ ],
+ [
+ "▁p",
+ "in"
+ ],
+ [
+ "▁pi",
+ "n"
+ ],
+ [
+ "▁",
+ "pin"
+ ],
+ [
+ "▁cr",
+ "ack"
+ ],
+ [
+ "TE",
+ "ST"
+ ],
+ [
+ "TES",
+ "T"
+ ],
+ [
+ "T",
+ "EST"
+ ],
+ [
+ "▁L",
+ "ocal"
+ ],
+ [
+ "▁Loc",
+ "al"
+ ],
+ [
+ "▁Lo",
+ "cal"
+ ],
+ [
+ "▁",
+ "Local"
+ ],
+ [
+ "▁R",
+ "ight"
+ ],
+ [
+ "▁",
+ "Right"
+ ],
+ [
+ "▁v",
+ "ast"
+ ],
+ [
+ "▁va",
+ "st"
+ ],
+ [
+ "▁vas",
+ "t"
+ ],
+ [
+ "▁f",
+ "aster"
+ ],
+ [
+ "▁fa",
+ "ster"
+ ],
+ [
+ "▁fast",
+ "er"
+ ],
+ [
+ "▁inst",
+ "itut"
+ ],
+ [
+ "▁ann",
+ "ual"
+ ],
+ [
+ "LA",
+ "N"
+ ],
+ [
+ "L",
+ "AN"
+ ],
+ [
+ "▁e",
+ "pisode"
+ ],
+ [
+ "▁epis",
+ "ode"
+ ],
+ [
+ "▁X",
+ "V"
+ ],
+ [
+ "▁del",
+ "ivery"
+ ],
+ [
+ "▁deliver",
+ "y"
+ ],
+ [
+ "t",
+ "l"
+ ],
+ [
+ "F",
+ "P"
+ ],
+ [
+ "ci",
+ "rc"
+ ],
+ [
+ "cir",
+ "c"
+ ],
+ [
+ "▁typ",
+ "ically"
+ ],
+ [
+ "▁typical",
+ "ly"
+ ],
+ [
+ "ig",
+ "o"
+ ],
+ [
+ "i",
+ "go"
+ ],
+ [
+ "▁int",
+ "el"
+ ],
+ [
+ "▁inte",
+ "l"
+ ],
+ [
+ "▁",
+ "intel"
+ ],
+ [
+ "na",
+ "t"
+ ],
+ [
+ "n",
+ "at"
+ ],
+ [
+ "x",
+ "b"
+ ],
+ [
+ "ст",
+ "ро"
+ ],
+ [
+ "с",
+ "тро"
+ ],
+ [
+ ")",
+ "-"
+ ],
+ [
+ "▁B",
+ "al"
+ ],
+ [
+ "▁Ba",
+ "l"
+ ],
+ [
+ "▁",
+ "Bal"
+ ],
+ [
+ "▁J",
+ "os"
+ ],
+ [
+ "▁Jo",
+ "s"
+ ],
+ [
+ "▁g",
+ "onna"
+ ],
+ [
+ "▁R",
+ "est"
+ ],
+ [
+ "▁Re",
+ "st"
+ ],
+ [
+ "▁Res",
+ "t"
+ ],
+ [
+ "▁",
+ "Rest"
+ ],
+ [
+ "jo",
+ "r"
+ ],
+ [
+ "j",
+ "or"
+ ],
+ [
+ "on",
+ "ia"
+ ],
+ [
+ "oni",
+ "a"
+ ],
+ [
+ "o",
+ "nia"
+ ],
+ [
+ "or",
+ "ship"
+ ],
+ [
+ "ors",
+ "hip"
+ ],
+ [
+ "ov",
+ "ery"
+ ],
+ [
+ "ove",
+ "ry"
+ ],
+ [
+ "over",
+ "y"
+ ],
+ [
+ "o",
+ "very"
+ ],
+ [
+ "LI",
+ "NE"
+ ],
+ [
+ "LIN",
+ "E"
+ ],
+ [
+ "L",
+ "INE"
+ ],
+ [
+ "]",
+ ":"
+ ],
+ [
+ "Que",
+ "ue"
+ ],
+ [
+ "▁com",
+ "pare"
+ ],
+ [
+ "▁comp",
+ "are"
+ ],
+ [
+ "▁compar",
+ "e"
+ ],
+ [
+ "▁",
+ "compare"
+ ],
+ [
+ "▁ap",
+ "artment"
+ ],
+ [
+ "▁apart",
+ "ment"
+ ],
+ [
+ "▁r",
+ "ul"
+ ],
+ [
+ "▁ru",
+ "l"
+ ],
+ [
+ "D",
+ "r"
+ ],
+ [
+ "gen",
+ "cy"
+ ],
+ [
+ "g",
+ "ency"
+ ],
+ [
+ "▁ob",
+ "viously"
+ ],
+ [
+ "▁obvious",
+ "ly"
+ ],
+ [
+ "zi",
+ "e"
+ ],
+ [
+ "z",
+ "ie"
+ ],
+ [
+ "yc",
+ "l"
+ ],
+ [
+ "y",
+ "cl"
+ ],
+ [
+ "fort",
+ "unately"
+ ],
+ [
+ "fortun",
+ "ately"
+ ],
+ [
+ "fortunate",
+ "ly"
+ ],
+ [
+ "▁ste",
+ "pped"
+ ],
+ [
+ "▁step",
+ "ped"
+ ],
+ [
+ "▁S",
+ "eg"
+ ],
+ [
+ "▁Se",
+ "g"
+ ],
+ [
+ "▁",
+ "Seg"
+ ],
+ [
+ "▁Wh",
+ "ich"
+ ],
+ [
+ "▁",
+ "Which"
+ ],
+ [
+ "▁P",
+ "C"
+ ],
+ [
+ "▁",
+ "PC"
+ ],
+ [
+ "▁a",
+ "st"
+ ],
+ [
+ "▁as",
+ "t"
+ ],
+ [
+ "▁",
+ "ast"
+ ],
+ [
+ "end",
+ "or"
+ ],
+ [
+ "endo",
+ "r"
+ ],
+ [
+ "▁per",
+ "mission"
+ ],
+ [
+ "▁perm",
+ "ission"
+ ],
+ [
+ "▁",
+ "permission"
+ ],
+ [
+ "CO",
+ "L"
+ ],
+ [
+ "C",
+ "OL"
+ ],
+ [
+ "▁T",
+ "EST"
+ ],
+ [
+ "▁TE",
+ "ST"
+ ],
+ [
+ "▁",
+ "TEST"
+ ],
+ [
+ "P",
+ "ay"
+ ],
+ [
+ "ère",
+ "s"
+ ],
+ [
+ "è",
+ "res"
+ ],
+ [
+ "▁stud",
+ "ied"
+ ],
+ [
+ "▁accom",
+ "pl"
+ ],
+ [
+ "▁accomp",
+ "l"
+ ],
+ [
+ "ro",
+ "le"
+ ],
+ [
+ "rol",
+ "e"
+ ],
+ [
+ "r",
+ "ole"
+ ],
+ [
+ "Wh",
+ "ere"
+ ],
+ [
+ "Whe",
+ "re"
+ ],
+ [
+ "W",
+ "here"
+ ],
+ [
+ "proto",
+ "buf"
+ ],
+ [
+ "met",
+ "adata"
+ ],
+ [
+ "meta",
+ "data"
+ ],
+ [
+ "Jo",
+ "b"
+ ],
+ [
+ "J",
+ "ob"
+ ],
+ [
+ "▁F",
+ "our"
+ ],
+ [
+ "▁Fou",
+ "r"
+ ],
+ [
+ "▁Fo",
+ "ur"
+ ],
+ [
+ "pl",
+ "ements"
+ ],
+ [
+ "ple",
+ "ments"
+ ],
+ [
+ "plement",
+ "s"
+ ],
+ [
+ "dis",
+ "able"
+ ],
+ [
+ "▁l",
+ "oud"
+ ],
+ [
+ "▁lo",
+ "ud"
+ ],
+ [
+ "▁lou",
+ "d"
+ ],
+ [
+ "▁happ",
+ "ening"
+ ],
+ [
+ "▁happen",
+ "ing"
+ ],
+ [
+ "▁U",
+ "sing"
+ ],
+ [
+ "▁Us",
+ "ing"
+ ],
+ [
+ "▁",
+ "Using"
+ ],
+ [
+ "ro",
+ "g"
+ ],
+ [
+ "r",
+ "og"
+ ],
+ [
+ "▁depend",
+ "s"
+ ],
+ [
+ "▁dep",
+ "ends"
+ ],
+ [
+ "í",
+ "m"
+ ],
+ [
+ "'",
+ "\\"
+ ],
+ [
+ "▁t",
+ "aught"
+ ],
+ [
+ "sh",
+ "ared"
+ ],
+ [
+ "sha",
+ "red"
+ ],
+ [
+ "share",
+ "d"
+ ],
+ [
+ "▁att",
+ "ributes"
+ ],
+ [
+ "▁attribute",
+ "s"
+ ],
+ [
+ "▁attribut",
+ "es"
+ ],
+ [
+ "▁",
+ "attributes"
+ ],
+ [
+ "▁A",
+ "ction"
+ ],
+ [
+ "▁Act",
+ "ion"
+ ],
+ [
+ "▁",
+ "Action"
+ ],
+ [
+ "▁d",
+ "ess"
+ ],
+ [
+ "▁de",
+ "ss"
+ ],
+ [
+ "▁des",
+ "s"
+ ],
+ [
+ "▁",
+ "dess"
+ ],
+ [
+ "▁h",
+ "ouses"
+ ],
+ [
+ "▁house",
+ "s"
+ ],
+ [
+ "▁hous",
+ "es"
+ ],
+ [
+ "▁ho",
+ "uses"
+ ],
+ [
+ "▁re",
+ "set"
+ ],
+ [
+ "▁res",
+ "et"
+ ],
+ [
+ "▁",
+ "reset"
+ ],
+ [
+ "▁b",
+ "ien"
+ ],
+ [
+ "▁bi",
+ "en"
+ ],
+ [
+ "▁ex",
+ "plicit"
+ ],
+ [
+ "▁expl",
+ "icit"
+ ],
+ [
+ "LO",
+ "W"
+ ],
+ [
+ "->",
+ "_"
+ ],
+ [
+ "▁P",
+ "M"
+ ],
+ [
+ "▁",
+ "PM"
+ ],
+ [
+ "C",
+ "ategory"
+ ],
+ [
+ "oi",
+ "ce"
+ ],
+ [
+ "o",
+ "ice"
+ ],
+ [
+ "in",
+ "to"
+ ],
+ [
+ "int",
+ "o"
+ ],
+ [
+ "▁m",
+ "ail"
+ ],
+ [
+ "▁ma",
+ "il"
+ ],
+ [
+ "▁mai",
+ "l"
+ ],
+ [
+ "▁",
+ "mail"
+ ],
+ [
+ "▁author",
+ "ity"
+ ],
+ [
+ "▁un",
+ "able"
+ ],
+ [
+ "▁una",
+ "ble"
+ ],
+ [
+ "file",
+ "name"
+ ],
+ [
+ "fil",
+ "ename"
+ ],
+ [
+ "é",
+ "k"
+ ],
+ [
+ "ле",
+ "й"
+ ],
+ [
+ "л",
+ "ей"
+ ],
+ [
+ "▁s",
+ "ector"
+ ],
+ [
+ "▁se",
+ "ctor"
+ ],
+ [
+ "▁sec",
+ "tor"
+ ],
+ [
+ "▁sect",
+ "or"
+ ],
+ [
+ "ap",
+ "point"
+ ],
+ [
+ "app",
+ "oint"
+ ],
+ [
+ "▁h",
+ "ang"
+ ],
+ [
+ "▁ha",
+ "ng"
+ ],
+ [
+ "▁han",
+ "g"
+ ],
+ [
+ "▁",
+ "hang"
+ ],
+ [
+ "▁c",
+ "el"
+ ],
+ [
+ "▁ce",
+ "l"
+ ],
+ [
+ "▁",
+ "cel"
+ ],
+ [
+ "rel",
+ "ated"
+ ],
+ [
+ "it",
+ "ate"
+ ],
+ [
+ "ita",
+ "te"
+ ],
+ [
+ "itat",
+ "e"
+ ],
+ [
+ "▁'",
+ "<"
+ ],
+ [
+ "am",
+ "ber"
+ ],
+ [
+ "amb",
+ "er"
+ ],
+ [
+ "a",
+ "mber"
+ ],
+ [
+ "▁c",
+ "heap"
+ ],
+ [
+ "▁che",
+ "ap"
+ ],
+ [
+ "▁en",
+ "abled"
+ ],
+ [
+ "▁enable",
+ "d"
+ ],
+ [
+ "▁",
+ "enabled"
+ ],
+ [
+ "▁di",
+ "vision"
+ ],
+ [
+ "▁div",
+ "ision"
+ ],
+ [
+ "▁divis",
+ "ion"
+ ],
+ [
+ "An",
+ "y"
+ ],
+ [
+ "A",
+ "ny"
+ ],
+ [
+ "▁h",
+ "ier"
+ ],
+ [
+ "▁hi",
+ "er"
+ ],
+ [
+ "▁H",
+ "ead"
+ ],
+ [
+ "▁He",
+ "ad"
+ ],
+ [
+ "▁",
+ "Head"
+ ],
+ [
+ "nt",
+ "ax"
+ ],
+ [
+ "n",
+ "tax"
+ ],
+ [
+ "ud",
+ "a"
+ ],
+ [
+ "u",
+ "da"
+ ],
+ [
+ "▁lim",
+ "itations"
+ ],
+ [
+ "▁limit",
+ "ations"
+ ],
+ [
+ "▁limitation",
+ "s"
+ ],
+ [
+ "▁st",
+ "udio"
+ ],
+ [
+ "▁stud",
+ "io"
+ ],
+ [
+ "med",
+ "ia"
+ ],
+ [
+ "medi",
+ "a"
+ ],
+ [
+ "m",
+ "edia"
+ ],
+ [
+ "▁cir",
+ "cle"
+ ],
+ [
+ "▁circ",
+ "le"
+ ],
+ [
+ "▁",
+ "circle"
+ ],
+ [
+ "но",
+ "ва"
+ ],
+ [
+ "нов",
+ "а"
+ ],
+ [
+ "▁l",
+ "aug"
+ ],
+ [
+ "▁la",
+ "ug"
+ ],
+ [
+ "ac",
+ "ts"
+ ],
+ [
+ "act",
+ "s"
+ ],
+ [
+ "▁В",
+ "о"
+ ],
+ [
+ "ó",
+ "d"
+ ],
+ [
+ "pl",
+ "ed"
+ ],
+ [
+ "ple",
+ "d"
+ ],
+ [
+ "p",
+ "led"
+ ],
+ [
+ "LO",
+ "C"
+ ],
+ [
+ "L",
+ "OC"
+ ],
+ [
+ "Ex",
+ "pr"
+ ],
+ [
+ "Exp",
+ "r"
+ ],
+ [
+ ">",
+ ":"
+ ],
+ [
+ "▁pr",
+ "és"
+ ],
+ [
+ "▁pré",
+ "s"
+ ],
+ [
+ "▁",
+ "prés"
+ ],
+ [
+ "▁laugh",
+ "ed"
+ ],
+ [
+ "▁laug",
+ "hed"
+ ],
+ [
+ "▁Th",
+ "ree"
+ ],
+ [
+ "▁",
+ "Three"
+ ],
+ [
+ "л",
+ "ы"
+ ],
+ [
+ "▁en",
+ "ds"
+ ],
+ [
+ "▁end",
+ "s"
+ ],
+ [
+ "▁",
+ "ends"
+ ],
+ [
+ "▁fund",
+ "ament"
+ ],
+ [
+ "▁in",
+ "her"
+ ],
+ [
+ "▁",
+ "inher"
+ ],
+ [
+ "▁l",
+ "iv"
+ ],
+ [
+ "▁li",
+ "v"
+ ],
+ [
+ "▁",
+ "liv"
+ ],
+ [
+ "bi",
+ "d"
+ ],
+ [
+ "b",
+ "id"
+ ],
+ [
+ "▁respons",
+ "ibility"
+ ],
+ [
+ "▁check",
+ "ed"
+ ],
+ [
+ "▁",
+ "checked"
+ ],
+ [
+ "▁P",
+ "ac"
+ ],
+ [
+ "▁Pa",
+ "c"
+ ],
+ [
+ "▁f",
+ "ault"
+ ],
+ [
+ "▁fa",
+ "ult"
+ ],
+ [
+ "▁y",
+ "ellow"
+ ],
+ [
+ "▁s",
+ "alt"
+ ],
+ [
+ "▁sa",
+ "lt"
+ ],
+ [
+ "▁sal",
+ "t"
+ ],
+ [
+ "▁Franc",
+ "isco"
+ ],
+ [
+ "▁Francis",
+ "co"
+ ],
+ [
+ "▁",
+ "^"
+ ],
+ [
+ "▁O",
+ "N"
+ ],
+ [
+ "▁",
+ "ON"
+ ],
+ [
+ "▁beaut",
+ "y"
+ ],
+ [
+ "y",
+ "g"
+ ],
+ [
+ "▁A",
+ "ff"
+ ],
+ [
+ "▁Af",
+ "f"
+ ],
+ [
+ "▁",
+ "Aff"
+ ],
+ [
+ "▁E",
+ "q"
+ ],
+ [
+ "▁",
+ "Eq"
+ ],
+ [
+ "▁mag",
+ "ic"
+ ],
+ [
+ "▁hand",
+ "ler"
+ ],
+ [
+ "▁handle",
+ "r"
+ ],
+ [
+ "▁",
+ "handler"
+ ],
+ [
+ "x",
+ "E"
+ ],
+ [
+ "▁numer",
+ "ous"
+ ],
+ [
+ "▁numero",
+ "us"
+ ],
+ [
+ "▁h",
+ "ole"
+ ],
+ [
+ "▁hol",
+ "e"
+ ],
+ [
+ "▁ho",
+ "le"
+ ],
+ [
+ "▁",
+ "hole"
+ ],
+ [
+ "▁ro",
+ "oms"
+ ],
+ [
+ "▁room",
+ "s"
+ ],
+ [
+ "▁",
+ "rooms"
+ ],
+ [
+ "cc",
+ "ión"
+ ],
+ [
+ "cció",
+ "n"
+ ],
+ [
+ "c",
+ "ción"
+ ],
+ [
+ "▁A",
+ "rm"
+ ],
+ [
+ "▁Ar",
+ "m"
+ ],
+ [
+ "▁",
+ "Arm"
+ ],
+ [
+ "per",
+ "son"
+ ],
+ [
+ "pers",
+ "on"
+ ],
+ [
+ "p",
+ "erson"
+ ],
+ [
+ "▁build",
+ "ings"
+ ],
+ [
+ "▁building",
+ "s"
+ ],
+ [
+ "▁p",
+ "late"
+ ],
+ [
+ "▁pl",
+ "ate"
+ ],
+ [
+ "▁plat",
+ "e"
+ ],
+ [
+ "ble",
+ "d"
+ ],
+ [
+ "bl",
+ "ed"
+ ],
+ [
+ "b",
+ "led"
+ ],
+ [
+ "er",
+ "rors"
+ ],
+ [
+ "err",
+ "ors"
+ ],
+ [
+ "error",
+ "s"
+ ],
+ [
+ "▁A",
+ "gain"
+ ],
+ [
+ "▁Ag",
+ "ain"
+ ],
+ [
+ "▁Def",
+ "ault"
+ ],
+ [
+ "▁",
+ "Default"
+ ],
+ [
+ "▁H",
+ "ard"
+ ],
+ [
+ "▁Har",
+ "d"
+ ],
+ [
+ "▁Ha",
+ "rd"
+ ],
+ [
+ "▁",
+ "Hard"
+ ],
+ [
+ "t",
+ "ó"
+ ],
+ [
+ "hu",
+ "s"
+ ],
+ [
+ "h",
+ "us"
+ ],
+ [
+ "▁dim",
+ "ension"
+ ],
+ [
+ "ial",
+ "e"
+ ],
+ [
+ "ia",
+ "le"
+ ],
+ [
+ "i",
+ "ale"
+ ],
+ [
+ "▁M",
+ "ult"
+ ],
+ [
+ "▁Mu",
+ "lt"
+ ],
+ [
+ "▁Mul",
+ "t"
+ ],
+ [
+ "▁",
+ "Mult"
+ ],
+ [
+ "▁Govern",
+ "ment"
+ ],
+ [
+ "Fun",
+ "c"
+ ],
+ [
+ "F",
+ "unc"
+ ],
+ [
+ "▁b",
+ "low"
+ ],
+ [
+ "▁bl",
+ "ow"
+ ],
+ [
+ "▁blo",
+ "w"
+ ],
+ [
+ "▁re",
+ "ct"
+ ],
+ [
+ "▁r",
+ "ect"
+ ],
+ [
+ "▁rec",
+ "t"
+ ],
+ [
+ "▁",
+ "rect"
+ ],
+ [
+ "er",
+ "ra"
+ ],
+ [
+ "err",
+ "a"
+ ],
+ [
+ "conne",
+ "ction"
+ ],
+ [
+ "connect",
+ "ion"
+ ],
+ [
+ "conn",
+ "ection"
+ ],
+ [
+ "▁pass",
+ "ing"
+ ],
+ [
+ "▁pas",
+ "sing"
+ ],
+ [
+ "ße",
+ "n"
+ ],
+ [
+ "ß",
+ "en"
+ ],
+ [
+ "ph",
+ "as"
+ ],
+ [
+ "pha",
+ "s"
+ ],
+ [
+ "p",
+ "has"
+ ],
+ [
+ "ens",
+ "ional"
+ ],
+ [
+ "ension",
+ "al"
+ ],
+ [
+ "re",
+ "cord"
+ ],
+ [
+ "rec",
+ "ord"
+ ],
+ [
+ "co",
+ "hol"
+ ],
+ [
+ "▁H",
+ "arry"
+ ],
+ [
+ "▁Har",
+ "ry"
+ ],
+ [
+ "▁Harr",
+ "y"
+ ],
+ [
+ "izont",
+ "al"
+ ],
+ [
+ "izon",
+ "tal"
+ ],
+ [
+ "▁f",
+ "inger"
+ ],
+ [
+ "▁fin",
+ "ger"
+ ],
+ [
+ "▁fing",
+ "er"
+ ],
+ [
+ "▁young",
+ "er"
+ ],
+ [
+ "▁S",
+ "C"
+ ],
+ [
+ "▁",
+ "SC"
+ ],
+ [
+ "oper",
+ "ation"
+ ],
+ [
+ "B",
+ "Y"
+ ],
+ [
+ "he",
+ "im"
+ ],
+ [
+ "▁B",
+ "ad"
+ ],
+ [
+ "▁Ba",
+ "d"
+ ],
+ [
+ "▁",
+ "Bad"
+ ],
+ [
+ "▁st",
+ "orm"
+ ],
+ [
+ "▁stor",
+ "m"
+ ],
+ [
+ "▁sto",
+ "rm"
+ ],
+ [
+ "▁",
+ "storm"
+ ],
+ [
+ "▁N",
+ "at"
+ ],
+ [
+ "▁Na",
+ "t"
+ ],
+ [
+ "▁bu",
+ "ying"
+ ],
+ [
+ "▁buy",
+ "ing"
+ ],
+ [
+ "▁S",
+ "ometimes"
+ ],
+ [
+ "▁Some",
+ "times"
+ ],
+ [
+ "▁С",
+ "та"
+ ],
+ [
+ "es",
+ "sed"
+ ],
+ [
+ "ess",
+ "ed"
+ ],
+ [
+ "esse",
+ "d"
+ ],
+ [
+ "▁da",
+ "mn"
+ ],
+ [
+ "▁dam",
+ "n"
+ ],
+ [
+ "▁m",
+ "eg"
+ ],
+ [
+ "▁me",
+ "g"
+ ],
+ [
+ "um",
+ "es"
+ ],
+ [
+ "ume",
+ "s"
+ ],
+ [
+ "u",
+ "mes"
+ ],
+ [
+ "ün",
+ "d"
+ ],
+ [
+ "ü",
+ "nd"
+ ],
+ [
+ "т",
+ "ра"
+ ],
+ [
+ "▁sil",
+ "ver"
+ ],
+ [
+ "w",
+ "d"
+ ],
+ [
+ "hid",
+ "den"
+ ],
+ [
+ "h",
+ "idden"
+ ],
+ [
+ "ar",
+ "do"
+ ],
+ [
+ "ard",
+ "o"
+ ],
+ [
+ "▁commun",
+ "ities"
+ ],
+ [
+ "▁d",
+ "iet"
+ ],
+ [
+ "▁di",
+ "et"
+ ],
+ [
+ "▁die",
+ "t"
+ ],
+ [
+ "ot",
+ "ted"
+ ],
+ [
+ "ott",
+ "ed"
+ ],
+ [
+ "otte",
+ "d"
+ ],
+ [
+ "▁b",
+ "at"
+ ],
+ [
+ "▁ba",
+ "t"
+ ],
+ [
+ "▁",
+ "bat"
+ ],
+ [
+ "an",
+ "cer"
+ ],
+ [
+ "ance",
+ "r"
+ ],
+ [
+ "anc",
+ "er"
+ ],
+ [
+ "▁f",
+ "mt"
+ ],
+ [
+ "▁",
+ "fmt"
+ ],
+ [
+ "▁P",
+ "en"
+ ],
+ [
+ "▁Pe",
+ "n"
+ ],
+ [
+ "▁",
+ "Pen"
+ ],
+ [
+ "▁t",
+ "il"
+ ],
+ [
+ "▁ti",
+ "l"
+ ],
+ [
+ "▁",
+ "til"
+ ],
+ [
+ "En",
+ "um"
+ ],
+ [
+ "E",
+ "num"
+ ],
+ [
+ "PA",
+ "TH"
+ ],
+ [
+ "P",
+ "ATH"
+ ],
+ [
+ "▁mat",
+ "ters"
+ ],
+ [
+ "▁matter",
+ "s"
+ ],
+ [
+ "▁matt",
+ "ers"
+ ],
+ [
+ "time",
+ "out"
+ ],
+ [
+ "--",
+ "----------"
+ ],
+ [
+ "----",
+ "--------"
+ ],
+ [
+ "--------",
+ "----"
+ ],
+ [
+ "---",
+ "---------"
+ ],
+ [
+ "-----",
+ "-------"
+ ],
+ [
+ "----------",
+ "--"
+ ],
+ [
+ "------",
+ "------"
+ ],
+ [
+ "---------",
+ "---"
+ ],
+ [
+ "-------",
+ "-----"
+ ],
+ [
+ "-----------",
+ "-"
+ ],
+ [
+ "-",
+ "-----------"
+ ],
+ [
+ "ka",
+ "n"
+ ],
+ [
+ "k",
+ "an"
+ ],
+ [
+ "▁Cor",
+ "por"
+ ],
+ [
+ "=\"",
+ "../../"
+ ],
+ [
+ "=\"../",
+ "../"
+ ],
+ [
+ "▁A",
+ "le"
+ ],
+ [
+ "▁Al",
+ "e"
+ ],
+ [
+ "hent",
+ "ication"
+ ],
+ [
+ "hentic",
+ "ation"
+ ],
+ [
+ "▁com",
+ "plic"
+ ],
+ [
+ "▁comp",
+ "lic"
+ ],
+ [
+ "▁compl",
+ "ic"
+ ],
+ [
+ "▁Se",
+ "curity"
+ ],
+ [
+ "▁Sec",
+ "urity"
+ ],
+ [
+ "▁",
+ "Security"
+ ],
+ [
+ "OF",
+ "F"
+ ],
+ [
+ "O",
+ "FF"
+ ],
+ [
+ "R",
+ "ad"
+ ],
+ [
+ "ap",
+ "se"
+ ],
+ [
+ "aps",
+ "e"
+ ],
+ [
+ "a",
+ "pse"
+ ],
+ [
+ "▁d",
+ "ance"
+ ],
+ [
+ "▁dan",
+ "ce"
+ ],
+ [
+ "▁perm",
+ "issions"
+ ],
+ [
+ "▁permission",
+ "s"
+ ],
+ [
+ "▁war",
+ "rant"
+ ],
+ [
+ "▁l",
+ "ad"
+ ],
+ [
+ "▁la",
+ "d"
+ ],
+ [
+ "▁",
+ "lad"
+ ],
+ [
+ "▁is",
+ "ol"
+ ],
+ [
+ "▁i",
+ "sol"
+ ],
+ [
+ "d",
+ "l"
+ ],
+ [
+ "▁A",
+ "u"
+ ],
+ [
+ "ye",
+ "s"
+ ],
+ [
+ "y",
+ "es"
+ ],
+ [
+ "▁t",
+ "v"
+ ],
+ [
+ "▁",
+ "tv"
+ ],
+ [
+ "▁pro",
+ "vider"
+ ],
+ [
+ "▁prov",
+ "ider"
+ ],
+ [
+ "▁provide",
+ "r"
+ ],
+ [
+ "▁",
+ "provider"
+ ],
+ [
+ "▁ter",
+ "rible"
+ ],
+ [
+ "▁terr",
+ "ible"
+ ],
+ [
+ "▁dep",
+ "artment"
+ ],
+ [
+ "▁depart",
+ "ment"
+ ],
+ [
+ "er",
+ "al"
+ ],
+ [
+ "era",
+ "l"
+ ],
+ [
+ "e",
+ "ral"
+ ],
+ [
+ "▁implement",
+ "ation"
+ ],
+ [
+ "S",
+ "R"
+ ],
+ [
+ "▁h",
+ "earing"
+ ],
+ [
+ "▁he",
+ "aring"
+ ],
+ [
+ "▁hear",
+ "ing"
+ ],
+ [
+ "▁K",
+ "n"
+ ],
+ [
+ "F",
+ "R"
+ ],
+ [
+ "t",
+ "v"
+ ],
+ [
+ "▁d",
+ "iss"
+ ],
+ [
+ "▁dis",
+ "s"
+ ],
+ [
+ "▁di",
+ "ss"
+ ],
+ [
+ "F",
+ "UN"
+ ],
+ [
+ "▁dur",
+ "ante"
+ ],
+ [
+ "▁durant",
+ "e"
+ ],
+ [
+ "os",
+ "is"
+ ],
+ [
+ "osi",
+ "s"
+ ],
+ [
+ "o",
+ "sis"
+ ],
+ [
+ "▁task",
+ "s"
+ ],
+ [
+ "▁",
+ "tasks"
+ ],
+ [
+ "▁B",
+ "lo"
+ ],
+ [
+ "▁Bl",
+ "o"
+ ],
+ [
+ "▁",
+ "Blo"
+ ],
+ [
+ "во",
+ "д"
+ ],
+ [
+ "▁br",
+ "anch"
+ ],
+ [
+ "▁",
+ "branch"
+ ],
+ [
+ "▁polit",
+ "ics"
+ ],
+ [
+ "▁E",
+ "lle"
+ ],
+ [
+ "▁El",
+ "le"
+ ],
+ [
+ "▁Ell",
+ "e"
+ ],
+ [
+ "▁lead",
+ "ership"
+ ],
+ [
+ "▁leader",
+ "ship"
+ ],
+ [
+ "▁leaders",
+ "hip"
+ ],
+ [
+ "ex",
+ "pr"
+ ],
+ [
+ "exp",
+ "r"
+ ],
+ [
+ "▁techn",
+ "iques"
+ ],
+ [
+ "▁technique",
+ "s"
+ ],
+ [
+ "pr",
+ "ec"
+ ],
+ [
+ "pre",
+ "c"
+ ],
+ [
+ "p",
+ "rec"
+ ],
+ [
+ "Sig",
+ "ma"
+ ],
+ [
+ "S",
+ "igma"
+ ],
+ [
+ "im",
+ "ately"
+ ],
+ [
+ "imate",
+ "ly"
+ ],
+ [
+ "imat",
+ "ely"
+ ],
+ [
+ "t",
+ "k"
+ ],
+ [
+ "ach",
+ "ment"
+ ],
+ [
+ "▁En",
+ "ter"
+ ],
+ [
+ "▁Ent",
+ "er"
+ ],
+ [
+ "▁",
+ "Enter"
+ ],
+ [
+ "▁cre",
+ "ative"
+ ],
+ [
+ "▁creat",
+ "ive"
+ ],
+ [
+ "▁з",
+ "на"
+ ],
+ [
+ "▁",
+ "зна"
+ ],
+ [
+ "ap",
+ "py"
+ ],
+ [
+ "app",
+ "y"
+ ],
+ [
+ "un",
+ "ched"
+ ],
+ [
+ "unch",
+ "ed"
+ ],
+ [
+ "unc",
+ "hed"
+ ],
+ [
+ "▁'",
+ "',"
+ ],
+ [
+ "▁''",
+ ","
+ ],
+ [
+ "on",
+ "der"
+ ],
+ [
+ "ond",
+ "er"
+ ],
+ [
+ "onde",
+ "r"
+ ],
+ [
+ "o",
+ "nder"
+ ],
+ [
+ "{",
+ "-"
+ ],
+ [
+ "NU",
+ "M"
+ ],
+ [
+ "N",
+ "UM"
+ ],
+ [
+ "▁n",
+ "arr"
+ ],
+ [
+ "▁na",
+ "rr"
+ ],
+ [
+ "▁nar",
+ "r"
+ ],
+ [
+ "Mem",
+ "ory"
+ ],
+ [
+ "▁win",
+ "ning"
+ ],
+ [
+ "▁",
+ "winning"
+ ],
+ [
+ "▁F",
+ "ollow"
+ ],
+ [
+ "▁Fol",
+ "low"
+ ],
+ [
+ "▁",
+ "Follow"
+ ],
+ [
+ "*/",
+ "\r"
+ ],
+ [
+ "vis",
+ "ion"
+ ],
+ [
+ "v",
+ "ision"
+ ],
+ [
+ "res",
+ "ents"
+ ],
+ [
+ "resent",
+ "s"
+ ],
+ [
+ "zi",
+ "one"
+ ],
+ [
+ "z",
+ "ione"
+ ],
+ [
+ "▁l",
+ "atter"
+ ],
+ [
+ "▁lat",
+ "ter"
+ ],
+ [
+ "▁requ",
+ "ests"
+ ],
+ [
+ "▁request",
+ "s"
+ ],
+ [
+ "▁",
+ "requests"
+ ],
+ [
+ "▁m",
+ "argin"
+ ],
+ [
+ "▁mar",
+ "gin"
+ ],
+ [
+ "▁marg",
+ "in"
+ ],
+ [
+ "▁",
+ "margin"
+ ],
+ [
+ "▁{",
+ "\""
+ ],
+ [
+ "▁",
+ "{\""
+ ],
+ [
+ "v",
+ "ideo"
+ ],
+ [
+ "c",
+ "n"
+ ],
+ [
+ "▁Im",
+ "age"
+ ],
+ [
+ "▁",
+ "Image"
+ ],
+ [
+ "T",
+ "im"
+ ],
+ [
+ "CON",
+ "FIG"
+ ],
+ [
+ "CONF",
+ "IG"
+ ],
+ [
+ "▁all",
+ "owing"
+ ],
+ [
+ "▁allow",
+ "ing"
+ ],
+ [
+ "▁comb",
+ "ined"
+ ],
+ [
+ "▁combine",
+ "d"
+ ],
+ [
+ "PU",
+ "T"
+ ],
+ [
+ "P",
+ "UT"
+ ],
+ [
+ "▁instance",
+ "of"
+ ],
+ [
+ "ig",
+ "in"
+ ],
+ [
+ "igi",
+ "n"
+ ],
+ [
+ "i",
+ "gin"
+ ],
+ [
+ "▁p",
+ "ero"
+ ],
+ [
+ "▁per",
+ "o"
+ ],
+ [
+ "▁pe",
+ "ro"
+ ],
+ [
+ "▁'",
+ "'"
+ ],
+ [
+ "▁",
+ "''"
+ ],
+ [
+ "▁conf",
+ "idence"
+ ],
+ [
+ "▁equ",
+ "ivalent"
+ ],
+ [
+ "▁equival",
+ "ent"
+ ],
+ [
+ "pa",
+ "d"
+ ],
+ [
+ "p",
+ "ad"
+ ],
+ [
+ "ef",
+ "fect"
+ ],
+ [
+ "eff",
+ "ect"
+ ],
+ [
+ "e",
+ "ffect"
+ ],
+ [
+ "R",
+ "X"
+ ],
+ [
+ "▁l",
+ "ang"
+ ],
+ [
+ "▁la",
+ "ng"
+ ],
+ [
+ "▁lan",
+ "g"
+ ],
+ [
+ "▁",
+ "lang"
+ ],
+ [
+ "str",
+ "ong"
+ ],
+ [
+ "▁b",
+ "ridge"
+ ],
+ [
+ "▁br",
+ "idge"
+ ],
+ [
+ "▁",
+ "bridge"
+ ],
+ [
+ "ay",
+ "a"
+ ],
+ [
+ "a",
+ "ya"
+ ],
+ [
+ "▁t",
+ "reated"
+ ],
+ [
+ "▁tre",
+ "ated"
+ ],
+ [
+ "▁treat",
+ "ed"
+ ],
+ [
+ "▁f",
+ "orth"
+ ],
+ [
+ "▁for",
+ "th"
+ ],
+ [
+ "▁fort",
+ "h"
+ ],
+ [
+ "S",
+ "W"
+ ],
+ [
+ "▁account",
+ "s"
+ ],
+ [
+ "▁P",
+ "O"
+ ],
+ [
+ "▁",
+ "PO"
+ ],
+ [
+ "▁list",
+ "ening"
+ ],
+ [
+ "▁listen",
+ "ing"
+ ],
+ [
+ "Ro",
+ "ute"
+ ],
+ [
+ "R",
+ "oute"
+ ],
+ [
+ "()",
+ "))"
+ ],
+ [
+ "())",
+ ")"
+ ],
+ [
+ "(",
+ ")))"
+ ],
+ [
+ "cp",
+ "y"
+ ],
+ [
+ "c",
+ "py"
+ ],
+ [
+ "▁re",
+ "form"
+ ],
+ [
+ "▁ref",
+ "orm"
+ ],
+ [
+ "▁g",
+ "ate"
+ ],
+ [
+ "▁ga",
+ "te"
+ ],
+ [
+ "▁",
+ "gate"
+ ],
+ [
+ "▁W",
+ "alk"
+ ],
+ [
+ "▁Wal",
+ "k"
+ ],
+ [
+ "▁",
+ "Walk"
+ ],
+ [
+ "▁some",
+ "how"
+ ],
+ [
+ "t",
+ "f"
+ ],
+ [
+ "▁l",
+ "ayout"
+ ],
+ [
+ "▁la",
+ "yout"
+ ],
+ [
+ "▁lay",
+ "out"
+ ],
+ [
+ "▁",
+ "layout"
+ ],
+ [
+ "um",
+ "in"
+ ],
+ [
+ "umi",
+ "n"
+ ],
+ [
+ "u",
+ "min"
+ ],
+ [
+ "▁consider",
+ "ing"
+ ],
+ [
+ "▁consid",
+ "ering"
+ ],
+ [
+ "▁pre",
+ "mi"
+ ],
+ [
+ "▁pr",
+ "emi"
+ ],
+ [
+ "▁prem",
+ "i"
+ ],
+ [
+ "▁M",
+ "om"
+ ],
+ [
+ "▁Mo",
+ "m"
+ ],
+ [
+ "at",
+ "han"
+ ],
+ [
+ "ath",
+ "an"
+ ],
+ [
+ "a",
+ "than"
+ ],
+ [
+ "Ge",
+ "n"
+ ],
+ [
+ "G",
+ "en"
+ ],
+ [
+ "▁plan",
+ "et"
+ ],
+ [
+ "▁plane",
+ "t"
+ ],
+ [
+ "am",
+ "ples"
+ ],
+ [
+ "amp",
+ "les"
+ ],
+ [
+ "ample",
+ "s"
+ ],
+ [
+ "▁M",
+ "O"
+ ],
+ [
+ "▁",
+ "MO"
+ ],
+ [
+ "sh",
+ "op"
+ ],
+ [
+ "s",
+ "hop"
+ ],
+ [
+ "▁prem",
+ "ier"
+ ],
+ [
+ "▁premi",
+ "er"
+ ],
+ [
+ "▁s",
+ "impl"
+ ],
+ [
+ "▁sim",
+ "pl"
+ ],
+ [
+ "▁s",
+ "egu"
+ ],
+ [
+ "▁se",
+ "gu"
+ ],
+ [
+ "▁seg",
+ "u"
+ ],
+ [
+ "L",
+ "Y"
+ ],
+ [
+ "Su",
+ "m"
+ ],
+ [
+ "S",
+ "um"
+ ],
+ [
+ "▁t",
+ "ables"
+ ],
+ [
+ "▁table",
+ "s"
+ ],
+ [
+ "▁tab",
+ "les"
+ ],
+ [
+ "▁ta",
+ "bles"
+ ],
+ [
+ "▁",
+ "tables"
+ ],
+ [
+ "sk",
+ "a"
+ ],
+ [
+ "s",
+ "ka"
+ ],
+ [
+ "▁",
+ "ž"
+ ],
+ [
+ "p",
+ "d"
+ ],
+ [
+ "▁s",
+ "ous"
+ ],
+ [
+ "▁so",
+ "us"
+ ],
+ [
+ "▁sou",
+ "s"
+ ],
+ [
+ "▁con",
+ "ference"
+ ],
+ [
+ "▁confer",
+ "ence"
+ ],
+ [
+ "▁D",
+ "at"
+ ],
+ [
+ "▁Da",
+ "t"
+ ],
+ [
+ "▁",
+ "Dat"
+ ],
+ [
+ "Sc",
+ "roll"
+ ],
+ [
+ "▁stand",
+ "ards"
+ ],
+ [
+ "▁standard",
+ "s"
+ ],
+ [
+ "▁г",
+ "ру"
+ ],
+ [
+ "es",
+ "se"
+ ],
+ [
+ "ess",
+ "e"
+ ],
+ [
+ "▁citiz",
+ "ens"
+ ],
+ [
+ "▁citizen",
+ "s"
+ ],
+ [
+ "▁occur",
+ "red"
+ ],
+ [
+ "▁dem",
+ "ocr"
+ ],
+ [
+ "▁demo",
+ "cr"
+ ],
+ [
+ "▁e",
+ "lev"
+ ],
+ [
+ "▁el",
+ "ev"
+ ],
+ [
+ "▁ele",
+ "v"
+ ],
+ [
+ "▁S",
+ "em"
+ ],
+ [
+ "▁Se",
+ "m"
+ ],
+ [
+ "▁",
+ "Sem"
+ ],
+ [
+ "ens",
+ "us"
+ ],
+ [
+ "he",
+ "aders"
+ ],
+ [
+ "head",
+ "ers"
+ ],
+ [
+ "header",
+ "s"
+ ],
+ [
+ "▁Ch",
+ "ris"
+ ],
+ [
+ "im",
+ "ento"
+ ],
+ [
+ "iment",
+ "o"
+ ],
+ [
+ "imen",
+ "to"
+ ],
+ [
+ "ko",
+ "m"
+ ],
+ [
+ "k",
+ "om"
+ ],
+ [
+ "Co",
+ "r"
+ ],
+ [
+ "C",
+ "or"
+ ],
+ [
+ "MI",
+ "N"
+ ],
+ [
+ "M",
+ "IN"
+ ],
+ [
+ "us",
+ "her"
+ ],
+ [
+ "ush",
+ "er"
+ ],
+ [
+ "Data",
+ "base"
+ ],
+ [
+ "Dat",
+ "abase"
+ ],
+ [
+ "▁f",
+ "ormal"
+ ],
+ [
+ "▁for",
+ "mal"
+ ],
+ [
+ "▁form",
+ "al"
+ ],
+ [
+ "▁forma",
+ "l"
+ ],
+ [
+ "ig",
+ "ne"
+ ],
+ [
+ "ign",
+ "e"
+ ],
+ [
+ "▁organ",
+ "izations"
+ ],
+ [
+ "▁organiz",
+ "ations"
+ ],
+ [
+ "▁organization",
+ "s"
+ ],
+ [
+ "▁I",
+ "re"
+ ],
+ [
+ "▁Ir",
+ "e"
+ ],
+ [
+ "X",
+ "ml"
+ ],
+ [
+ "и",
+ "з"
+ ],
+ [
+ "▁p",
+ "ray"
+ ],
+ [
+ "▁pr",
+ "ay"
+ ],
+ [
+ "▁pra",
+ "y"
+ ],
+ [
+ "▁b",
+ "omb"
+ ],
+ [
+ "▁bo",
+ "mb"
+ ],
+ [
+ "▁bom",
+ "b"
+ ],
+ [
+ "▁m",
+ "and"
+ ],
+ [
+ "▁man",
+ "d"
+ ],
+ [
+ "▁ma",
+ "nd"
+ ],
+ [
+ "▁",
+ "mand"
+ ],
+ [
+ "er",
+ "ts"
+ ],
+ [
+ "ert",
+ "s"
+ ],
+ [
+ "▁c",
+ "lock"
+ ],
+ [
+ "▁cl",
+ "ock"
+ ],
+ [
+ "▁clo",
+ "ck"
+ ],
+ [
+ "▁",
+ "clock"
+ ],
+ [
+ "▁b",
+ "uck"
+ ],
+ [
+ "▁bu",
+ "ck"
+ ],
+ [
+ "ва",
+ "ли"
+ ],
+ [
+ "вал",
+ "и"
+ ],
+ [
+ "в",
+ "али"
+ ],
+ [
+ "en",
+ "sch"
+ ],
+ [
+ "ens",
+ "ch"
+ ],
+ [
+ "▁v",
+ "olt"
+ ],
+ [
+ "▁vo",
+ "lt"
+ ],
+ [
+ "▁vol",
+ "t"
+ ],
+ [
+ "▁",
+ "volt"
+ ],
+ [
+ "▁fil",
+ "ms"
+ ],
+ [
+ "▁film",
+ "s"
+ ],
+ [
+ "▁pl",
+ "ants"
+ ],
+ [
+ "▁plan",
+ "ts"
+ ],
+ [
+ "▁plant",
+ "s"
+ ],
+ [
+ "in",
+ "ode"
+ ],
+ [
+ "ino",
+ "de"
+ ],
+ [
+ "i",
+ "node"
+ ],
+ [
+ "Bo",
+ "olean"
+ ],
+ [
+ "▁restaur",
+ "ant"
+ ],
+ [
+ "ía",
+ "n"
+ ],
+ [
+ "í",
+ "an"
+ ],
+ [
+ "▁de",
+ "but"
+ ],
+ [
+ "▁deb",
+ "ut"
+ ],
+ [
+ "page",
+ "s"
+ ],
+ [
+ "pa",
+ "ges"
+ ],
+ [
+ "pag",
+ "es"
+ ],
+ [
+ "p",
+ "ages"
+ ],
+ [
+ "▁wor",
+ "dt"
+ ],
+ [
+ "▁word",
+ "t"
+ ],
+ [
+ "▁Б",
+ "а"
+ ],
+ [
+ "▁great",
+ "est"
+ ],
+ [
+ "(\"",
+ "/"
+ ],
+ [
+ "▁c",
+ "opyright"
+ ],
+ [
+ "▁copy",
+ "right"
+ ],
+ [
+ "▁",
+ "copyright"
+ ],
+ [
+ "▁r",
+ "it"
+ ],
+ [
+ "▁ri",
+ "t"
+ ],
+ [
+ "▁",
+ "rit"
+ ],
+ [
+ "size",
+ "of"
+ ],
+ [
+ "Tr",
+ "ace"
+ ],
+ [
+ "Tra",
+ "ce"
+ ],
+ [
+ "ue",
+ "nt"
+ ],
+ [
+ "uen",
+ "t"
+ ],
+ [
+ "u",
+ "ent"
+ ],
+ [
+ "ту",
+ "р"
+ ],
+ [
+ "т",
+ "ур"
+ ],
+ [
+ "▁k",
+ "o"
+ ],
+ [
+ "▁",
+ "ko"
+ ],
+ [
+ ":",
+ "\\"
+ ],
+ [
+ "▁b",
+ "igger"
+ ],
+ [
+ "▁big",
+ "ger"
+ ],
+ [
+ "▁perfect",
+ "ly"
+ ],
+ [
+ "ten",
+ "ance"
+ ],
+ [
+ "MA",
+ "SK"
+ ],
+ [
+ "M",
+ "ASK"
+ ],
+ [
+ "r",
+ "é"
+ ],
+ [
+ "▁e",
+ "tt"
+ ],
+ [
+ "▁et",
+ "t"
+ ],
+ [
+ "▁",
+ "ett"
+ ],
+ [
+ "▁n",
+ "ose"
+ ],
+ [
+ "▁no",
+ "se"
+ ],
+ [
+ "▁nos",
+ "e"
+ ],
+ [
+ "▁c",
+ "raft"
+ ],
+ [
+ "▁cr",
+ "aft"
+ ],
+ [
+ "▁",
+ "craft"
+ ],
+ [
+ "it",
+ "eral"
+ ],
+ [
+ "ite",
+ "ral"
+ ],
+ [
+ "iter",
+ "al"
+ ],
+ [
+ "▁discuss",
+ "ed"
+ ],
+ [
+ "▁Jew",
+ "ish"
+ ],
+ [
+ "C",
+ "ap"
+ ],
+ [
+ "▁Un",
+ "less"
+ ],
+ [
+ "▁Jack",
+ "son"
+ ],
+ [
+ "Att",
+ "ributes"
+ ],
+ [
+ "Attribute",
+ "s"
+ ],
+ [
+ "Attrib",
+ "utes"
+ ],
+ [
+ "▁l",
+ "unch"
+ ],
+ [
+ "▁lun",
+ "ch"
+ ],
+ [
+ "ö",
+ "l"
+ ],
+ [
+ "at",
+ "r"
+ ],
+ [
+ "a",
+ "tr"
+ ],
+ [
+ "▁pay",
+ "ing"
+ ],
+ [
+ "▁pa",
+ "ying"
+ ],
+ [
+ "Par",
+ "se"
+ ],
+ [
+ "Pars",
+ "e"
+ ],
+ [
+ "P",
+ "arse"
+ ],
+ [
+ "()",
+ "\r"
+ ],
+ [
+ "(",
+ ")\r"
+ ],
+ [
+ "la",
+ "d"
+ ],
+ [
+ "l",
+ "ad"
+ ],
+ [
+ "▁r",
+ "are"
+ ],
+ [
+ "▁ra",
+ "re"
+ ],
+ [
+ "▁[",
+ "];"
+ ],
+ [
+ "▁[]",
+ ";"
+ ],
+ [
+ "▁",
+ "[];"
+ ],
+ [
+ "st",
+ "one"
+ ],
+ [
+ "ston",
+ "e"
+ ],
+ [
+ "sto",
+ "ne"
+ ],
+ [
+ "▁u",
+ "nc"
+ ],
+ [
+ "▁un",
+ "c"
+ ],
+ [
+ "▁",
+ "unc"
+ ],
+ [
+ "▁def",
+ "ense"
+ ],
+ [
+ "▁defens",
+ "e"
+ ],
+ [
+ "}",
+ "+"
+ ],
+ [
+ "▁Gl",
+ "obal"
+ ],
+ [
+ "▁",
+ "Global"
+ ],
+ [
+ "▁Sov",
+ "iet"
+ ],
+ [
+ "▁Austral",
+ "ian"
+ ],
+ [
+ "▁Australia",
+ "n"
+ ],
+ [
+ "▁g",
+ "li"
+ ],
+ [
+ "▁gl",
+ "i"
+ ],
+ [
+ "var",
+ "iant"
+ ],
+ [
+ "vari",
+ "ant"
+ ],
+ [
+ "▁R",
+ "on"
+ ],
+ [
+ "▁Ro",
+ "n"
+ ],
+ [
+ "▁lo",
+ "an"
+ ],
+ [
+ "St",
+ "ep"
+ ],
+ [
+ "Ste",
+ "p"
+ ],
+ [
+ "me",
+ "mber"
+ ],
+ [
+ "mem",
+ "ber"
+ ],
+ [
+ "m",
+ "ember"
+ ],
+ [
+ "Sc",
+ "h"
+ ],
+ [
+ "S",
+ "ch"
+ ],
+ [
+ "▁Commit",
+ "tee"
+ ],
+ [
+ "▁s",
+ "pending"
+ ],
+ [
+ "▁sp",
+ "ending"
+ ],
+ [
+ "▁spend",
+ "ing"
+ ],
+ [
+ "▁T",
+ "ri"
+ ],
+ [
+ "▁Tr",
+ "i"
+ ],
+ [
+ "▁",
+ "Tri"
+ ],
+ [
+ "▁J",
+ "ournal"
+ ],
+ [
+ "▁Jour",
+ "nal"
+ ],
+ [
+ "▁",
+ "Journal"
+ ],
+ [
+ "▁s",
+ "ugar"
+ ],
+ [
+ "▁su",
+ "gar"
+ ],
+ [
+ "▁sug",
+ "ar"
+ ],
+ [
+ "el",
+ "ly"
+ ],
+ [
+ "ell",
+ "y"
+ ],
+ [
+ "HT",
+ "ML"
+ ],
+ [
+ "▁ad",
+ "vent"
+ ],
+ [
+ "▁adv",
+ "ent"
+ ],
+ [
+ "win",
+ "g"
+ ],
+ [
+ "wi",
+ "ng"
+ ],
+ [
+ "w",
+ "ing"
+ ],
+ [
+ "▁Wh",
+ "ether"
+ ],
+ [
+ "▁Whe",
+ "ther"
+ ],
+ [
+ "or",
+ "ation"
+ ],
+ [
+ "▁N",
+ "E"
+ ],
+ [
+ "▁",
+ "NE"
+ ],
+ [
+ "iv",
+ "eness"
+ ],
+ [
+ "ive",
+ "ness"
+ ],
+ [
+ "iven",
+ "ess"
+ ],
+ [
+ "▁h",
+ "av"
+ ],
+ [
+ "▁ha",
+ "v"
+ ],
+ [
+ "▁",
+ "hav"
+ ],
+ [
+ "▁con",
+ "scious"
+ ],
+ [
+ "▁",
+ "conscious"
+ ],
+ [
+ "ee",
+ "n"
+ ],
+ [
+ "e",
+ "en"
+ ],
+ [
+ "Sym",
+ "bol"
+ ],
+ [
+ "S",
+ "ymbol"
+ ],
+ [
+ "▁к",
+ "у"
+ ],
+ [
+ "▁",
+ "ку"
+ ],
+ [
+ "Log",
+ "ger"
+ ],
+ [
+ "▁L",
+ "ittle"
+ ],
+ [
+ "▁Lit",
+ "tle"
+ ],
+ [
+ "wide",
+ "t"
+ ],
+ [
+ "wi",
+ "det"
+ ],
+ [
+ "wid",
+ "et"
+ ],
+ [
+ "oc",
+ "ation"
+ ],
+ [
+ "pi",
+ "n"
+ ],
+ [
+ "p",
+ "in"
+ ],
+ [
+ "▁sym",
+ "met"
+ ],
+ [
+ "▁A",
+ "D"
+ ],
+ [
+ "▁",
+ "AD"
+ ],
+ [
+ "▁pos",
+ "ts"
+ ],
+ [
+ "▁po",
+ "sts"
+ ],
+ [
+ "▁post",
+ "s"
+ ],
+ [
+ "▁",
+ "posts"
+ ],
+ [
+ "sh",
+ "al"
+ ],
+ [
+ "sha",
+ "l"
+ ],
+ [
+ "s",
+ "hal"
+ ],
+ [
+ "▁Con",
+ "f"
+ ],
+ [
+ "▁Co",
+ "nf"
+ ],
+ [
+ "▁",
+ "Conf"
+ ],
+ [
+ "▁ch",
+ "ose"
+ ],
+ [
+ "▁cho",
+ "se"
+ ],
+ [
+ "ma",
+ "l"
+ ],
+ [
+ "m",
+ "al"
+ ],
+ [
+ "ul",
+ "o"
+ ],
+ [
+ "u",
+ "lo"
+ ],
+ [
+ "▁M",
+ "ethod"
+ ],
+ [
+ "▁",
+ "Method"
+ ],
+ [
+ "▁miss",
+ "ed"
+ ],
+ [
+ "▁mis",
+ "sed"
+ ],
+ [
+ "Re",
+ "move"
+ ],
+ [
+ "Rem",
+ "ove"
+ ],
+ [
+ "Aut",
+ "o"
+ ],
+ [
+ "A",
+ "uto"
+ ],
+ [
+ "VAL",
+ "UE"
+ ],
+ [
+ "th",
+ "let"
+ ],
+ [
+ "▁For",
+ "ce"
+ ],
+ [
+ "▁",
+ "Force"
+ ],
+ [
+ "p",
+ "f"
+ ],
+ [
+ "▁",
+ "Я"
+ ],
+ [
+ "la",
+ "te"
+ ],
+ [
+ "lat",
+ "e"
+ ],
+ [
+ "l",
+ "ate"
+ ],
+ [
+ "▁p",
+ "ul"
+ ],
+ [
+ "▁pu",
+ "l"
+ ],
+ [
+ "▁",
+ "pul"
+ ],
+ [
+ "Po",
+ "p"
+ ],
+ [
+ "P",
+ "op"
+ ],
+ [
+ "▁adv",
+ "anced"
+ ],
+ [
+ "▁advance",
+ "d"
+ ],
+ [
+ "air",
+ "es"
+ ],
+ [
+ "ai",
+ "res"
+ ],
+ [
+ "aire",
+ "s"
+ ],
+ [
+ "a",
+ "ires"
+ ],
+ [
+ "res",
+ "sed"
+ ],
+ [
+ "ress",
+ "ed"
+ ],
+ [
+ "resse",
+ "d"
+ ],
+ [
+ "r",
+ "essed"
+ ],
+ [
+ "AM",
+ "E"
+ ],
+ [
+ "A",
+ "ME"
+ ],
+ [
+ "be",
+ "ll"
+ ],
+ [
+ "bel",
+ "l"
+ ],
+ [
+ "b",
+ "ell"
+ ],
+ [
+ "ac",
+ "hing"
+ ],
+ [
+ "ach",
+ "ing"
+ ],
+ [
+ "achi",
+ "ng"
+ ],
+ [
+ "a",
+ "ching"
+ ],
+ [
+ "i",
+ "ć"
+ ],
+ [
+ "ec",
+ "ho"
+ ],
+ [
+ "ech",
+ "o"
+ ],
+ [
+ "e",
+ "cho"
+ ],
+ [
+ "H",
+ "S"
+ ],
+ [
+ "▁fun",
+ "ny"
+ ],
+ [
+ "ри",
+ "и"
+ ],
+ [
+ "▁e",
+ "er"
+ ],
+ [
+ "▁ve",
+ "get"
+ ],
+ [
+ "▁four",
+ "th"
+ ],
+ [
+ "c",
+ "f"
+ ],
+ [
+ "trans",
+ "form"
+ ],
+ [
+ "▁g",
+ "rown"
+ ],
+ [
+ "▁gr",
+ "own"
+ ],
+ [
+ "▁grow",
+ "n"
+ ],
+ [
+ "▁gro",
+ "wn"
+ ],
+ [
+ "▁Mc",
+ "C"
+ ],
+ [
+ "si",
+ "te"
+ ],
+ [
+ "s",
+ "ite"
+ ],
+ [
+ "▁b",
+ "eneath"
+ ],
+ [
+ "▁be",
+ "neath"
+ ],
+ [
+ "▁s",
+ "hell"
+ ],
+ [
+ "▁sh",
+ "ell"
+ ],
+ [
+ "▁she",
+ "ll"
+ ],
+ [
+ "▁shel",
+ "l"
+ ],
+ [
+ "▁",
+ "shell"
+ ],
+ [
+ "x",
+ "d"
+ ],
+ [
+ "Pl",
+ "ay"
+ ],
+ [
+ "P",
+ "lay"
+ ],
+ [
+ "sh",
+ "ort"
+ ],
+ [
+ "Ro",
+ "le"
+ ],
+ [
+ "R",
+ "ole"
+ ],
+ [
+ "▁relig",
+ "ion"
+ ],
+ [
+ "in",
+ "ator"
+ ],
+ [
+ "ina",
+ "tor"
+ ],
+ [
+ "}",
+ ""
+ ],
+ [
+ "▁El",
+ "iz"
+ ],
+ [
+ "▁Eli",
+ "z"
+ ],
+ [
+ "M",
+ "icrosoft"
+ ],
+ [
+ "▁v",
+ "ez"
+ ],
+ [
+ "▁ve",
+ "z"
+ ],
+ [
+ "▁",
+ "vez"
+ ],
+ [
+ "▁ра",
+ "бо"
+ ],
+ [
+ "▁",
+ "рабо"
+ ],
+ [
+ "re",
+ "ich"
+ ],
+ [
+ "rei",
+ "ch"
+ ],
+ [
+ "ve",
+ "t"
+ ],
+ [
+ "v",
+ "et"
+ ],
+ [
+ "en",
+ "um"
+ ],
+ [
+ "enu",
+ "m"
+ ],
+ [
+ "e",
+ "num"
+ ],
+ [
+ "▁w",
+ "elcome"
+ ],
+ [
+ "▁wel",
+ "come"
+ ],
+ [
+ "name",
+ "nt"
+ ],
+ [
+ "na",
+ "ment"
+ ],
+ [
+ "nam",
+ "ent"
+ ],
+ [
+ "n",
+ "ament"
+ ],
+ [
+ "▁j",
+ "an"
+ ],
+ [
+ "▁ja",
+ "n"
+ ],
+ [
+ "▁",
+ "jan"
+ ],
+ [
+ "▁c",
+ "ycle"
+ ],
+ [
+ "▁cy",
+ "cle"
+ ],
+ [
+ "▁cycl",
+ "e"
+ ],
+ [
+ "▁",
+ "cycle"
+ ],
+ [
+ "▁a",
+ "cknow"
+ ],
+ [
+ "▁ac",
+ "know"
+ ],
+ [
+ "▁w",
+ "ound"
+ ],
+ [
+ "▁wo",
+ "und"
+ ],
+ [
+ "id",
+ "i"
+ ],
+ [
+ "i",
+ "di"
+ ],
+ [
+ "▁poss",
+ "ibility"
+ ],
+ [
+ "an",
+ "notation"
+ ],
+ [
+ "annot",
+ "ation"
+ ],
+ [
+ "▁techn",
+ "ical"
+ ],
+ [
+ "▁f",
+ "old"
+ ],
+ [
+ "▁fol",
+ "d"
+ ],
+ [
+ "▁fo",
+ "ld"
+ ],
+ [
+ "▁",
+ "fold"
+ ],
+ [
+ "e",
+ "h"
+ ],
+ [
+ "ist",
+ "ence"
+ ],
+ [
+ "isten",
+ "ce"
+ ],
+ [
+ "▁re",
+ "ply"
+ ],
+ [
+ "▁rep",
+ "ly"
+ ],
+ [
+ "▁repl",
+ "y"
+ ],
+ [
+ "▁",
+ "reply"
+ ],
+ [
+ "et",
+ "es"
+ ],
+ [
+ "ete",
+ "s"
+ ],
+ [
+ "e",
+ "tes"
+ ],
+ [
+ "▁dec",
+ "ades"
+ ],
+ [
+ "▁decade",
+ "s"
+ ],
+ [
+ "wa",
+ "n"
+ ],
+ [
+ "w",
+ "an"
+ ],
+ [
+ "▁к",
+ "ра"
+ ],
+ [
+ "▁",
+ "кра"
+ ],
+ [
+ "▁L",
+ "ab"
+ ],
+ [
+ "▁La",
+ "b"
+ ],
+ [
+ "▁u",
+ "nf"
+ ],
+ [
+ "▁un",
+ "f"
+ ],
+ [
+ "▁im",
+ "per"
+ ],
+ [
+ "▁imp",
+ "er"
+ ],
+ [
+ "▁",
+ "imper"
+ ],
+ [
+ "▁b",
+ "ug"
+ ],
+ [
+ "▁bu",
+ "g"
+ ],
+ [
+ "▁",
+ "bug"
+ ],
+ [
+ "▁Th",
+ "ough"
+ ],
+ [
+ "th",
+ "rows"
+ ],
+ [
+ "throw",
+ "s"
+ ],
+ [
+ "Vis",
+ "ible"
+ ],
+ [
+ "V",
+ "isible"
+ ],
+ [
+ "pr",
+ "ev"
+ ],
+ [
+ "pre",
+ "v"
+ ],
+ [
+ "p",
+ "rev"
+ ],
+ [
+ "▁T",
+ "y"
+ ],
+ [
+ "▁",
+ "Ty"
+ ],
+ [
+ "▁de",
+ "pending"
+ ],
+ [
+ "▁depend",
+ "ing"
+ ],
+ [
+ "▁dep",
+ "ending"
+ ],
+ [
+ "▁pol",
+ "icies"
+ ],
+ [
+ "▁polic",
+ "ies"
+ ],
+ [
+ "an",
+ "dy"
+ ],
+ [
+ "and",
+ "y"
+ ],
+ [
+ "▁Ital",
+ "ian"
+ ],
+ [
+ "▁Italia",
+ "n"
+ ],
+ [
+ "um",
+ "a"
+ ],
+ [
+ "u",
+ "ma"
+ ],
+ [
+ "▁sign",
+ "s"
+ ],
+ [
+ "▁sig",
+ "ns"
+ ],
+ [
+ "▁Th",
+ "rough"
+ ],
+ [
+ "б",
+ "ы"
+ ],
+ [
+ "bo",
+ "t"
+ ],
+ [
+ "b",
+ "ot"
+ ],
+ [
+ "▁pub",
+ "lish"
+ ],
+ [
+ "▁publi",
+ "sh"
+ ],
+ [
+ "▁",
+ "publish"
+ ],
+ [
+ ")*",
+ "*"
+ ],
+ [
+ ")",
+ "**"
+ ],
+ [
+ "AT",
+ "TR"
+ ],
+ [
+ "ATT",
+ "R"
+ ],
+ [
+ "ir",
+ "al"
+ ],
+ [
+ "ira",
+ "l"
+ ],
+ [
+ "i",
+ "ral"
+ ],
+ [
+ "V",
+ "T"
+ ],
+ [
+ "▁recogn",
+ "ized"
+ ],
+ [
+ "▁recognize",
+ "d"
+ ],
+ [
+ "▁L",
+ "ind"
+ ],
+ [
+ "▁Lin",
+ "d"
+ ],
+ [
+ "▁Li",
+ "nd"
+ ],
+ [
+ "ect",
+ "ion"
+ ],
+ [
+ "e",
+ "ction"
+ ],
+ [
+ "▁rel",
+ "atively"
+ ],
+ [
+ "▁relative",
+ "ly"
+ ],
+ [
+ "▁relativ",
+ "ely"
+ ],
+ [
+ "▁A",
+ "h"
+ ],
+ [
+ "▁",
+ "Ah"
+ ],
+ [
+ "▁D",
+ "ig"
+ ],
+ [
+ "▁Di",
+ "g"
+ ],
+ [
+ "▁",
+ "Dig"
+ ],
+ [
+ "ц",
+ "ь"
+ ],
+ [
+ "ic",
+ "ket"
+ ],
+ [
+ "ick",
+ "et"
+ ],
+ [
+ "▁specific",
+ "ally"
+ ],
+ [
+ "no",
+ "st"
+ ],
+ [
+ "nos",
+ "t"
+ ],
+ [
+ "n",
+ "ost"
+ ],
+ [
+ "▁g",
+ "rass"
+ ],
+ [
+ "▁gr",
+ "ass"
+ ],
+ [
+ "▁gra",
+ "ss"
+ ],
+ [
+ "▁gras",
+ "s"
+ ],
+ [
+ "▁c",
+ "auses"
+ ],
+ [
+ "▁caus",
+ "es"
+ ],
+ [
+ "▁cause",
+ "s"
+ ],
+ [
+ "▁ca",
+ "uses"
+ ],
+ [
+ "т",
+ "во"
+ ],
+ [
+ "ut",
+ "ter"
+ ],
+ [
+ "utt",
+ "er"
+ ],
+ [
+ "▁F",
+ "estival"
+ ],
+ [
+ "▁Fest",
+ "ival"
+ ],
+ [
+ "gr",
+ "eg"
+ ],
+ [
+ "gre",
+ "g"
+ ],
+ [
+ "g",
+ "reg"
+ ],
+ [
+ "▁weap",
+ "ons"
+ ],
+ [
+ "▁weapon",
+ "s"
+ ],
+ [
+ "▁s",
+ "ir"
+ ],
+ [
+ "▁si",
+ "r"
+ ],
+ [
+ "▁Virgin",
+ "ia"
+ ],
+ [
+ "lo",
+ "gin"
+ ],
+ [
+ "log",
+ "in"
+ ],
+ [
+ "▁s",
+ "chedul"
+ ],
+ [
+ "▁sched",
+ "ul"
+ ],
+ [
+ "сь",
+ "кого"
+ ],
+ [
+ "сько",
+ "го"
+ ],
+ [
+ "▁l",
+ "osing"
+ ],
+ [
+ "▁lo",
+ "sing"
+ ],
+ [
+ "▁los",
+ "ing"
+ ],
+ [
+ "▁E",
+ "urop"
+ ],
+ [
+ "▁Euro",
+ "p"
+ ],
+ [
+ "▁Eu",
+ "rop"
+ ],
+ [
+ "\">",
+ "<"
+ ],
+ [
+ "\"",
+ "><"
+ ],
+ [
+ "as",
+ "p"
+ ],
+ [
+ "a",
+ "sp"
+ ],
+ [
+ "aj",
+ "o"
+ ],
+ [
+ "a",
+ "jo"
+ ],
+ [
+ "ex",
+ "ports"
+ ],
+ [
+ "exp",
+ "orts"
+ ],
+ [
+ "export",
+ "s"
+ ],
+ [
+ "▁N",
+ "ode"
+ ],
+ [
+ "▁No",
+ "de"
+ ],
+ [
+ "▁",
+ "Node"
+ ],
+ [
+ "▁j",
+ "ako"
+ ],
+ [
+ "▁ja",
+ "ko"
+ ],
+ [
+ "▁jak",
+ "o"
+ ],
+ [
+ "▁y",
+ "a"
+ ],
+ [
+ "▁",
+ "ya"
+ ],
+ [
+ "▁success",
+ "fully"
+ ],
+ [
+ "▁successful",
+ "ly"
+ ],
+ [
+ "▁friend",
+ "ly"
+ ],
+ [
+ "▁",
+ "friendly"
+ ],
+ [
+ "buf",
+ "f"
+ ],
+ [
+ "bu",
+ "ff"
+ ],
+ [
+ "b",
+ "uff"
+ ],
+ [
+ "DE",
+ "FAULT"
+ ],
+ [
+ "▁pre",
+ "gn"
+ ],
+ [
+ "▁preg",
+ "n"
+ ],
+ [
+ "Requ",
+ "ired"
+ ],
+ [
+ "Require",
+ "d"
+ ],
+ [
+ "▁b",
+ "inary"
+ ],
+ [
+ "▁bin",
+ "ary"
+ ],
+ [
+ "▁",
+ "binary"
+ ],
+ [
+ "is",
+ "ting"
+ ],
+ [
+ "ist",
+ "ing"
+ ],
+ [
+ "isti",
+ "ng"
+ ],
+ [
+ "▁st",
+ "ared"
+ ],
+ [
+ "▁star",
+ "ed"
+ ],
+ [
+ "▁stare",
+ "d"
+ ],
+ [
+ "▁sta",
+ "red"
+ ],
+ [
+ "▁circum",
+ "stances"
+ ],
+ [
+ "▁х",
+ "о"
+ ],
+ [
+ "▁",
+ "хо"
+ ],
+ [
+ "re",
+ "i"
+ ],
+ [
+ "r",
+ "ei"
+ ],
+ [
+ "▁Г",
+ "о"
+ ],
+ [
+ "Trans",
+ "form"
+ ],
+ [
+ "cn",
+ "t"
+ ],
+ [
+ "c",
+ "nt"
+ ],
+ [
+ "▁E",
+ "xt"
+ ],
+ [
+ "▁Ex",
+ "t"
+ ],
+ [
+ "▁",
+ "Ext"
+ ],
+ [
+ "re",
+ "port"
+ ],
+ [
+ "rep",
+ "ort"
+ ],
+ [
+ "repo",
+ "rt"
+ ],
+ [
+ "VER",
+ "SION"
+ ],
+ [
+ "▁an",
+ "aly"
+ ],
+ [
+ "▁anal",
+ "y"
+ ],
+ [
+ "▁",
+ "analy"
+ ],
+ [
+ "▁M",
+ "arg"
+ ],
+ [
+ "▁Mar",
+ "g"
+ ],
+ [
+ "▁Ma",
+ "rg"
+ ],
+ [
+ "▁al",
+ "leg"
+ ],
+ [
+ "▁all",
+ "eg"
+ ],
+ [
+ "▁alle",
+ "g"
+ ],
+ [
+ "build",
+ "er"
+ ],
+ [
+ "b",
+ "uilder"
+ ],
+ [
+ "To",
+ "String"
+ ],
+ [
+ "La",
+ "yer"
+ ],
+ [
+ "L",
+ "ayer"
+ ],
+ [
+ "ís",
+ "t"
+ ],
+ [
+ "í",
+ "st"
+ ],
+ [
+ "Pro",
+ "p"
+ ],
+ [
+ "Pr",
+ "op"
+ ],
+ [
+ "P",
+ "rop"
+ ],
+ [
+ "▁E",
+ "mp"
+ ],
+ [
+ "▁Em",
+ "p"
+ ],
+ [
+ "▁",
+ "Emp"
+ ],
+ [
+ "}",
+ "]"
+ ],
+ [
+ "▁s",
+ "elling"
+ ],
+ [
+ "▁sell",
+ "ing"
+ ],
+ [
+ "▁sel",
+ "ling"
+ ],
+ [
+ "▁",
+ "selling"
+ ],
+ [
+ "▁que",
+ "ue"
+ ],
+ [
+ "▁",
+ "queue"
+ ],
+ [
+ "▁ser",
+ "iously"
+ ],
+ [
+ "▁serious",
+ "ly"
+ ],
+ [
+ "▁L",
+ "ead"
+ ],
+ [
+ "▁Le",
+ "ad"
+ ],
+ [
+ "▁",
+ "Lead"
+ ],
+ [
+ "text",
+ "it"
+ ],
+ [
+ "tex",
+ "tit"
+ ],
+ [
+ "test",
+ "ing"
+ ],
+ [
+ "tes",
+ "ting"
+ ],
+ [
+ "▁П",
+ "ре"
+ ],
+ [
+ "se",
+ "curity"
+ ],
+ [
+ "sec",
+ "urity"
+ ],
+ [
+ "ia",
+ "ł"
+ ],
+ [
+ "i",
+ "ał"
+ ],
+ [
+ "ú",
+ "n"
+ ],
+ [
+ "ch",
+ "ip"
+ ],
+ [
+ "chi",
+ "p"
+ ],
+ [
+ "c",
+ "hip"
+ ],
+ [
+ "▁c",
+ "andidate"
+ ],
+ [
+ "▁candid",
+ "ate"
+ ],
+ [
+ "▁min",
+ "ister"
+ ],
+ [
+ "▁mini",
+ "ster"
+ ],
+ [
+ "▁minist",
+ "er"
+ ],
+ [
+ "▁",
+ "minister"
+ ],
+ [
+ "er",
+ "ia"
+ ],
+ [
+ "eri",
+ "a"
+ ],
+ [
+ "e",
+ "ria"
+ ],
+ [
+ "▁H",
+ "et"
+ ],
+ [
+ "▁He",
+ "t"
+ ],
+ [
+ "ди",
+ "н"
+ ],
+ [
+ "д",
+ "ин"
+ ],
+ [
+ "▁Brit",
+ "ain"
+ ],
+ [
+ "▁b",
+ "arely"
+ ],
+ [
+ "▁bar",
+ "ely"
+ ],
+ [
+ "▁bare",
+ "ly"
+ ],
+ [
+ "▁s",
+ "ty"
+ ],
+ [
+ "▁st",
+ "y"
+ ],
+ [
+ "▁",
+ "sty"
+ ],
+ [
+ "▁Span",
+ "ish"
+ ],
+ [
+ "▁V",
+ "en"
+ ],
+ [
+ "▁Ve",
+ "n"
+ ],
+ [
+ "time",
+ "r"
+ ],
+ [
+ "ti",
+ "mer"
+ ],
+ [
+ "tim",
+ "er"
+ ],
+ [
+ "t",
+ "imer"
+ ],
+ [
+ "кі",
+ "в"
+ ],
+ [
+ "к",
+ "ів"
+ ],
+ [
+ "▁document",
+ "s"
+ ],
+ [
+ "▁doc",
+ "uments"
+ ],
+ [
+ "('",
+ "."
+ ],
+ [
+ "(",
+ "'."
+ ],
+ [
+ "▁d",
+ "ebug"
+ ],
+ [
+ "▁de",
+ "bug"
+ ],
+ [
+ "▁deb",
+ "ug"
+ ],
+ [
+ "▁",
+ "debug"
+ ],
+ [
+ "▁cont",
+ "ro"
+ ],
+ [
+ "▁contr",
+ "o"
+ ],
+ [
+ "сто",
+ "я"
+ ],
+ [
+ "▁j",
+ "oy"
+ ],
+ [
+ "▁jo",
+ "y"
+ ],
+ [
+ "▁",
+ "joy"
+ ],
+ [
+ "S",
+ "n"
+ ],
+ [
+ "In",
+ "v"
+ ],
+ [
+ "I",
+ "nv"
+ ],
+ [
+ "▁pro",
+ "tocol"
+ ],
+ [
+ "▁proto",
+ "col"
+ ],
+ [
+ "▁prot",
+ "ocol"
+ ],
+ [
+ "▁",
+ "protocol"
+ ],
+ [
+ "▁f",
+ "aces"
+ ],
+ [
+ "▁face",
+ "s"
+ ],
+ [
+ "▁fac",
+ "es"
+ ],
+ [
+ "▁fa",
+ "ces"
+ ],
+ [
+ "▁",
+ "faces"
+ ],
+ [
+ "▁Des",
+ "pite"
+ ],
+ [
+ "se",
+ "d"
+ ],
+ [
+ "s",
+ "ed"
+ ],
+ [
+ "Con",
+ "f"
+ ],
+ [
+ "Co",
+ "nf"
+ ],
+ [
+ "AR",
+ "G"
+ ],
+ [
+ "A",
+ "RG"
+ ],
+ [
+ "▁e",
+ "volution"
+ ],
+ [
+ "▁ev",
+ "olution"
+ ],
+ [
+ "▁t",
+ "od"
+ ],
+ [
+ "▁to",
+ "d"
+ ],
+ [
+ "▁P",
+ "romise"
+ ],
+ [
+ "▁Prom",
+ "ise"
+ ],
+ [
+ "▁",
+ "Promise"
+ ],
+ [
+ "▁pos",
+ "ted"
+ ],
+ [
+ "▁po",
+ "sted"
+ ],
+ [
+ "▁post",
+ "ed"
+ ],
+ [
+ "Per",
+ "m"
+ ],
+ [
+ "Pe",
+ "rm"
+ ],
+ [
+ "P",
+ "erm"
+ ],
+ [
+ "be",
+ "t"
+ ],
+ [
+ "b",
+ "et"
+ ],
+ [
+ "An",
+ "g"
+ ],
+ [
+ "A",
+ "ng"
+ ],
+ [
+ "J",
+ "ust"
+ ],
+ [
+ "▁r",
+ "um"
+ ],
+ [
+ "▁ru",
+ "m"
+ ],
+ [
+ "▁",
+ "rum"
+ ],
+ [
+ "la",
+ "yer"
+ ],
+ [
+ "lay",
+ "er"
+ ],
+ [
+ "l",
+ "ayer"
+ ],
+ [
+ "▁beh",
+ "avi"
+ ],
+ [
+ "▁behav",
+ "i"
+ ],
+ [
+ "ip",
+ "ping"
+ ],
+ [
+ "ipp",
+ "ing"
+ ],
+ [
+ "ippi",
+ "ng"
+ ],
+ [
+ "i",
+ "pping"
+ ],
+ [
+ "▁d",
+ "ynam"
+ ],
+ [
+ "▁dy",
+ "nam"
+ ],
+ [
+ "▁dyn",
+ "am"
+ ],
+ [
+ "▁sch",
+ "eme"
+ ],
+ [
+ "▁sche",
+ "me"
+ ],
+ [
+ "▁",
+ "scheme"
+ ],
+ [
+ "▁pro",
+ "to"
+ ],
+ [
+ "▁pr",
+ "oto"
+ ],
+ [
+ "▁prot",
+ "o"
+ ],
+ [
+ "▁",
+ "proto"
+ ],
+ [
+ ")",
+ "/"
+ ],
+ [
+ "Col",
+ "lections"
+ ],
+ [
+ "Collection",
+ "s"
+ ],
+ [
+ "Collect",
+ "ions"
+ ],
+ [
+ "ri",
+ "ev"
+ ],
+ [
+ "rie",
+ "v"
+ ],
+ [
+ "r",
+ "iev"
+ ],
+ [
+ "▁C",
+ "lick"
+ ],
+ [
+ "▁Cl",
+ "ick"
+ ],
+ [
+ "▁",
+ "Click"
+ ],
+ [
+ "▁u",
+ "ns"
+ ],
+ [
+ "▁un",
+ "s"
+ ],
+ [
+ "▁",
+ "uns"
+ ],
+ [
+ "wide",
+ "tilde"
+ ],
+ [
+ "widet",
+ "ilde"
+ ],
+ [
+ "▁remember",
+ "ed"
+ ],
+ [
+ "г",
+ "і"
+ ],
+ [
+ "in",
+ "ates"
+ ],
+ [
+ "ina",
+ "tes"
+ ],
+ [
+ "inate",
+ "s"
+ ],
+ [
+ "▁incor",
+ "por"
+ ],
+ [
+ "▁De",
+ "scription"
+ ],
+ [
+ "▁Des",
+ "cription"
+ ],
+ [
+ "▁",
+ "Description"
+ ],
+ [
+ "▁pre",
+ "pare"
+ ],
+ [
+ "▁prep",
+ "are"
+ ],
+ [
+ "▁prepar",
+ "e"
+ ],
+ [
+ "▁",
+ "prepare"
+ ],
+ [
+ "▁F",
+ "inal"
+ ],
+ [
+ "▁Fin",
+ "al"
+ ],
+ [
+ "▁Fi",
+ "nal"
+ ],
+ [
+ "▁",
+ "Final"
+ ],
+ [
+ "u",
+ "ation"
+ ],
+ [
+ "▁Qu",
+ "een"
+ ],
+ [
+ "▁Que",
+ "en"
+ ],
+ [
+ ">",
+ ";"
+ ],
+ [
+ "▁autom",
+ "atically"
+ ],
+ [
+ "▁automatic",
+ "ally"
+ ],
+ [
+ "▁sh",
+ "arp"
+ ],
+ [
+ "▁shar",
+ "p"
+ ],
+ [
+ "▁sha",
+ "rp"
+ ],
+ [
+ "▁me",
+ "at"
+ ],
+ [
+ "at",
+ "eur"
+ ],
+ [
+ "ate",
+ "ur"
+ ],
+ [
+ "as",
+ "tern"
+ ],
+ [
+ "ast",
+ "ern"
+ ],
+ [
+ "aster",
+ "n"
+ ],
+ [
+ "aste",
+ "rn"
+ ],
+ [
+ "▁st",
+ "uck"
+ ],
+ [
+ "ASS",
+ "ERT"
+ ],
+ [
+ "▁pl",
+ "anned"
+ ],
+ [
+ "▁plan",
+ "ned"
+ ],
+ [
+ "do",
+ "ts"
+ ],
+ [
+ "dot",
+ "s"
+ ],
+ [
+ "d",
+ "ots"
+ ],
+ [
+ "ook",
+ "ie"
+ ],
+ [
+ "oo",
+ "kie"
+ ],
+ [
+ "▁His",
+ "tor"
+ ],
+ [
+ "▁Hist",
+ "or"
+ ],
+ [
+ "▁re",
+ "views"
+ ],
+ [
+ "▁review",
+ "s"
+ ],
+ [
+ "IM",
+ "P"
+ ],
+ [
+ "I",
+ "MP"
+ ],
+ [
+ "▁answ",
+ "ered"
+ ],
+ [
+ "▁answer",
+ "ed"
+ ],
+ [
+ "To",
+ "tal"
+ ],
+ [
+ "T",
+ "otal"
+ ],
+ [
+ "▁s",
+ "au"
+ ],
+ [
+ "▁sa",
+ "u"
+ ],
+ [
+ "▁Me",
+ "xico"
+ ],
+ [
+ "▁Mex",
+ "ico"
+ ],
+ [
+ "contin",
+ "ue"
+ ],
+ [
+ "▁App",
+ "le"
+ ],
+ [
+ "▁Ap",
+ "ple"
+ ],
+ [
+ "like",
+ "ly"
+ ],
+ [
+ "lik",
+ "ely"
+ ],
+ [
+ "з",
+ "ва"
+ ],
+ [
+ "us",
+ "ers"
+ ],
+ [
+ "use",
+ "rs"
+ ],
+ [
+ "user",
+ "s"
+ ],
+ [
+ "▁ident",
+ "ified"
+ ],
+ [
+ "▁L",
+ "ev"
+ ],
+ [
+ "▁Le",
+ "v"
+ ],
+ [
+ "▁m",
+ "ol"
+ ],
+ [
+ "▁mo",
+ "l"
+ ],
+ [
+ "▁Is",
+ "lam"
+ ],
+ [
+ "▁com",
+ "mitted"
+ ],
+ [
+ "▁comm",
+ "itted"
+ ],
+ [
+ "▁commit",
+ "ted"
+ ],
+ [
+ "wr",
+ "it"
+ ],
+ [
+ "w",
+ "rit"
+ ],
+ [
+ "бе",
+ "р"
+ ],
+ [
+ "б",
+ "ер"
+ ],
+ [
+ "ri",
+ "ft"
+ ],
+ [
+ "rif",
+ "t"
+ ],
+ [
+ "r",
+ "ift"
+ ],
+ [
+ "▁inter",
+ "rupt"
+ ],
+ [
+ "▁",
+ "interrupt"
+ ],
+ [
+ "▁read",
+ "only"
+ ],
+ [
+ "sch",
+ "ema"
+ ],
+ [
+ "sche",
+ "ma"
+ ],
+ [
+ "s",
+ "chema"
+ ],
+ [
+ "S",
+ "m"
+ ],
+ [
+ "D",
+ "ouble"
+ ],
+ [
+ "az",
+ "a"
+ ],
+ [
+ "a",
+ "za"
+ ],
+ [
+ "▁H",
+ "al"
+ ],
+ [
+ "▁Ha",
+ "l"
+ ],
+ [
+ "▁",
+ "Hal"
+ ],
+ [
+ "Mo",
+ "ve"
+ ],
+ [
+ "M",
+ "ove"
+ ],
+ [
+ "▁S",
+ "eries"
+ ],
+ [
+ "▁Se",
+ "ries"
+ ],
+ [
+ "▁Ser",
+ "ies"
+ ],
+ [
+ "▁Serie",
+ "s"
+ ],
+ [
+ "▁",
+ "Series"
+ ],
+ [
+ "in",
+ "line"
+ ],
+ [
+ "▁кото",
+ "ры"
+ ],
+ [
+ "so",
+ "c"
+ ],
+ [
+ "s",
+ "oc"
+ ],
+ [
+ "▁t",
+ "ent"
+ ],
+ [
+ "▁te",
+ "nt"
+ ],
+ [
+ "▁ten",
+ "t"
+ ],
+ [
+ "▁a",
+ "mer"
+ ],
+ [
+ "▁am",
+ "er"
+ ],
+ [
+ "▁",
+ "amer"
+ ],
+ [
+ "ak",
+ "i"
+ ],
+ [
+ "a",
+ "ki"
+ ],
+ [
+ "▁l",
+ "ady"
+ ],
+ [
+ "▁la",
+ "dy"
+ ],
+ [
+ "▁lad",
+ "y"
+ ],
+ [
+ "▁t",
+ "ired"
+ ],
+ [
+ "▁ti",
+ "red"
+ ],
+ [
+ "▁tire",
+ "d"
+ ],
+ [
+ "▁tir",
+ "ed"
+ ],
+ [
+ "if",
+ "i"
+ ],
+ [
+ "i",
+ "fi"
+ ],
+ [
+ "▁m",
+ "ême"
+ ],
+ [
+ "▁",
+ "même"
+ ],
+ [
+ "ou",
+ "ver"
+ ],
+ [
+ "▁a",
+ "side"
+ ],
+ [
+ "▁as",
+ "ide"
+ ],
+ [
+ "Di",
+ "d"
+ ],
+ [
+ "D",
+ "id"
+ ],
+ [
+ "',",
+ "\r"
+ ],
+ [
+ "'",
+ ",\r"
+ ],
+ [
+ "▁br",
+ "inging"
+ ],
+ [
+ "▁bring",
+ "ing"
+ ],
+ [
+ "Draw",
+ "ing"
+ ],
+ [
+ "ar",
+ "o"
+ ],
+ [
+ "a",
+ "ro"
+ ],
+ [
+ "▁R",
+ "h"
+ ],
+ [
+ "▁N",
+ "az"
+ ],
+ [
+ "▁Na",
+ "z"
+ ],
+ [
+ "es",
+ "so"
+ ],
+ [
+ "ess",
+ "o"
+ ],
+ [
+ "▁re",
+ "action"
+ ],
+ [
+ "▁react",
+ "ion"
+ ],
+ [
+ "mit",
+ "ted"
+ ],
+ [
+ "mitt",
+ "ed"
+ ],
+ [
+ "m",
+ "itted"
+ ],
+ [
+ "▁abs",
+ "olute"
+ ],
+ [
+ "▁absolut",
+ "e"
+ ],
+ [
+ "▁",
+ "absolute"
+ ],
+ [
+ "ha",
+ "ust"
+ ],
+ [
+ "haus",
+ "t"
+ ],
+ [
+ "((",
+ ")"
+ ],
+ [
+ "(",
+ "()"
+ ],
+ [
+ "▁T",
+ "ask"
+ ],
+ [
+ "▁Ta",
+ "sk"
+ ],
+ [
+ "▁",
+ "Task"
+ ],
+ [
+ "ER",
+ "S"
+ ],
+ [
+ "E",
+ "RS"
+ ],
+ [
+ "▁^",
+ "{"
+ ],
+ [
+ "▁",
+ "^{"
+ ],
+ [
+ "V",
+ "D"
+ ],
+ [
+ "▁t",
+ "one"
+ ],
+ [
+ "▁to",
+ "ne"
+ ],
+ [
+ "▁ton",
+ "e"
+ ],
+ [
+ "dis",
+ "t"
+ ],
+ [
+ "di",
+ "st"
+ ],
+ [
+ "d",
+ "ist"
+ ],
+ [
+ "v",
+ "s"
+ ],
+ [
+ "▁whe",
+ "el"
+ ],
+ [
+ "▁",
+ "wheel"
+ ],
+ [
+ "▁administr",
+ "ation"
+ ],
+ [
+ "▁admin",
+ "istration"
+ ],
+ [
+ "▁inter",
+ "ests"
+ ],
+ [
+ "▁interest",
+ "s"
+ ],
+ [
+ "▁point",
+ "er"
+ ],
+ [
+ "▁po",
+ "inter"
+ ],
+ [
+ "▁",
+ "pointer"
+ ],
+ [
+ "▁en",
+ "counter"
+ ],
+ [
+ "▁enc",
+ "ounter"
+ ],
+ [
+ "ave",
+ "r"
+ ],
+ [
+ "av",
+ "er"
+ ],
+ [
+ "a",
+ "ver"
+ ],
+ [
+ "▁n",
+ "ord"
+ ],
+ [
+ "▁no",
+ "rd"
+ ],
+ [
+ "▁nor",
+ "d"
+ ],
+ [
+ "ke",
+ "t"
+ ],
+ [
+ "k",
+ "et"
+ ],
+ [
+ "▁b",
+ "each"
+ ],
+ [
+ "▁be",
+ "ach"
+ ],
+ [
+ "▁enjoy",
+ "ed"
+ ],
+ [
+ "cont",
+ "ains"
+ ],
+ [
+ "▁app",
+ "end"
+ ],
+ [
+ "▁ap",
+ "pend"
+ ],
+ [
+ "▁appe",
+ "nd"
+ ],
+ [
+ "▁",
+ "append"
+ ],
+ [
+ "W",
+ "ait"
+ ],
+ [
+ "▁s",
+ "quad"
+ ],
+ [
+ "▁squ",
+ "ad"
+ ],
+ [
+ "ze",
+ "l"
+ ],
+ [
+ "z",
+ "el"
+ ],
+ [
+ "▁med",
+ "ium"
+ ],
+ [
+ "▁medi",
+ "um"
+ ],
+ [
+ "▁",
+ "medium"
+ ],
+ [
+ "▁s",
+ "ending"
+ ],
+ [
+ "▁send",
+ "ing"
+ ],
+ [
+ "▁sen",
+ "ding"
+ ],
+ [
+ "▁L",
+ "ady"
+ ],
+ [
+ "▁La",
+ "dy"
+ ],
+ [
+ "▁Lad",
+ "y"
+ ],
+ [
+ "ç",
+ "ões"
+ ],
+ [
+ "▁dest",
+ "ination"
+ ],
+ [
+ "▁destin",
+ "ation"
+ ],
+ [
+ "▁",
+ "destination"
+ ],
+ [
+ "ny",
+ "ch"
+ ],
+ [
+ "n",
+ "ych"
+ ],
+ [
+ "▁conf",
+ "lict"
+ ],
+ [
+ "▁conflic",
+ "t"
+ ],
+ [
+ "▁L",
+ "y"
+ ],
+ [
+ "▁v",
+ "ul"
+ ],
+ [
+ "▁vu",
+ "l"
+ ],
+ [
+ "▁bas",
+ "ically"
+ ],
+ [
+ "▁basic",
+ "ally"
+ ],
+ [
+ "re",
+ "ated"
+ ],
+ [
+ "reat",
+ "ed"
+ ],
+ [
+ "reate",
+ "d"
+ ],
+ [
+ "rea",
+ "ted"
+ ],
+ [
+ "bl",
+ "ack"
+ ],
+ [
+ "ug",
+ "ins"
+ ],
+ [
+ "ugin",
+ "s"
+ ],
+ [
+ "▁cal",
+ "m"
+ ],
+ [
+ "▁ca",
+ "lm"
+ ],
+ [
+ "ér",
+ "ie"
+ ],
+ [
+ "éri",
+ "e"
+ ],
+ [
+ "é",
+ "rie"
+ ],
+ [
+ "ha",
+ "r"
+ ],
+ [
+ "h",
+ "ar"
+ ],
+ [
+ "ла",
+ "н"
+ ],
+ [
+ "л",
+ "ан"
+ ],
+ [
+ "▁С",
+ "е"
+ ],
+ [
+ "w",
+ "atch"
+ ],
+ [
+ "▁P",
+ "ut"
+ ],
+ [
+ "▁Pu",
+ "t"
+ ],
+ [
+ "▁",
+ "Put"
+ ],
+ [
+ "▁d",
+ "ump"
+ ],
+ [
+ "▁du",
+ "mp"
+ ],
+ [
+ "▁",
+ "dump"
+ ],
+ [
+ "ac",
+ "her"
+ ],
+ [
+ "ach",
+ "er"
+ ],
+ [
+ "ache",
+ "r"
+ ],
+ [
+ "a",
+ "cher"
+ ],
+ [
+ "sc",
+ "roll"
+ ],
+ [
+ "scr",
+ "oll"
+ ],
+ [
+ "▁cl",
+ "aimed"
+ ],
+ [
+ "▁claim",
+ "ed"
+ ],
+ [
+ "▁",
+ "claimed"
+ ],
+ [
+ "▁Cont",
+ "rol"
+ ],
+ [
+ "▁",
+ "Control"
+ ],
+ [
+ "▁bl",
+ "ind"
+ ],
+ [
+ "en",
+ "ti"
+ ],
+ [
+ "ent",
+ "i"
+ ],
+ [
+ "▁Ke",
+ "ep"
+ ],
+ [
+ "▁",
+ "Keep"
+ ],
+ [
+ "▁Develop",
+ "ment"
+ ],
+ [
+ "im",
+ "ages"
+ ],
+ [
+ "image",
+ "s"
+ ],
+ [
+ "ima",
+ "ges"
+ ],
+ [
+ "imag",
+ "es"
+ ],
+ [
+ "▁t",
+ "ough"
+ ],
+ [
+ "▁to",
+ "ugh"
+ ],
+ [
+ "▁tou",
+ "gh"
+ ],
+ [
+ "ge",
+ "bra"
+ ],
+ [
+ "geb",
+ "ra"
+ ],
+ [
+ "▁se",
+ "pt"
+ ],
+ [
+ "▁sep",
+ "t"
+ ],
+ [
+ "he",
+ "w"
+ ],
+ [
+ "h",
+ "ew"
+ ],
+ [
+ "▁s",
+ "kill"
+ ],
+ [
+ "▁sk",
+ "ill"
+ ],
+ [
+ "▁ski",
+ "ll"
+ ],
+ [
+ "▁",
+ "skill"
+ ],
+ [
+ "▁T",
+ "ay"
+ ],
+ [
+ "▁Ta",
+ "y"
+ ],
+ [
+ "▁k",
+ "tó"
+ ],
+ [
+ "ow",
+ "ner"
+ ],
+ [
+ "own",
+ "er"
+ ],
+ [
+ "par",
+ "e"
+ ],
+ [
+ "pa",
+ "re"
+ ],
+ [
+ "p",
+ "are"
+ ],
+ [
+ "▁f",
+ "ee"
+ ],
+ [
+ "▁fe",
+ "e"
+ ],
+ [
+ "▁",
+ "fee"
+ ],
+ [
+ "▁contin",
+ "ues"
+ ],
+ [
+ "▁continue",
+ "s"
+ ],
+ [
+ "▁continu",
+ "es"
+ ],
+ [
+ "▁k",
+ "an"
+ ],
+ [
+ "▁ka",
+ "n"
+ ],
+ [
+ "▁",
+ "kan"
+ ],
+ [
+ "be",
+ "s"
+ ],
+ [
+ "b",
+ "es"
+ ],
+ [
+ "▁c",
+ "ha"
+ ],
+ [
+ "▁ch",
+ "a"
+ ],
+ [
+ "▁",
+ "cha"
+ ],
+ [
+ "ov",
+ "o"
+ ],
+ [
+ "o",
+ "vo"
+ ],
+ [
+ "▁N",
+ "ight"
+ ],
+ [
+ "▁Ni",
+ "ght"
+ ],
+ [
+ "ict",
+ "ure"
+ ],
+ [
+ "sh",
+ "ire"
+ ],
+ [
+ "s",
+ "hire"
+ ],
+ [
+ "▁es",
+ "say"
+ ],
+ [
+ "▁ess",
+ "ay"
+ ],
+ [
+ "▁sup",
+ "pose"
+ ],
+ [
+ "▁supp",
+ "ose"
+ ],
+ [
+ "et",
+ "ic"
+ ],
+ [
+ "eti",
+ "c"
+ ],
+ [
+ "Ar",
+ "t"
+ ],
+ [
+ "A",
+ "rt"
+ ],
+ [
+ "ac",
+ "on"
+ ],
+ [
+ "aco",
+ "n"
+ ],
+ [
+ "a",
+ "con"
+ ],
+ [
+ "ll",
+ "a"
+ ],
+ [
+ "l",
+ "la"
+ ],
+ [
+ "word",
+ "s"
+ ],
+ [
+ "wor",
+ "ds"
+ ],
+ [
+ "w",
+ "ords"
+ ],
+ [
+ "▁compar",
+ "ison"
+ ],
+ [
+ "▁B",
+ "E"
+ ],
+ [
+ "▁",
+ "BE"
+ ],
+ [
+ "▁challeng",
+ "es"
+ ],
+ [
+ "▁challenge",
+ "s"
+ ],
+ [
+ "▁o",
+ "l"
+ ],
+ [
+ "▁",
+ "ol"
+ ],
+ [
+ "cite",
+ "p"
+ ],
+ [
+ "cit",
+ "ep"
+ ],
+ [
+ "▁F",
+ "oot"
+ ],
+ [
+ "▁Fo",
+ "ot"
+ ],
+ [
+ "▁",
+ "Foot"
+ ],
+ [
+ "▁S",
+ "uch"
+ ],
+ [
+ "▁Su",
+ "ch"
+ ],
+ [
+ "▁",
+ "Such"
+ ],
+ [
+ "▁p",
+ "apers"
+ ],
+ [
+ "▁paper",
+ "s"
+ ],
+ [
+ "▁pa",
+ "pers"
+ ],
+ [
+ "▁pap",
+ "ers"
+ ],
+ [
+ "act",
+ "iv"
+ ],
+ [
+ "qu",
+ "er"
+ ],
+ [
+ "que",
+ "r"
+ ],
+ [
+ "q",
+ "uer"
+ ],
+ [
+ "т",
+ "я"
+ ],
+ [
+ "▁Т",
+ "о"
+ ],
+ [
+ "сь",
+ "кий"
+ ],
+ [
+ "th",
+ "ur"
+ ],
+ [
+ "do",
+ "ne"
+ ],
+ [
+ "don",
+ "e"
+ ],
+ [
+ "d",
+ "one"
+ ],
+ [
+ "▁sh",
+ "ock"
+ ],
+ [
+ "▁ded",
+ "icated"
+ ],
+ [
+ "▁dedic",
+ "ated"
+ ],
+ [
+ "▁cor",
+ "respond"
+ ],
+ [
+ "▁correspon",
+ "d"
+ ],
+ [
+ "Se",
+ "cond"
+ ],
+ [
+ "Sec",
+ "ond"
+ ],
+ [
+ "▁b",
+ "ull"
+ ],
+ [
+ "▁bu",
+ "ll"
+ ],
+ [
+ "▁bul",
+ "l"
+ ],
+ [
+ "li",
+ "fe"
+ ],
+ [
+ "lif",
+ "e"
+ ],
+ [
+ "l",
+ "ife"
+ ],
+ [
+ "ind",
+ "ent"
+ ],
+ [
+ "inde",
+ "nt"
+ ],
+ [
+ "inden",
+ "t"
+ ],
+ [
+ "▁fig",
+ "ures"
+ ],
+ [
+ "▁figure",
+ "s"
+ ],
+ [
+ "▁And",
+ "rew"
+ ],
+ [
+ "▁Andre",
+ "w"
+ ],
+ [
+ "▁Andr",
+ "ew"
+ ],
+ [
+ "is",
+ "p"
+ ],
+ [
+ "i",
+ "sp"
+ ],
+ [
+ "▁fav",
+ "our"
+ ],
+ [
+ "зд",
+ "а"
+ ],
+ [
+ "з",
+ "да"
+ ],
+ [
+ "▁E",
+ "lect"
+ ],
+ [
+ "▁El",
+ "ect"
+ ],
+ [
+ "▁Ele",
+ "ct"
+ ],
+ [
+ "F",
+ "ull"
+ ],
+ [
+ "▁near",
+ "by"
+ ],
+ [
+ "▁Reg",
+ "ister"
+ ],
+ [
+ "▁",
+ "Register"
+ ],
+ [
+ "Sc",
+ "ale"
+ ],
+ [
+ "Scal",
+ "e"
+ ],
+ [
+ "ic",
+ "ations"
+ ],
+ [
+ "ication",
+ "s"
+ ],
+ [
+ "и",
+ "н"
+ ],
+ [
+ "▁A",
+ "M"
+ ],
+ [
+ "▁",
+ "AM"
+ ],
+ [
+ "pa",
+ "ir"
+ ],
+ [
+ "p",
+ "air"
+ ],
+ [
+ "▁pers",
+ "pective"
+ ],
+ [
+ "▁n",
+ "os"
+ ],
+ [
+ "▁no",
+ "s"
+ ],
+ [
+ "▁",
+ "nos"
+ ],
+ [
+ "ap",
+ "a"
+ ],
+ [
+ "a",
+ "pa"
+ ],
+ [
+ "ost",
+ "ał"
+ ],
+ [
+ "osta",
+ "ł"
+ ],
+ [
+ "▁P",
+ "ers"
+ ],
+ [
+ "▁Per",
+ "s"
+ ],
+ [
+ "▁Pe",
+ "rs"
+ ],
+ [
+ "▁",
+ "Pers"
+ ],
+ [
+ "ic",
+ "er"
+ ],
+ [
+ "ice",
+ "r"
+ ],
+ [
+ "i",
+ "cer"
+ ],
+ [
+ "▁pl",
+ "astic"
+ ],
+ [
+ "до",
+ "в"
+ ],
+ [
+ "д",
+ "ов"
+ ],
+ [
+ "ci",
+ "ples"
+ ],
+ [
+ "cipl",
+ "es"
+ ],
+ [
+ "cip",
+ "les"
+ ],
+ [
+ "z",
+ "ą"
+ ],
+ [
+ "cl",
+ "os"
+ ],
+ [
+ "c",
+ "los"
+ ],
+ [
+ "▁у",
+ "ча"
+ ],
+ [
+ "▁",
+ "Á"
+ ],
+ [
+ "pl",
+ "ugin"
+ ],
+ [
+ "plug",
+ "in"
+ ],
+ [
+ "▁an",
+ "gle"
+ ],
+ [
+ "▁ang",
+ "le"
+ ],
+ [
+ "▁angl",
+ "e"
+ ],
+ [
+ "▁",
+ "angle"
+ ],
+ [
+ "▁com",
+ "mission"
+ ],
+ [
+ "▁comm",
+ "ission"
+ ],
+ [
+ "▁fun",
+ "ds"
+ ],
+ [
+ "▁fund",
+ "s"
+ ],
+ [
+ "▁in",
+ "du"
+ ],
+ [
+ "▁ind",
+ "u"
+ ],
+ [
+ "▁d",
+ "rawn"
+ ],
+ [
+ "▁dr",
+ "awn"
+ ],
+ [
+ "▁draw",
+ "n"
+ ],
+ [
+ "á",
+ "m"
+ ],
+ [
+ "▁develop",
+ "ing"
+ ],
+ [
+ "▁seg",
+ "ment"
+ ],
+ [
+ "▁",
+ "segment"
+ ],
+ [
+ "is",
+ "me"
+ ],
+ [
+ "ism",
+ "e"
+ ],
+ [
+ "sc",
+ "r"
+ ],
+ [
+ "s",
+ "cr"
+ ],
+ [
+ "▁l",
+ "ies"
+ ],
+ [
+ "▁li",
+ "es"
+ ],
+ [
+ "▁lie",
+ "s"
+ ],
+ [
+ "▁I",
+ "L"
+ ],
+ [
+ "▁",
+ "IL"
+ ],
+ [
+ "▁a",
+ "pi"
+ ],
+ [
+ "▁ap",
+ "i"
+ ],
+ [
+ "▁",
+ "api"
+ ],
+ [
+ "Ext",
+ "ension"
+ ],
+ [
+ "▁s",
+ "cal"
+ ],
+ [
+ "▁sc",
+ "al"
+ ],
+ [
+ "▁",
+ "scal"
+ ],
+ [
+ "inst",
+ "all"
+ ],
+ [
+ "▁We",
+ "ek"
+ ],
+ [
+ "▁",
+ "Week"
+ ],
+ [
+ "▁gen",
+ "tle"
+ ],
+ [
+ "▁gent",
+ "le"
+ ],
+ [
+ "▁Canad",
+ "ian"
+ ],
+ [
+ "▁d",
+ "ialog"
+ ],
+ [
+ "▁dial",
+ "og"
+ ],
+ [
+ "▁dia",
+ "log"
+ ],
+ [
+ "▁",
+ "dialog"
+ ],
+ [
+ "▁art",
+ "icles"
+ ],
+ [
+ "▁article",
+ "s"
+ ],
+ [
+ "▁artic",
+ "les"
+ ],
+ [
+ "The",
+ "me"
+ ],
+ [
+ "Th",
+ "eme"
+ ],
+ [
+ "S",
+ "M"
+ ],
+ [
+ "▁B",
+ "ul"
+ ],
+ [
+ "▁Bu",
+ "l"
+ ],
+ [
+ "▁",
+ "Bul"
+ ],
+ [
+ "▁l",
+ "eur"
+ ],
+ [
+ "▁le",
+ "ur"
+ ],
+ [
+ "▁s",
+ "tom"
+ ],
+ [
+ "▁st",
+ "om"
+ ],
+ [
+ "▁sto",
+ "m"
+ ],
+ [
+ "Pl",
+ "ugin"
+ ],
+ [
+ "▁по",
+ "сле"
+ ],
+ [
+ "▁пос",
+ "ле"
+ ],
+ [
+ "▁st",
+ "ead"
+ ],
+ [
+ "▁ste",
+ "ad"
+ ],
+ [
+ "▁",
+ "stead"
+ ],
+ [
+ "▁",
+ "ś"
+ ],
+ [
+ "ip",
+ "her"
+ ],
+ [
+ "iph",
+ "er"
+ ],
+ [
+ "i",
+ "pher"
+ ],
+ [
+ "▁pr",
+ "ze"
+ ],
+ [
+ "▁prz",
+ "e"
+ ],
+ [
+ "▁d",
+ "raft"
+ ],
+ [
+ "▁dr",
+ "aft"
+ ],
+ [
+ "▁",
+ "draft"
+ ],
+ [
+ "bot",
+ "tom"
+ ],
+ [
+ "b",
+ "ottom"
+ ],
+ [
+ "▁{",
+ "};"
+ ],
+ [
+ "▁{}",
+ ";"
+ ],
+ [
+ "▁stay",
+ "ed"
+ ],
+ [
+ "fe",
+ "ature"
+ ],
+ [
+ "feat",
+ "ure"
+ ],
+ [
+ "▁v",
+ "ot"
+ ],
+ [
+ "▁vo",
+ "t"
+ ],
+ [
+ "▁fab",
+ "ric"
+ ],
+ [
+ "ç",
+ "a"
+ ],
+ [
+ "('",
+ "#"
+ ],
+ [
+ "re",
+ "a"
+ ],
+ [
+ "r",
+ "ea"
+ ],
+ [
+ "▁re",
+ "put"
+ ],
+ [
+ "▁rep",
+ "ut"
+ ],
+ [
+ "▁C",
+ "ir"
+ ],
+ [
+ "▁Ci",
+ "r"
+ ],
+ [
+ "▁",
+ "Cir"
+ ],
+ [
+ "▁A",
+ "L"
+ ],
+ [
+ "▁",
+ "AL"
+ ],
+ [
+ "▁assert",
+ "Equals"
+ ],
+ [
+ "▁",
+ "assertEquals"
+ ],
+ [
+ "result",
+ "s"
+ ],
+ [
+ "▁C",
+ "ross"
+ ],
+ [
+ "▁Cr",
+ "oss"
+ ],
+ [
+ "▁Cro",
+ "ss"
+ ],
+ [
+ "▁",
+ "Cross"
+ ],
+ [
+ "urs",
+ "day"
+ ],
+ [
+ "▁a",
+ "udio"
+ ],
+ [
+ "▁aud",
+ "io"
+ ],
+ [
+ "▁",
+ "audio"
+ ],
+ [
+ "▁g",
+ "ap"
+ ],
+ [
+ "▁ga",
+ "p"
+ ],
+ [
+ "▁stre",
+ "ets"
+ ],
+ [
+ "▁street",
+ "s"
+ ],
+ [
+ "▁scient",
+ "ific"
+ ],
+ [
+ "pl",
+ "atform"
+ ],
+ [
+ "▁a",
+ "uss"
+ ],
+ [
+ "▁au",
+ "ss"
+ ],
+ [
+ "▁aus",
+ "s"
+ ],
+ [
+ "▁C",
+ "ro"
+ ],
+ [
+ "▁Cr",
+ "o"
+ ],
+ [
+ "▁part",
+ "ial"
+ ],
+ [
+ "▁parti",
+ "al"
+ ],
+ [
+ "▁",
+ "partial"
+ ],
+ [
+ "un",
+ "c"
+ ],
+ [
+ "u",
+ "nc"
+ ],
+ [
+ "▁cho",
+ "ices"
+ ],
+ [
+ "▁choice",
+ "s"
+ ],
+ [
+ "▁и",
+ "ли"
+ ],
+ [
+ "pr",
+ "ed"
+ ],
+ [
+ "pre",
+ "d"
+ ],
+ [
+ "p",
+ "red"
+ ],
+ [
+ "▁he",
+ "ads"
+ ],
+ [
+ "▁head",
+ "s"
+ ],
+ [
+ "▁",
+ "heads"
+ ],
+ [
+ "ter",
+ "day"
+ ],
+ [
+ "▁N",
+ "ick"
+ ],
+ [
+ "▁Nic",
+ "k"
+ ],
+ [
+ "▁Ni",
+ "ck"
+ ],
+ [
+ "▁we",
+ "ird"
+ ],
+ [
+ "as",
+ "ant"
+ ],
+ [
+ "asa",
+ "nt"
+ ],
+ [
+ "▁represent",
+ "ed"
+ ],
+ [
+ "▁п",
+ "и"
+ ],
+ [
+ "▁",
+ "пи"
+ ],
+ [
+ "D",
+ "P"
+ ],
+ [
+ "or",
+ "ders"
+ ],
+ [
+ "ord",
+ "ers"
+ ],
+ [
+ "order",
+ "s"
+ ],
+ [
+ "cl",
+ "ock"
+ ],
+ [
+ "c",
+ "lock"
+ ],
+ [
+ "▁H",
+ "o"
+ ],
+ [
+ "ar",
+ "ters"
+ ],
+ [
+ "art",
+ "ers"
+ ],
+ [
+ "arter",
+ "s"
+ ],
+ [
+ "arte",
+ "rs"
+ ],
+ [
+ "C",
+ "md"
+ ],
+ [
+ "og",
+ "a"
+ ],
+ [
+ "o",
+ "ga"
+ ],
+ [
+ "Key",
+ "s"
+ ],
+ [
+ "Ke",
+ "ys"
+ ],
+ [
+ "Re",
+ "port"
+ ],
+ [
+ "Rep",
+ "ort"
+ ],
+ [
+ "Repo",
+ "rt"
+ ],
+ [
+ "▁V",
+ "ill"
+ ],
+ [
+ "▁Vi",
+ "ll"
+ ],
+ [
+ "▁Vil",
+ "l"
+ ],
+ [
+ "▁M",
+ "u"
+ ],
+ [
+ "▁",
+ "Mu"
+ ],
+ [
+ "▁own",
+ "ed"
+ ],
+ [
+ "▁",
+ "owned"
+ ],
+ [
+ "SU",
+ "CCESS"
+ ],
+ [
+ "▁type",
+ "of"
+ ],
+ [
+ "▁",
+ "typeof"
+ ],
+ [
+ "hd",
+ "r"
+ ],
+ [
+ "h",
+ "dr"
+ ],
+ [
+ "ua",
+ "ble"
+ ],
+ [
+ "u",
+ "able"
+ ],
+ [
+ "▁neighbor",
+ "hood"
+ ],
+ [
+ "▁A",
+ "P"
+ ],
+ [
+ "▁",
+ "AP"
+ ],
+ [
+ "▁result",
+ "ing"
+ ],
+ [
+ "▁sh",
+ "adow"
+ ],
+ [
+ "▁",
+ "shadow"
+ ],
+ [
+ "STR",
+ "ING"
+ ],
+ [
+ "▁video",
+ "s"
+ ],
+ [
+ "▁vide",
+ "os"
+ ],
+ [
+ "ле",
+ "ння"
+ ],
+ [
+ "лен",
+ "ня"
+ ],
+ [
+ "ex",
+ "pect"
+ ],
+ [
+ "exp",
+ "ect"
+ ],
+ [
+ "▁Val",
+ "ley"
+ ],
+ [
+ "▁Vall",
+ "ey"
+ ],
+ [
+ "▁g",
+ "oto"
+ ],
+ [
+ "▁go",
+ "to"
+ ],
+ [
+ "▁got",
+ "o"
+ ],
+ [
+ "▁",
+ "goto"
+ ],
+ [
+ "▁S",
+ "her"
+ ],
+ [
+ "▁She",
+ "r"
+ ],
+ [
+ "▁Sh",
+ "er"
+ ],
+ [
+ "fr",
+ "astr"
+ ],
+ [
+ "▁oper",
+ "ating"
+ ],
+ [
+ "▁opera",
+ "ting"
+ ],
+ [
+ "▁э",
+ "то"
+ ],
+ [
+ "▁License",
+ "d"
+ ],
+ [
+ "▁Lic",
+ "ensed"
+ ],
+ [
+ "Var",
+ "iable"
+ ],
+ [
+ "Vari",
+ "able"
+ ],
+ [
+ "▁P",
+ "R"
+ ],
+ [
+ "▁",
+ "PR"
+ ],
+ [
+ "▁H",
+ "ans"
+ ],
+ [
+ "▁Ha",
+ "ns"
+ ],
+ [
+ "▁Han",
+ "s"
+ ],
+ [
+ "cl",
+ "one"
+ ],
+ [
+ "▁G",
+ "esch"
+ ],
+ [
+ "▁Ge",
+ "sch"
+ ],
+ [
+ "▁Ges",
+ "ch"
+ ],
+ [
+ "▁B",
+ "and"
+ ],
+ [
+ "▁Ba",
+ "nd"
+ ],
+ [
+ "▁Ban",
+ "d"
+ ],
+ [
+ "▁",
+ "Band"
+ ],
+ [
+ "...",
+ "....."
+ ],
+ [
+ "....",
+ "...."
+ ],
+ [
+ ".....",
+ "..."
+ ],
+ [
+ "ui",
+ "ng"
+ ],
+ [
+ "u",
+ "ing"
+ ],
+ [
+ "▁hundred",
+ "s"
+ ],
+ [
+ "▁о",
+ "к"
+ ],
+ [
+ "▁emot",
+ "ional"
+ ],
+ [
+ "▁emotion",
+ "al"
+ ],
+ [
+ "▁Ind",
+ "ust"
+ ],
+ [
+ ")",
+ "+"
+ ],
+ [
+ "▁Egy",
+ "pt"
+ ],
+ [
+ "▁fr",
+ "anç"
+ ],
+ [
+ "▁",
+ "š"
+ ],
+ [
+ "▁f",
+ "asc"
+ ],
+ [
+ "▁fa",
+ "sc"
+ ],
+ [
+ "on",
+ "to"
+ ],
+ [
+ "ont",
+ "o"
+ ],
+ [
+ "▁A",
+ "dam"
+ ],
+ [
+ "▁Ad",
+ "am"
+ ],
+ [
+ "▁l",
+ "aid"
+ ],
+ [
+ "▁la",
+ "id"
+ ],
+ [
+ "▁r",
+ "ig"
+ ],
+ [
+ "▁ri",
+ "g"
+ ],
+ [
+ "▁",
+ "rig"
+ ],
+ [
+ "▁det",
+ "ailed"
+ ],
+ [
+ "▁detail",
+ "ed"
+ ],
+ [
+ "▁im",
+ "plements"
+ ],
+ [
+ "▁implement",
+ "s"
+ ],
+ [
+ "▁impl",
+ "ements"
+ ],
+ [
+ "▁univers",
+ "ity"
+ ],
+ [
+ "▁H",
+ "y"
+ ],
+ [
+ "▁",
+ "Hy"
+ ],
+ [
+ "▁g",
+ "rid"
+ ],
+ [
+ "▁gr",
+ "id"
+ ],
+ [
+ "▁gri",
+ "d"
+ ],
+ [
+ "▁",
+ "grid"
+ ],
+ [
+ "▁reg",
+ "ions"
+ ],
+ [
+ "▁region",
+ "s"
+ ],
+ [
+ "St",
+ "op"
+ ],
+ [
+ "S",
+ "top"
+ ],
+ [
+ "▁s",
+ "lot"
+ ],
+ [
+ "▁sl",
+ "ot"
+ ],
+ [
+ "▁",
+ "slot"
+ ],
+ [
+ "▁ang",
+ "ry"
+ ],
+ [
+ "▁-",
+ "="
+ ],
+ [
+ "▁wait",
+ "ed"
+ ],
+ [
+ "▁wa",
+ "ited"
+ ],
+ [
+ "Ver",
+ "t"
+ ],
+ [
+ "V",
+ "ert"
+ ],
+ [
+ "\":",
+ "\""
+ ],
+ [
+ "\"",
+ ":\""
+ ],
+ [
+ "▁e",
+ "lem"
+ ],
+ [
+ "▁el",
+ "em"
+ ],
+ [
+ "▁ele",
+ "m"
+ ],
+ [
+ "▁",
+ "elem"
+ ],
+ [
+ "▁r",
+ "ég"
+ ],
+ [
+ "▁ré",
+ "g"
+ ],
+ [
+ "ow",
+ "ed"
+ ],
+ [
+ "owe",
+ "d"
+ ],
+ [
+ "o",
+ "wed"
+ ],
+ [
+ "Mem",
+ "ber"
+ ],
+ [
+ "Me",
+ "mber"
+ ],
+ [
+ "M",
+ "ember"
+ ],
+ [
+ "▁r",
+ "atio"
+ ],
+ [
+ "▁rat",
+ "io"
+ ],
+ [
+ "▁",
+ "ratio"
+ ],
+ [
+ "is",
+ "en"
+ ],
+ [
+ "ise",
+ "n"
+ ],
+ [
+ "i",
+ "sen"
+ ],
+ [
+ "▁L",
+ "em"
+ ],
+ [
+ "▁Le",
+ "m"
+ ],
+ [
+ "ge",
+ "ry"
+ ],
+ [
+ "ger",
+ "y"
+ ],
+ [
+ "g",
+ "ery"
+ ],
+ [
+ "▁c",
+ "ream"
+ ],
+ [
+ "▁cre",
+ "am"
+ ],
+ [
+ "▁ét",
+ "ait"
+ ],
+ [
+ "▁",
+ "était"
+ ],
+ [
+ "▁g",
+ "eb"
+ ],
+ [
+ "▁ge",
+ "b"
+ ],
+ [
+ "▁",
+ "geb"
+ ],
+ [
+ "un",
+ "ique"
+ ],
+ [
+ "uni",
+ "que"
+ ],
+ [
+ "▁D",
+ "eb"
+ ],
+ [
+ "▁De",
+ "b"
+ ],
+ [
+ "▁f",
+ "actory"
+ ],
+ [
+ "▁fact",
+ "ory"
+ ],
+ [
+ "▁factor",
+ "y"
+ ],
+ [
+ "▁",
+ "factory"
+ ],
+ [
+ "ż",
+ "e"
+ ],
+ [
+ "d",
+ "ialog"
+ ],
+ [
+ "▁Con",
+ "fig"
+ ],
+ [
+ "▁Conf",
+ "ig"
+ ],
+ [
+ "▁",
+ "Config"
+ ],
+ [
+ "Sy",
+ "nc"
+ ],
+ [
+ "S",
+ "ync"
+ ],
+ [
+ "an",
+ "gers"
+ ],
+ [
+ "ang",
+ "ers"
+ ],
+ [
+ "ange",
+ "rs"
+ ],
+ [
+ "anger",
+ "s"
+ ],
+ [
+ "▁gover",
+ "ning"
+ ],
+ [
+ "▁govern",
+ "ing"
+ ],
+ [
+ "▁H",
+ "un"
+ ],
+ [
+ "▁Hu",
+ "n"
+ ],
+ [
+ "Sp",
+ "ace"
+ ],
+ [
+ "S",
+ "pace"
+ ],
+ [
+ "▁j",
+ "est"
+ ],
+ [
+ "▁je",
+ "st"
+ ],
+ [
+ "ic",
+ "ious"
+ ],
+ [
+ "ici",
+ "ous"
+ ],
+ [
+ "icio",
+ "us"
+ ],
+ [
+ "▁em",
+ "phas"
+ ],
+ [
+ "▁emp",
+ "has"
+ ],
+ [
+ "um",
+ "ps"
+ ],
+ [
+ "ump",
+ "s"
+ ],
+ [
+ "▁E",
+ "sp"
+ ],
+ [
+ "▁Es",
+ "p"
+ ],
+ [
+ "▁",
+ "Esp"
+ ],
+ [
+ "▁s",
+ "ul"
+ ],
+ [
+ "▁su",
+ "l"
+ ],
+ [
+ "▁histor",
+ "ical"
+ ],
+ [
+ "▁historic",
+ "al"
+ ],
+ [
+ "ij",
+ "a"
+ ],
+ [
+ "i",
+ "ja"
+ ],
+ [
+ "▁l",
+ "ying"
+ ],
+ [
+ "▁ly",
+ "ing"
+ ],
+ [
+ "▁",
+ "lying"
+ ],
+ [
+ "▁St",
+ "eve"
+ ],
+ [
+ "▁Ste",
+ "ve"
+ ],
+ [
+ "▁me",
+ "asures"
+ ],
+ [
+ "▁measure",
+ "s"
+ ],
+ [
+ "▁meas",
+ "ures"
+ ],
+ [
+ "os",
+ "to"
+ ],
+ [
+ "ost",
+ "o"
+ ],
+ [
+ "o",
+ "sto"
+ ],
+ [
+ "?",
+ "”"
+ ],
+ [
+ "▁p",
+ "ocket"
+ ],
+ [
+ "▁poc",
+ "ket"
+ ],
+ [
+ "▁S",
+ "at"
+ ],
+ [
+ "▁Sa",
+ "t"
+ ],
+ [
+ "▁p",
+ "itch"
+ ],
+ [
+ "▁pit",
+ "ch"
+ ],
+ [
+ "▁n",
+ "atur"
+ ],
+ [
+ "▁nat",
+ "ur"
+ ],
+ [
+ "▁hum",
+ "ans"
+ ],
+ [
+ "▁human",
+ "s"
+ ],
+ [
+ "▁Sim",
+ "on"
+ ],
+ [
+ "▁Si",
+ "mon"
+ ],
+ [
+ "ad",
+ "ores"
+ ],
+ [
+ "ado",
+ "res"
+ ],
+ [
+ "ador",
+ "es"
+ ],
+ [
+ "(\"",
+ "\\"
+ ],
+ [
+ "(",
+ "\"\\"
+ ],
+ [
+ "in",
+ "king"
+ ],
+ [
+ "ink",
+ "ing"
+ ],
+ [
+ "▁ex",
+ "pos"
+ ],
+ [
+ "▁exp",
+ "os"
+ ],
+ [
+ "mat",
+ "erial"
+ ],
+ [
+ "mate",
+ "rial"
+ ],
+ [
+ "m",
+ "aterial"
+ ],
+ [
+ "▁app",
+ "arently"
+ ],
+ [
+ "▁apparent",
+ "ly"
+ ],
+ [
+ "▁appar",
+ "ently"
+ ],
+ [
+ "▁C",
+ "amb"
+ ],
+ [
+ "▁Cam",
+ "b"
+ ],
+ [
+ "▁Ca",
+ "mb"
+ ],
+ [
+ "▁B",
+ "ox"
+ ],
+ [
+ "▁Bo",
+ "x"
+ ],
+ [
+ "▁",
+ "Box"
+ ],
+ [
+ "▁s",
+ "paces"
+ ],
+ [
+ "▁sp",
+ "aces"
+ ],
+ [
+ "▁space",
+ "s"
+ ],
+ [
+ "ex",
+ "ists"
+ ],
+ [
+ "exist",
+ "s"
+ ],
+ [
+ "▁act",
+ "ing"
+ ],
+ [
+ "▁ac",
+ "ting"
+ ],
+ [
+ "OR",
+ "Y"
+ ],
+ [
+ "зо",
+ "ва"
+ ],
+ [
+ "Go",
+ "od"
+ ],
+ [
+ "G",
+ "ood"
+ ],
+ [
+ "ien",
+ "ne"
+ ],
+ [
+ "i",
+ "enne"
+ ],
+ [
+ "▁William",
+ "s"
+ ],
+ [
+ "▁f",
+ "ruit"
+ ],
+ [
+ "▁fr",
+ "uit"
+ ],
+ [
+ "▁fru",
+ "it"
+ ],
+ [
+ "ie",
+ "ra"
+ ],
+ [
+ "ier",
+ "a"
+ ],
+ [
+ "i",
+ "era"
+ ],
+ [
+ "▁L",
+ "im"
+ ],
+ [
+ "▁Li",
+ "m"
+ ],
+ [
+ "▁",
+ "Lim"
+ ],
+ [
+ "▁t",
+ "rait"
+ ],
+ [
+ "▁tr",
+ "ait"
+ ],
+ [
+ "▁tra",
+ "it"
+ ],
+ [
+ "▁",
+ "trait"
+ ],
+ [
+ "▁art",
+ "ists"
+ ],
+ [
+ "▁artist",
+ "s"
+ ],
+ [
+ "▁ab",
+ "sor"
+ ],
+ [
+ "▁abs",
+ "or"
+ ],
+ [
+ "ra",
+ "it"
+ ],
+ [
+ "rai",
+ "t"
+ ],
+ [
+ "r",
+ "ait"
+ ],
+ [
+ "LO",
+ "AD"
+ ],
+ [
+ "▁mov",
+ "ies"
+ ],
+ [
+ "▁movie",
+ "s"
+ ],
+ [
+ "▁d",
+ "ynamic"
+ ],
+ [
+ "▁dynam",
+ "ic"
+ ],
+ [
+ "▁dyn",
+ "amic"
+ ],
+ [
+ "▁",
+ "dynamic"
+ ],
+ [
+ "as",
+ "ts"
+ ],
+ [
+ "ast",
+ "s"
+ ],
+ [
+ "a",
+ "sts"
+ ],
+ [
+ "▁In",
+ "teger"
+ ],
+ [
+ "▁",
+ "Integer"
+ ],
+ [
+ "▁sm",
+ "oke"
+ ],
+ [
+ "п",
+ "і"
+ ],
+ [
+ "an",
+ "gel"
+ ],
+ [
+ "ang",
+ "el"
+ ],
+ [
+ "ange",
+ "l"
+ ],
+ [
+ ">(",
+ "\""
+ ],
+ [
+ ">",
+ "(\""
+ ],
+ [
+ "▁in",
+ "strument"
+ ],
+ [
+ "▁instr",
+ "ument"
+ ],
+ [
+ "▁f",
+ "uel"
+ ],
+ [
+ "▁fue",
+ "l"
+ ],
+ [
+ "▁fu",
+ "el"
+ ],
+ [
+ "но",
+ "ї"
+ ],
+ [
+ "atal",
+ "ogue"
+ ],
+ [
+ "atalog",
+ "ue"
+ ],
+ [
+ "▁s",
+ "erial"
+ ],
+ [
+ "▁se",
+ "rial"
+ ],
+ [
+ "▁ser",
+ "ial"
+ ],
+ [
+ "▁",
+ "serial"
+ ],
+ [
+ "File",
+ "s"
+ ],
+ [
+ "Fil",
+ "es"
+ ],
+ [
+ "Fi",
+ "les"
+ ],
+ [
+ "F",
+ "iles"
+ ],
+ [
+ "▁bath",
+ "room"
+ ],
+ [
+ "il",
+ "o"
+ ],
+ [
+ "i",
+ "lo"
+ ],
+ [
+ "es",
+ "to"
+ ],
+ [
+ "est",
+ "o"
+ ],
+ [
+ "e",
+ "sto"
+ ],
+ [
+ "▁p",
+ "m"
+ ],
+ [
+ "▁",
+ "pm"
+ ],
+ [
+ "ent",
+ "ials"
+ ],
+ [
+ "ential",
+ "s"
+ ],
+ [
+ "enti",
+ "als"
+ ],
+ [
+ "▁On",
+ "line"
+ ],
+ [
+ "wh",
+ "ite"
+ ],
+ [
+ "▁t",
+ "ips"
+ ],
+ [
+ "▁tip",
+ "s"
+ ],
+ [
+ "▁ti",
+ "ps"
+ ],
+ [
+ "▁cap",
+ "able"
+ ],
+ [
+ "Fi",
+ "g"
+ ],
+ [
+ "F",
+ "ig"
+ ],
+ [
+ "T",
+ "V"
+ ],
+ [
+ "▁о",
+ "н"
+ ],
+ [
+ "▁",
+ "он"
+ ],
+ [
+ "k",
+ "é"
+ ],
+ [
+ "bit",
+ "r"
+ ],
+ [
+ "bi",
+ "tr"
+ ],
+ [
+ "b",
+ "itr"
+ ],
+ [
+ "Map",
+ "ping"
+ ],
+ [
+ "Ma",
+ "pping"
+ ],
+ [
+ "M",
+ "apping"
+ ],
+ [
+ "▁t",
+ "ak"
+ ],
+ [
+ "▁ta",
+ "k"
+ ],
+ [
+ "ю",
+ "щи"
+ ],
+ [
+ "в",
+ "ля"
+ ],
+ [
+ ")\"",
+ ","
+ ],
+ [
+ ")",
+ "\","
+ ],
+ [
+ "▁K",
+ "arl"
+ ],
+ [
+ "▁Kar",
+ "l"
+ ],
+ [
+ "▁Ka",
+ "rl"
+ ],
+ [
+ "▁H",
+ "uman"
+ ],
+ [
+ "▁Hu",
+ "man"
+ ],
+ [
+ "▁Hum",
+ "an"
+ ],
+ [
+ "▁P",
+ "ot"
+ ],
+ [
+ "▁Po",
+ "t"
+ ],
+ [
+ "▁rep",
+ "resents"
+ ],
+ [
+ "▁represent",
+ "s"
+ ],
+ [
+ "▁cons",
+ "istent"
+ ],
+ [
+ "▁consist",
+ "ent"
+ ],
+ [
+ "_",
+ "("
+ ],
+ [
+ "we",
+ "n"
+ ],
+ [
+ "w",
+ "en"
+ ],
+ [
+ "▁R",
+ "ose"
+ ],
+ [
+ "▁Ro",
+ "se"
+ ],
+ [
+ "▁Ros",
+ "e"
+ ],
+ [
+ "la",
+ "w"
+ ],
+ [
+ "l",
+ "aw"
+ ],
+ [
+ "▁F",
+ "ROM"
+ ],
+ [
+ "▁FR",
+ "OM"
+ ],
+ [
+ "▁",
+ "FROM"
+ ],
+ [
+ "▁beg",
+ "ins"
+ ],
+ [
+ "▁begin",
+ "s"
+ ],
+ [
+ "▁e",
+ "dit"
+ ],
+ [
+ "▁ed",
+ "it"
+ ],
+ [
+ "▁",
+ "edit"
+ ],
+ [
+ "▁mount",
+ "ain"
+ ],
+ [
+ "▁ch",
+ "apter"
+ ],
+ [
+ "▁chap",
+ "ter"
+ ],
+ [
+ "▁wonder",
+ "ed"
+ ],
+ [
+ "▁indust",
+ "rial"
+ ],
+ [
+ "▁M",
+ "ajor"
+ ],
+ [
+ "▁Ma",
+ "jor"
+ ],
+ [
+ "▁Maj",
+ "or"
+ ],
+ [
+ "▁g",
+ "es"
+ ],
+ [
+ "▁ge",
+ "s"
+ ],
+ [
+ "▁",
+ "ges"
+ ],
+ [
+ "▁direct",
+ "ed"
+ ],
+ [
+ "▁dire",
+ "cted"
+ ],
+ [
+ "er",
+ "os"
+ ],
+ [
+ "ero",
+ "s"
+ ],
+ [
+ "e",
+ "ros"
+ ],
+ [
+ "▁W",
+ "ild"
+ ],
+ [
+ "▁Wil",
+ "d"
+ ],
+ [
+ "▁Wi",
+ "ld"
+ ],
+ [
+ "li",
+ "ament"
+ ],
+ [
+ "lia",
+ "ment"
+ ],
+ [
+ "Bo",
+ "ok"
+ ],
+ [
+ "B",
+ "ook"
+ ],
+ [
+ "user",
+ "name"
+ ],
+ [
+ "ho",
+ "t"
+ ],
+ [
+ "h",
+ "ot"
+ ],
+ [
+ "▁n",
+ "am"
+ ],
+ [
+ "▁na",
+ "m"
+ ],
+ [
+ "▁",
+ "nam"
+ ],
+ [
+ "▁le",
+ "ague"
+ ],
+ [
+ "br",
+ "a"
+ ],
+ [
+ "b",
+ "ra"
+ ],
+ [
+ "ко",
+ "н"
+ ],
+ [
+ "к",
+ "он"
+ ],
+ [
+ "▁T",
+ "al"
+ ],
+ [
+ "▁Ta",
+ "l"
+ ],
+ [
+ "▁В",
+ "а"
+ ],
+ [
+ "▁ex",
+ "ports"
+ ],
+ [
+ "▁exp",
+ "orts"
+ ],
+ [
+ "▁export",
+ "s"
+ ],
+ [
+ "▁",
+ "exports"
+ ],
+ [
+ "(",
+ "@"
+ ],
+ [
+ "▁sh",
+ "aring"
+ ],
+ [
+ "▁shar",
+ "ing"
+ ],
+ [
+ "▁sha",
+ "ring"
+ ],
+ [
+ "▁T",
+ "ro"
+ ],
+ [
+ "▁Tr",
+ "o"
+ ],
+ [
+ "ś",
+ "ć"
+ ],
+ [
+ "ues",
+ "day"
+ ],
+ [
+ "yl",
+ "v"
+ ],
+ [
+ "y",
+ "lv"
+ ],
+ [
+ "▁gu",
+ "itar"
+ ],
+ [
+ "el",
+ "en"
+ ],
+ [
+ "ele",
+ "n"
+ ],
+ [
+ "e",
+ "len"
+ ],
+ [
+ "Se",
+ "lection"
+ ],
+ [
+ "Select",
+ "ion"
+ ],
+ [
+ "S",
+ "election"
+ ],
+ [
+ "▁conf",
+ "ident"
+ ],
+ [
+ "ry",
+ "pto"
+ ],
+ [
+ "rypt",
+ "o"
+ ],
+ [
+ "▁h",
+ "ors"
+ ],
+ [
+ "▁hor",
+ "s"
+ ],
+ [
+ "▁ho",
+ "rs"
+ ],
+ [
+ "ed",
+ "itor"
+ ],
+ [
+ "edit",
+ "or"
+ ],
+ [
+ "edi",
+ "tor"
+ ],
+ [
+ "▁should",
+ "ers"
+ ],
+ [
+ "▁shoulder",
+ "s"
+ ],
+ [
+ "get",
+ "Name"
+ ],
+ [
+ "en",
+ "cing"
+ ],
+ [
+ "enc",
+ "ing"
+ ],
+ [
+ "enci",
+ "ng"
+ ],
+ [
+ "SE",
+ "LECT"
+ ],
+ [
+ "SEL",
+ "ECT"
+ ],
+ [
+ "в",
+ "ши"
+ ],
+ [
+ "▁kind",
+ "s"
+ ],
+ [
+ "▁kin",
+ "ds"
+ ],
+ [
+ "▁W",
+ "el"
+ ],
+ [
+ "▁We",
+ "l"
+ ],
+ [
+ "▁pur",
+ "poses"
+ ],
+ [
+ "▁purpose",
+ "s"
+ ],
+ [
+ "Mat",
+ "rix"
+ ],
+ [
+ "in",
+ "valid"
+ ],
+ [
+ "▁own",
+ "ers"
+ ],
+ [
+ "▁owner",
+ "s"
+ ],
+ [
+ "▁",
+ "owners"
+ ],
+ [
+ "▁Rec",
+ "ords"
+ ],
+ [
+ "▁Record",
+ "s"
+ ],
+ [
+ "▁",
+ "Records"
+ ],
+ [
+ "▁Pro",
+ "cess"
+ ],
+ [
+ "▁",
+ "Process"
+ ],
+ [
+ "▁c",
+ "hat"
+ ],
+ [
+ "▁ch",
+ "at"
+ ],
+ [
+ "▁cha",
+ "t"
+ ],
+ [
+ "▁",
+ "chat"
+ ],
+ [
+ "▁D",
+ "or"
+ ],
+ [
+ "▁Do",
+ "r"
+ ],
+ [
+ "▁b",
+ "in"
+ ],
+ [
+ "▁bi",
+ "n"
+ ],
+ [
+ "▁",
+ "bin"
+ ],
+ [
+ "re",
+ "dit"
+ ],
+ [
+ "red",
+ "it"
+ ],
+ [
+ "r",
+ "edit"
+ ],
+ [
+ "oi",
+ "re"
+ ],
+ [
+ "oir",
+ "e"
+ ],
+ [
+ "o",
+ "ire"
+ ],
+ [
+ "▁T",
+ "otal"
+ ],
+ [
+ "▁To",
+ "tal"
+ ],
+ [
+ "▁Tot",
+ "al"
+ ],
+ [
+ "▁",
+ "Total"
+ ],
+ [
+ "▁F",
+ "amily"
+ ],
+ [
+ "▁Famil",
+ "y"
+ ],
+ [
+ "▁",
+ "Family"
+ ],
+ [
+ "AR",
+ "Y"
+ ],
+ [
+ "▁b",
+ "read"
+ ],
+ [
+ "▁br",
+ "ead"
+ ],
+ [
+ "▁bre",
+ "ad"
+ ],
+ [
+ "▁",
+ "bread"
+ ],
+ [
+ "▁com",
+ "pre"
+ ],
+ [
+ "▁comp",
+ "re"
+ ],
+ [
+ "▁compr",
+ "e"
+ ],
+ [
+ "▁sh",
+ "oes"
+ ],
+ [
+ "▁shoe",
+ "s"
+ ],
+ [
+ "▁r",
+ "az"
+ ],
+ [
+ "▁ra",
+ "z"
+ ],
+ [
+ "▁",
+ "raz"
+ ],
+ [
+ "▁tr",
+ "ace"
+ ],
+ [
+ "▁tra",
+ "ce"
+ ],
+ [
+ "▁",
+ "trace"
+ ],
+ [
+ "ne",
+ "j"
+ ],
+ [
+ "n",
+ "ej"
+ ],
+ [
+ "or",
+ "ted"
+ ],
+ [
+ "ort",
+ "ed"
+ ],
+ [
+ "orte",
+ "d"
+ ],
+ [
+ "h",
+ "n"
+ ],
+ [
+ "▁pro",
+ "cedure"
+ ],
+ [
+ "▁proced",
+ "ure"
+ ],
+ [
+ "pro",
+ "perties"
+ ],
+ [
+ "pl",
+ "ier"
+ ],
+ [
+ "▁h",
+ "ero"
+ ],
+ [
+ "▁he",
+ "ro"
+ ],
+ [
+ "▁her",
+ "o"
+ ],
+ [
+ "▁",
+ "hero"
+ ],
+ [
+ "pan",
+ "el"
+ ],
+ [
+ "pa",
+ "nel"
+ ],
+ [
+ "p",
+ "anel"
+ ],
+ [
+ "▁mark",
+ "ed"
+ ],
+ [
+ "▁mar",
+ "ked"
+ ],
+ [
+ "▁wor",
+ "ried"
+ ],
+ [
+ "\\",
+ "|"
+ ],
+ [
+ "pt",
+ "s"
+ ],
+ [
+ "p",
+ "ts"
+ ],
+ [
+ "▁S",
+ "upport"
+ ],
+ [
+ "▁Sup",
+ "port"
+ ],
+ [
+ "▁Supp",
+ "ort"
+ ],
+ [
+ "▁",
+ "Support"
+ ],
+ [
+ "▁ser",
+ "ving"
+ ],
+ [
+ "▁serv",
+ "ing"
+ ],
+ [
+ "F",
+ "ail"
+ ],
+ [
+ "▁dis",
+ "appoint"
+ ],
+ [
+ "▁Sc",
+ "ot"
+ ],
+ [
+ "▁ple",
+ "asure"
+ ],
+ [
+ "▁j",
+ "udge"
+ ],
+ [
+ "▁jud",
+ "ge"
+ ],
+ [
+ "▁judg",
+ "e"
+ ],
+ [
+ "ze",
+ "ich"
+ ],
+ [
+ "▁for",
+ "ever"
+ ],
+ [
+ "▁fore",
+ "ver"
+ ],
+ [
+ "▁Ze",
+ "it"
+ ],
+ [
+ "uo",
+ "us"
+ ],
+ [
+ "u",
+ "ous"
+ ],
+ [
+ "in",
+ "ent"
+ ],
+ [
+ "ine",
+ "nt"
+ ],
+ [
+ "inen",
+ "t"
+ ],
+ [
+ "i",
+ "nent"
+ ],
+ [
+ "▁d",
+ "w"
+ ],
+ [
+ "▁",
+ "dw"
+ ],
+ [
+ "▁w",
+ "aren"
+ ],
+ [
+ "▁war",
+ "en"
+ ],
+ [
+ "▁wa",
+ "ren"
+ ],
+ [
+ "▁ware",
+ "n"
+ ],
+ [
+ "▁fl",
+ "ash"
+ ],
+ [
+ "▁",
+ "flash"
+ ],
+ [
+ "▁tro",
+ "ops"
+ ],
+ [
+ "▁dr",
+ "ugs"
+ ],
+ [
+ "▁dru",
+ "gs"
+ ],
+ [
+ "▁drug",
+ "s"
+ ],
+ [
+ "▁d",
+ "iam"
+ ],
+ [
+ "▁di",
+ "am"
+ ],
+ [
+ "▁dia",
+ "m"
+ ],
+ [
+ ".",
+ "~"
+ ],
+ [
+ "im",
+ "p"
+ ],
+ [
+ "i",
+ "mp"
+ ],
+ [
+ "in",
+ "ned"
+ ],
+ [
+ "inn",
+ "ed"
+ ],
+ [
+ "▁E",
+ "V"
+ ],
+ [
+ "▁",
+ "EV"
+ ],
+ [
+ "St",
+ "ruct"
+ ],
+ [
+ "Str",
+ "uct"
+ ],
+ [
+ "▁just",
+ "ice"
+ ],
+ [
+ "▁offic",
+ "ials"
+ ],
+ [
+ "▁official",
+ "s"
+ ],
+ [
+ "ff",
+ "ff"
+ ],
+ [
+ "fff",
+ "f"
+ ],
+ [
+ "f",
+ "fff"
+ ],
+ [
+ "▁Com",
+ "mon"
+ ],
+ [
+ "▁Comm",
+ "on"
+ ],
+ [
+ "▁",
+ "Common"
+ ],
+ [
+ "▁C",
+ "at"
+ ],
+ [
+ "▁Ca",
+ "t"
+ ],
+ [
+ "▁",
+ "Cat"
+ ],
+ [
+ "▁tom",
+ "orrow"
+ ],
+ [
+ "▁é",
+ "l"
+ ],
+ [
+ "▁",
+ "él"
+ ],
+ [
+ "Text",
+ "ure"
+ ],
+ [
+ "Te",
+ "xture"
+ ],
+ [
+ "qp",
+ "oint"
+ ],
+ [
+ "q",
+ "point"
+ ],
+ [
+ "▁F",
+ "ried"
+ ],
+ [
+ "▁Fr",
+ "ied"
+ ],
+ [
+ "▁T",
+ "erm"
+ ],
+ [
+ "▁Te",
+ "rm"
+ ],
+ [
+ "▁Ter",
+ "m"
+ ],
+ [
+ "▁",
+ "Term"
+ ],
+ [
+ "pgf",
+ "qpoint"
+ ],
+ [
+ "▁n",
+ "em"
+ ],
+ [
+ "▁ne",
+ "m"
+ ],
+ [
+ "▁",
+ "nem"
+ ],
+ [
+ "no",
+ "rm"
+ ],
+ [
+ "nor",
+ "m"
+ ],
+ [
+ "n",
+ "orm"
+ ],
+ [
+ "▁hard",
+ "ly"
+ ],
+ [
+ "od",
+ "a"
+ ],
+ [
+ "o",
+ "da"
+ ],
+ [
+ "ze",
+ "ta"
+ ],
+ [
+ "zet",
+ "a"
+ ],
+ [
+ "z",
+ "eta"
+ ],
+ [
+ "em",
+ "ic"
+ ],
+ [
+ "emi",
+ "c"
+ ],
+ [
+ "e",
+ "mic"
+ ],
+ [
+ "▁по",
+ "лу"
+ ],
+ [
+ "▁пол",
+ "у"
+ ],
+ [
+ "▁lo",
+ "aded"
+ ],
+ [
+ "▁load",
+ "ed"
+ ],
+ [
+ "▁",
+ "loaded"
+ ],
+ [
+ "ke",
+ "s"
+ ],
+ [
+ "k",
+ "es"
+ ],
+ [
+ "ci",
+ "ó"
+ ],
+ [
+ "c",
+ "ió"
+ ],
+ [
+ "▁f",
+ "ool"
+ ],
+ [
+ "▁fo",
+ "ol"
+ ],
+ [
+ "▁foo",
+ "l"
+ ],
+ [
+ "▁t",
+ "rick"
+ ],
+ [
+ "▁tr",
+ "ick"
+ ],
+ [
+ "▁tri",
+ "ck"
+ ],
+ [
+ "▁d",
+ "st"
+ ],
+ [
+ "▁ds",
+ "t"
+ ],
+ [
+ "▁",
+ "dst"
+ ],
+ [
+ "Fin",
+ "d"
+ ],
+ [
+ "Fi",
+ "nd"
+ ],
+ [
+ "F",
+ "ind"
+ ],
+ [
+ "▁в",
+ "се"
+ ],
+ [
+ "}}",
+ ","
+ ],
+ [
+ "}",
+ "},"
+ ],
+ [
+ "▁frame",
+ "work"
+ ],
+ [
+ "▁",
+ "framework"
+ ],
+ [
+ "▁mer",
+ "ely"
+ ],
+ [
+ "▁mere",
+ "ly"
+ ],
+ [
+ "▁un",
+ "ion"
+ ],
+ [
+ "▁",
+ "union"
+ ],
+ [
+ "▁Ed",
+ "ward"
+ ],
+ [
+ "ri",
+ "f"
+ ],
+ [
+ "r",
+ "if"
+ ],
+ [
+ "Fl",
+ "ag"
+ ],
+ [
+ "F",
+ "lag"
+ ],
+ [
+ "▁cris",
+ "is"
+ ],
+ [
+ "▁cri",
+ "sis"
+ ],
+ [
+ "▁fin",
+ "ite"
+ ],
+ [
+ "▁",
+ "finite"
+ ],
+ [
+ "▁l",
+ "ol"
+ ],
+ [
+ "▁lo",
+ "l"
+ ],
+ [
+ "▁K",
+ "im"
+ ],
+ [
+ "▁Ki",
+ "m"
+ ],
+ [
+ "на",
+ "та"
+ ],
+ [
+ "sin",
+ "ce"
+ ],
+ [
+ "s",
+ "ince"
+ ],
+ [
+ "▁com",
+ "pat"
+ ],
+ [
+ "▁comp",
+ "at"
+ ],
+ [
+ "▁",
+ "compat"
+ ],
+ [
+ "▁p",
+ "ert"
+ ],
+ [
+ "▁per",
+ "t"
+ ],
+ [
+ "▁pe",
+ "rt"
+ ],
+ [
+ "▁",
+ "pert"
+ ],
+ [
+ "ib",
+ "ilities"
+ ],
+ [
+ "ibil",
+ "ities"
+ ],
+ [
+ "▁tamb",
+ "ién"
+ ],
+ [
+ "ib",
+ "li"
+ ],
+ [
+ "▁t",
+ "een"
+ ],
+ [
+ "▁te",
+ "en"
+ ],
+ [
+ "▁",
+ "teen"
+ ],
+ [
+ "▁sym",
+ "pt"
+ ],
+ [
+ "or",
+ "al"
+ ],
+ [
+ "ora",
+ "l"
+ ],
+ [
+ "o",
+ "ral"
+ ],
+ [
+ "de",
+ "rs"
+ ],
+ [
+ "der",
+ "s"
+ ],
+ [
+ "d",
+ "ers"
+ ],
+ [
+ "ot",
+ "te"
+ ],
+ [
+ "ott",
+ "e"
+ ],
+ [
+ "п",
+ "ри"
+ ],
+ [
+ "▁J",
+ "ane"
+ ],
+ [
+ "▁Jan",
+ "e"
+ ],
+ [
+ "▁Ja",
+ "ne"
+ ],
+ [
+ "▁original",
+ "ly"
+ ],
+ [
+ "▁origin",
+ "ally"
+ ],
+ [
+ "▁thro",
+ "at"
+ ],
+ [
+ "ma",
+ "g"
+ ],
+ [
+ "m",
+ "ag"
+ ],
+ [
+ "su",
+ "p"
+ ],
+ [
+ "s",
+ "up"
+ ],
+ [
+ "un",
+ "i"
+ ],
+ [
+ "u",
+ "ni"
+ ],
+ [
+ "$",
+ "$"
+ ],
+ [
+ "▁L",
+ "ibrary"
+ ],
+ [
+ "▁",
+ "Library"
+ ],
+ [
+ "▁att",
+ "acks"
+ ],
+ [
+ "▁attack",
+ "s"
+ ],
+ [
+ "in",
+ "gen"
+ ],
+ [
+ "ing",
+ "en"
+ ],
+ [
+ "inge",
+ "n"
+ ],
+ [
+ "('",
+ "/"
+ ],
+ [
+ "▁h",
+ "es"
+ ],
+ [
+ "▁he",
+ "s"
+ ],
+ [
+ "▁",
+ "hes"
+ ],
+ [
+ "co",
+ "in"
+ ],
+ [
+ "c",
+ "oin"
+ ],
+ [
+ "oun",
+ "ce"
+ ],
+ [
+ "▁Academ",
+ "y"
+ ],
+ [
+ "MOD",
+ "ULE"
+ ],
+ [
+ "is",
+ "ms"
+ ],
+ [
+ "ism",
+ "s"
+ ],
+ [
+ "▁A",
+ "dv"
+ ],
+ [
+ "▁Ad",
+ "v"
+ ],
+ [
+ "▁",
+ "Adv"
+ ],
+ [
+ "▁B",
+ "ol"
+ ],
+ [
+ "▁Bo",
+ "l"
+ ],
+ [
+ "▁inc",
+ "ident"
+ ],
+ [
+ ")^",
+ "{"
+ ],
+ [
+ ")",
+ "^{"
+ ],
+ [
+ "▁b",
+ "ij"
+ ],
+ [
+ "▁bi",
+ "j"
+ ],
+ [
+ "▁R",
+ "ome"
+ ],
+ [
+ "▁Rom",
+ "e"
+ ],
+ [
+ "▁Ro",
+ "me"
+ ],
+ [
+ "▁It",
+ "aly"
+ ],
+ [
+ "▁Ital",
+ "y"
+ ],
+ [
+ "ev",
+ "ents"
+ ],
+ [
+ "event",
+ "s"
+ ],
+ [
+ "even",
+ "ts"
+ ],
+ [
+ "▁F",
+ "ern"
+ ],
+ [
+ "▁Fe",
+ "rn"
+ ],
+ [
+ "▁Fer",
+ "n"
+ ],
+ [
+ "▁b",
+ "er"
+ ],
+ [
+ "▁be",
+ "r"
+ ],
+ [
+ "▁",
+ "ber"
+ ],
+ [
+ "▁sil",
+ "ent"
+ ],
+ [
+ "▁p",
+ "ier"
+ ],
+ [
+ "▁pie",
+ "r"
+ ],
+ [
+ "▁pi",
+ "er"
+ ],
+ [
+ "▁Y",
+ "O"
+ ],
+ [
+ "▁pl",
+ "ain"
+ ],
+ [
+ "▁",
+ "plain"
+ ],
+ [
+ "B",
+ "as"
+ ],
+ [
+ "▁p",
+ "ill"
+ ],
+ [
+ "▁pi",
+ "ll"
+ ],
+ [
+ "▁pil",
+ "l"
+ ],
+ [
+ "ra",
+ "se"
+ ],
+ [
+ "ras",
+ "e"
+ ],
+ [
+ "r",
+ "ase"
+ ],
+ [
+ "▁car",
+ "rying"
+ ],
+ [
+ "▁carry",
+ "ing"
+ ],
+ [
+ "▁re",
+ "sp"
+ ],
+ [
+ "▁r",
+ "esp"
+ ],
+ [
+ "▁res",
+ "p"
+ ],
+ [
+ "▁",
+ "resp"
+ ],
+ [
+ "ну",
+ "ю"
+ ],
+ [
+ "▁typ",
+ "ical"
+ ],
+ [
+ "Wrap",
+ "per"
+ ],
+ [
+ "W",
+ "rapper"
+ ],
+ [
+ "▁g",
+ "au"
+ ],
+ [
+ "▁ga",
+ "u"
+ ],
+ [
+ "▁chem",
+ "ical"
+ ],
+ [
+ "▁h",
+ "al"
+ ],
+ [
+ "▁ha",
+ "l"
+ ],
+ [
+ "▁",
+ "hal"
+ ],
+ [
+ "th",
+ "row"
+ ],
+ [
+ "Cl",
+ "uster"
+ ],
+ [
+ "▁G",
+ "ab"
+ ],
+ [
+ "▁Ga",
+ "b"
+ ],
+ [
+ "▁G",
+ "irl"
+ ],
+ [
+ "▁Gi",
+ "rl"
+ ],
+ [
+ "▁Gir",
+ "l"
+ ],
+ [
+ "qu",
+ "ir"
+ ],
+ [
+ "▁A",
+ "rg"
+ ],
+ [
+ "▁Ar",
+ "g"
+ ],
+ [
+ "▁",
+ "Arg"
+ ],
+ [
+ "▁rel",
+ "ief"
+ ],
+ [
+ "▁relie",
+ "f"
+ ],
+ [
+ "▁reli",
+ "ef"
+ ],
+ [
+ "▁В",
+ "е"
+ ],
+ [
+ "d",
+ "m"
+ ],
+ [
+ "▁fr",
+ "ustr"
+ ],
+ [
+ "▁fru",
+ "str"
+ ],
+ [
+ "\\",
+ "%"
+ ],
+ [
+ "▁st",
+ "ores"
+ ],
+ [
+ "▁store",
+ "s"
+ ],
+ [
+ "▁stor",
+ "es"
+ ],
+ [
+ "▁sto",
+ "res"
+ ],
+ [
+ "▁bott",
+ "le"
+ ],
+ [
+ "▁bot",
+ "tle"
+ ],
+ [
+ "▁L",
+ "ew"
+ ],
+ [
+ "▁Le",
+ "w"
+ ],
+ [
+ "tw",
+ "o"
+ ],
+ [
+ "t",
+ "wo"
+ ],
+ [
+ "st",
+ "ad"
+ ],
+ [
+ "sta",
+ "d"
+ ],
+ [
+ "▁che",
+ "ek"
+ ],
+ [
+ "▁concern",
+ "s"
+ ],
+ [
+ "▁concer",
+ "ns"
+ ],
+ [
+ "▁help",
+ "ful"
+ ],
+ [
+ "▁co",
+ "verage"
+ ],
+ [
+ "▁cover",
+ "age"
+ ],
+ [
+ "is",
+ "i"
+ ],
+ [
+ "i",
+ "si"
+ ],
+ [
+ "AD",
+ "D"
+ ],
+ [
+ "A",
+ "DD"
+ ],
+ [
+ "as",
+ "ync"
+ ],
+ [
+ "asy",
+ "nc"
+ ],
+ [
+ "a",
+ "sync"
+ ],
+ [
+ "▁approxim",
+ "ately"
+ ],
+ [
+ "▁approx",
+ "imately"
+ ],
+ [
+ "▁approximate",
+ "ly"
+ ],
+ [
+ "if",
+ "fer"
+ ],
+ [
+ "iff",
+ "er"
+ ],
+ [
+ "iffe",
+ "r"
+ ],
+ [
+ "ho",
+ "ok"
+ ],
+ [
+ "h",
+ "ook"
+ ],
+ [
+ "▁e",
+ "num"
+ ],
+ [
+ "▁en",
+ "um"
+ ],
+ [
+ "▁",
+ "enum"
+ ],
+ [
+ "ov",
+ "á"
+ ],
+ [
+ "o",
+ "vá"
+ ],
+ [
+ "▁e",
+ "vil"
+ ],
+ [
+ "▁ev",
+ "il"
+ ],
+ [
+ "▁const",
+ "antly"
+ ],
+ [
+ "▁constant",
+ "ly"
+ ],
+ [
+ "ap",
+ "ply"
+ ],
+ [
+ "app",
+ "ly"
+ ],
+ [
+ "▁si",
+ "è"
+ ],
+ [
+ "▁pract",
+ "ices"
+ ],
+ [
+ "▁practice",
+ "s"
+ ],
+ [
+ "▁te",
+ "achers"
+ ],
+ [
+ "▁teach",
+ "ers"
+ ],
+ [
+ "▁teacher",
+ "s"
+ ],
+ [
+ "▁S",
+ "n"
+ ],
+ [
+ "▁",
+ "Sn"
+ ],
+ [
+ "▁A",
+ "wards"
+ ],
+ [
+ "▁Award",
+ "s"
+ ],
+ [
+ "▁Aw",
+ "ards"
+ ],
+ [
+ "▁sub",
+ "stant"
+ ],
+ [
+ "▁subst",
+ "ant"
+ ],
+ [
+ "▁$",
+ "."
+ ],
+ [
+ "▁",
+ "$."
+ ],
+ [
+ "d",
+ "k"
+ ],
+ [
+ "▁m",
+ "ob"
+ ],
+ [
+ "▁mo",
+ "b"
+ ],
+ [
+ "▁",
+ "mob"
+ ],
+ [
+ "▁ing",
+ "red"
+ ],
+ [
+ "ve",
+ "re"
+ ],
+ [
+ "ver",
+ "e"
+ ],
+ [
+ "v",
+ "ere"
+ ],
+ [
+ "Mult",
+ "i"
+ ],
+ [
+ "пе",
+ "р"
+ ],
+ [
+ "п",
+ "ер"
+ ],
+ [
+ "st",
+ "al"
+ ],
+ [
+ "sta",
+ "l"
+ ],
+ [
+ "s",
+ "tal"
+ ],
+ [
+ "ya",
+ "rd"
+ ],
+ [
+ "yar",
+ "d"
+ ],
+ [
+ "y",
+ "ard"
+ ],
+ [
+ "requ",
+ "ired"
+ ],
+ [
+ "require",
+ "d"
+ ],
+ [
+ "ve",
+ "ment"
+ ],
+ [
+ "v",
+ "ement"
+ ],
+ [
+ "▁int",
+ "elligence"
+ ],
+ [
+ "▁intellig",
+ "ence"
+ ],
+ [
+ "▁th",
+ "inks"
+ ],
+ [
+ "▁think",
+ "s"
+ ],
+ [
+ "▁thin",
+ "ks"
+ ],
+ [
+ "▁person",
+ "ally"
+ ],
+ [
+ "▁personal",
+ "ly"
+ ],
+ [
+ "▁tr",
+ "ained"
+ ],
+ [
+ "▁tra",
+ "ined"
+ ],
+ [
+ "▁train",
+ "ed"
+ ],
+ [
+ "▁",
+ "trained"
+ ],
+ [
+ "or",
+ "ney"
+ ],
+ [
+ "orn",
+ "ey"
+ ],
+ [
+ "orne",
+ "y"
+ ],
+ [
+ ")",
+ ""
+ ],
+ [
+ "gg",
+ "ed"
+ ],
+ [
+ "g",
+ "ged"
+ ],
+ [
+ "E",
+ "INVAL"
+ ],
+ [
+ "ar",
+ "na"
+ ],
+ [
+ "arn",
+ "a"
+ ],
+ [
+ "▁Ham",
+ "ilton"
+ ],
+ [
+ "mer",
+ "ce"
+ ],
+ [
+ "ek",
+ "t"
+ ],
+ [
+ "e",
+ "kt"
+ ],
+ [
+ "O",
+ "F"
+ ],
+ [
+ ")",
+ "["
+ ],
+ [
+ "ru",
+ "g"
+ ],
+ [
+ "r",
+ "ug"
+ ],
+ [
+ "ic",
+ "ión"
+ ],
+ [
+ "ici",
+ "ón"
+ ],
+ [
+ "ició",
+ "n"
+ ],
+ [
+ "i",
+ "ción"
+ ],
+ [
+ "▁sur",
+ "vey"
+ ],
+ [
+ "▁surv",
+ "ey"
+ ],
+ [
+ "▁surve",
+ "y"
+ ],
+ [
+ "nes",
+ "day"
+ ],
+ [
+ "▁p",
+ "ag"
+ ],
+ [
+ "▁pa",
+ "g"
+ ],
+ [
+ "▁",
+ "pag"
+ ],
+ [
+ "▁bound",
+ "ary"
+ ],
+ [
+ "▁quant",
+ "um"
+ ],
+ [
+ "▁draw",
+ "ing"
+ ],
+ [
+ "▁vol",
+ "unte"
+ ],
+ [
+ "▁volunt",
+ "e"
+ ],
+ [
+ "▁W",
+ "ord"
+ ],
+ [
+ "▁Wo",
+ "rd"
+ ],
+ [
+ "▁Wor",
+ "d"
+ ],
+ [
+ "▁",
+ "Word"
+ ],
+ [
+ "sk",
+ "y"
+ ],
+ [
+ "s",
+ "ky"
+ ],
+ [
+ "▁G",
+ "reg"
+ ],
+ [
+ "▁Gr",
+ "eg"
+ ],
+ [
+ "▁Gre",
+ "g"
+ ],
+ [
+ "co",
+ "ll"
+ ],
+ [
+ "col",
+ "l"
+ ],
+ [
+ "c",
+ "oll"
+ ],
+ [
+ "hi",
+ "de"
+ ],
+ [
+ "hid",
+ "e"
+ ],
+ [
+ "h",
+ "ide"
+ ],
+ [
+ "▁sw",
+ "im"
+ ],
+ [
+ "▁reve",
+ "aled"
+ ],
+ [
+ "▁reveal",
+ "ed"
+ ],
+ [
+ "ad",
+ "v"
+ ],
+ [
+ "a",
+ "dv"
+ ],
+ [
+ "д",
+ "я"
+ ],
+ [
+ ".\"",
+ ");"
+ ],
+ [
+ ".\")",
+ ";"
+ ],
+ [
+ ".",
+ "\");"
+ ],
+ [
+ "▁ex",
+ "plan"
+ ],
+ [
+ "▁expl",
+ "an"
+ ],
+ [
+ "▁exp",
+ "lan"
+ ],
+ [
+ "▁Cur",
+ "rent"
+ ],
+ [
+ "▁",
+ "Current"
+ ],
+ [
+ "▁got",
+ "ten"
+ ],
+ [
+ "▁f",
+ "alling"
+ ],
+ [
+ "▁fall",
+ "ing"
+ ],
+ [
+ "▁fal",
+ "ling"
+ ],
+ [
+ "▁cont",
+ "ained"
+ ],
+ [
+ "▁contain",
+ "ed"
+ ],
+ [
+ "UN",
+ "D"
+ ],
+ [
+ "U",
+ "ND"
+ ],
+ [
+ "▁Sh",
+ "ould"
+ ],
+ [
+ "▁",
+ "Should"
+ ],
+ [
+ "▁k",
+ "illing"
+ ],
+ [
+ "▁kill",
+ "ing"
+ ],
+ [
+ "▁kil",
+ "ling"
+ ],
+ [
+ "▁aspect",
+ "s"
+ ],
+ [
+ "ic",
+ "ted"
+ ],
+ [
+ "ict",
+ "ed"
+ ],
+ [
+ "i",
+ "cted"
+ ],
+ [
+ "▁P",
+ "aram"
+ ],
+ [
+ "▁Par",
+ "am"
+ ],
+ [
+ "▁Pa",
+ "ram"
+ ],
+ [
+ "▁Para",
+ "m"
+ ],
+ [
+ "▁",
+ "Param"
+ ],
+ [
+ "\",",
+ "\r"
+ ],
+ [
+ "\"",
+ ",\r"
+ ],
+ [
+ "TI",
+ "ON"
+ ],
+ [
+ "T",
+ "ION"
+ ],
+ [
+ "))",
+ ";\r"
+ ],
+ [
+ "));",
+ "\r"
+ ],
+ [
+ ")",
+ ");\r"
+ ],
+ [
+ "▁I",
+ "ran"
+ ],
+ [
+ "▁Ir",
+ "an"
+ ],
+ [
+ "▁Ira",
+ "n"
+ ],
+ [
+ "be",
+ "it"
+ ],
+ [
+ "▁B",
+ "u"
+ ],
+ [
+ "▁",
+ "Bu"
+ ],
+ [
+ "▁[",
+ "],"
+ ],
+ [
+ "▁[]",
+ ","
+ ],
+ [
+ "▁",
+ "[],"
+ ],
+ [
+ "SS",
+ "ION"
+ ],
+ [
+ "S",
+ "SION"
+ ],
+ [
+ "▁M",
+ "ah"
+ ],
+ [
+ "▁Ma",
+ "h"
+ ],
+ [
+ "▁res",
+ "olution"
+ ],
+ [
+ "▁b",
+ "oss"
+ ],
+ [
+ "▁bo",
+ "ss"
+ ],
+ [
+ "▁bos",
+ "s"
+ ],
+ [
+ "l",
+ "g"
+ ],
+ [
+ "ch",
+ "or"
+ ],
+ [
+ "cho",
+ "r"
+ ],
+ [
+ "c",
+ "hor"
+ ],
+ [
+ "▁Un",
+ "ter"
+ ],
+ [
+ "▁de",
+ "bt"
+ ],
+ [
+ "▁deb",
+ "t"
+ ],
+ [
+ "▁v",
+ "id"
+ ],
+ [
+ "▁vi",
+ "d"
+ ],
+ [
+ "▁",
+ "vid"
+ ],
+ [
+ "gi",
+ "e"
+ ],
+ [
+ "g",
+ "ie"
+ ],
+ [
+ "▁u",
+ "no"
+ ],
+ [
+ "▁un",
+ "o"
+ ],
+ [
+ "▁",
+ "uno"
+ ],
+ [
+ "C",
+ "B"
+ ],
+ [
+ "pl",
+ "om"
+ ],
+ [
+ "plo",
+ "m"
+ ],
+ [
+ "LIC",
+ "ENSE"
+ ],
+ [
+ "L",
+ "ICENSE"
+ ],
+ [
+ "▁K",
+ "enn"
+ ],
+ [
+ "▁Ke",
+ "nn"
+ ],
+ [
+ "▁Ken",
+ "n"
+ ],
+ [
+ "▁fin",
+ "ns"
+ ],
+ [
+ "ON",
+ "G"
+ ],
+ [
+ "O",
+ "NG"
+ ],
+ [
+ "▁some",
+ "what"
+ ],
+ [
+ "▁a",
+ "ctor"
+ ],
+ [
+ "▁act",
+ "or"
+ ],
+ [
+ "▁ac",
+ "tor"
+ ],
+ [
+ "▁",
+ "actor"
+ ],
+ [
+ "▁St",
+ "atus"
+ ],
+ [
+ "▁Stat",
+ "us"
+ ],
+ [
+ "▁",
+ "Status"
+ ],
+ [
+ "▁prob",
+ "ability"
+ ],
+ [
+ "f",
+ "b"
+ ],
+ [
+ "▁ch",
+ "art"
+ ],
+ [
+ "▁char",
+ "t"
+ ],
+ [
+ "▁cha",
+ "rt"
+ ],
+ [
+ "▁",
+ "chart"
+ ],
+ [
+ "▁st",
+ "ands"
+ ],
+ [
+ "▁stand",
+ "s"
+ ],
+ [
+ "▁stan",
+ "ds"
+ ],
+ [
+ "pol",
+ "icy"
+ ],
+ [
+ "▁o",
+ "nder"
+ ],
+ [
+ "▁on",
+ "der"
+ ],
+ [
+ "▁onde",
+ "r"
+ ],
+ [
+ "▁",
+ "onder"
+ ],
+ [
+ "tab",
+ "ular"
+ ],
+ [
+ "▁A",
+ "sh"
+ ],
+ [
+ "▁As",
+ "h"
+ ],
+ [
+ "▁bo",
+ "ost"
+ ],
+ [
+ "▁",
+ "boost"
+ ],
+ [
+ "▁des",
+ "per"
+ ],
+ [
+ "▁desp",
+ "er"
+ ],
+ [
+ "mon",
+ "th"
+ ],
+ [
+ "mont",
+ "h"
+ ],
+ [
+ "▁al",
+ "ert"
+ ],
+ [
+ "▁ale",
+ "rt"
+ ],
+ [
+ "▁",
+ "alert"
+ ],
+ [
+ "▁su",
+ "ite"
+ ],
+ [
+ "▁suit",
+ "e"
+ ],
+ [
+ "▁",
+ "suite"
+ ],
+ [
+ "▁g",
+ "én"
+ ],
+ [
+ "▁gé",
+ "n"
+ ],
+ [
+ "▁v",
+ "acc"
+ ],
+ [
+ "▁va",
+ "cc"
+ ],
+ [
+ "▁vac",
+ "c"
+ ],
+ [
+ "▁H",
+ "as"
+ ],
+ [
+ "▁Ha",
+ "s"
+ ],
+ [
+ "▁",
+ "Has"
+ ],
+ [
+ "Ma",
+ "sk"
+ ],
+ [
+ "M",
+ "ask"
+ ],
+ [
+ "▁Th",
+ "ursday"
+ ],
+ [
+ "▁pro",
+ "ved"
+ ],
+ [
+ "▁pr",
+ "oved"
+ ],
+ [
+ "▁prov",
+ "ed"
+ ],
+ [
+ "▁prove",
+ "d"
+ ],
+ [
+ "▁N",
+ "el"
+ ],
+ [
+ "▁Ne",
+ "l"
+ ],
+ [
+ "▁m",
+ "oral"
+ ],
+ [
+ "▁mor",
+ "al"
+ ],
+ [
+ "▁mo",
+ "ral"
+ ],
+ [
+ "▁j",
+ "a"
+ ],
+ [
+ "▁",
+ "ja"
+ ],
+ [
+ "au",
+ "er"
+ ],
+ [
+ "a",
+ "uer"
+ ],
+ [
+ "co",
+ "dec"
+ ],
+ [
+ "code",
+ "c"
+ ],
+ [
+ "cod",
+ "ec"
+ ],
+ [
+ "▁in",
+ "stant"
+ ],
+ [
+ "▁inst",
+ "ant"
+ ],
+ [
+ "am",
+ "ps"
+ ],
+ [
+ "amp",
+ "s"
+ ],
+ [
+ "▁mil",
+ "k"
+ ],
+ [
+ "WO",
+ "RD"
+ ],
+ [
+ "W",
+ "ORD"
+ ],
+ [
+ "▁",
+ "Ö"
+ ],
+ [
+ "Em",
+ "ail"
+ ],
+ [
+ "E",
+ "mail"
+ ],
+ [
+ "Element",
+ "s"
+ ],
+ [
+ "El",
+ "ements"
+ ],
+ [
+ "Elem",
+ "ents"
+ ],
+ [
+ "▁for",
+ "ma"
+ ],
+ [
+ "▁form",
+ "a"
+ ],
+ [
+ "Fr",
+ "ee"
+ ],
+ [
+ "F",
+ "ree"
+ ],
+ [
+ "MA",
+ "P"
+ ],
+ [
+ "M",
+ "AP"
+ ],
+ [
+ "▁",
+ "Ж"
+ ],
+ [
+ "sy",
+ "m"
+ ],
+ [
+ "s",
+ "ym"
+ ],
+ [
+ "▁т",
+ "и"
+ ],
+ [
+ "▁",
+ "ти"
+ ],
+ [
+ "▁E",
+ "conom"
+ ],
+ [
+ "▁Ec",
+ "onom"
+ ],
+ [
+ "▁V",
+ "i"
+ ],
+ [
+ "▁",
+ "Vi"
+ ],
+ [
+ "▁Col",
+ "umb"
+ ],
+ [
+ "▁_",
+ ","
+ ],
+ [
+ "▁",
+ "_,"
+ ],
+ [
+ "or",
+ "et"
+ ],
+ [
+ "ore",
+ "t"
+ ],
+ [
+ "o",
+ "ret"
+ ],
+ [
+ "Se",
+ "qu"
+ ],
+ [
+ "Seq",
+ "u"
+ ],
+ [
+ "S",
+ "equ"
+ ],
+ [
+ "pl",
+ "an"
+ ],
+ [
+ "p",
+ "lan"
+ ],
+ [
+ "▁f",
+ "requency"
+ ],
+ [
+ "▁frequ",
+ "ency"
+ ],
+ [
+ "▁",
+ "frequency"
+ ],
+ [
+ "ir",
+ "ement"
+ ],
+ [
+ "ire",
+ "ment"
+ ],
+ [
+ "▁ass",
+ "umed"
+ ],
+ [
+ "▁assum",
+ "ed"
+ ],
+ [
+ "▁assume",
+ "d"
+ ],
+ [
+ "▁C",
+ "a"
+ ],
+ [
+ "▁B",
+ "it"
+ ],
+ [
+ "▁Bi",
+ "t"
+ ],
+ [
+ "▁",
+ "Bit"
+ ],
+ [
+ "▁ко",
+ "ман"
+ ],
+ [
+ "▁ком",
+ "ан"
+ ],
+ [
+ "▁sm",
+ "ell"
+ ],
+ [
+ "Se",
+ "curity"
+ ],
+ [
+ "Sec",
+ "urity"
+ ],
+ [
+ "▁a",
+ "qu"
+ ],
+ [
+ "▁",
+ "aqu"
+ ],
+ [
+ "oo",
+ "r"
+ ],
+ [
+ "o",
+ "or"
+ ],
+ [
+ "pr",
+ "ice"
+ ],
+ [
+ "p",
+ "rice"
+ ],
+ [
+ "in",
+ "ity"
+ ],
+ [
+ "init",
+ "y"
+ ],
+ [
+ "ini",
+ "ty"
+ ],
+ [
+ "▁a",
+ "xis"
+ ],
+ [
+ "▁ax",
+ "is"
+ ],
+ [
+ "▁",
+ "axis"
+ ],
+ [
+ "re",
+ "lease"
+ ],
+ [
+ "▁res",
+ "olve"
+ ],
+ [
+ "▁",
+ "resolve"
+ ],
+ [
+ "▁t",
+ "ears"
+ ],
+ [
+ "▁te",
+ "ars"
+ ],
+ [
+ "▁tea",
+ "rs"
+ ],
+ [
+ "▁tear",
+ "s"
+ ],
+ [
+ "▁b",
+ "other"
+ ],
+ [
+ "▁bo",
+ "ther"
+ ],
+ [
+ "▁both",
+ "er"
+ ],
+ [
+ "▁bot",
+ "her"
+ ],
+ [
+ "▁Comm",
+ "unity"
+ ],
+ [
+ "▁Commun",
+ "ity"
+ ],
+ [
+ "▁register",
+ "ed"
+ ],
+ [
+ "▁re",
+ "volution"
+ ],
+ [
+ "▁rev",
+ "olution"
+ ],
+ [
+ "▁revol",
+ "ution"
+ ],
+ [
+ "?",
+ "."
+ ],
+ [
+ "▁version",
+ "s"
+ ],
+ [
+ "▁vers",
+ "ions"
+ ],
+ [
+ "▁",
+ "versions"
+ ],
+ [
+ "%%",
+ "%%"
+ ],
+ [
+ "yd",
+ "ro"
+ ],
+ [
+ "y",
+ "dro"
+ ],
+ [
+ "Su",
+ "ccess"
+ ],
+ [
+ "▁W",
+ "in"
+ ],
+ [
+ "▁Wi",
+ "n"
+ ],
+ [
+ "▁",
+ "Win"
+ ],
+ [
+ "▁B",
+ "oy"
+ ],
+ [
+ "▁Bo",
+ "y"
+ ],
+ [
+ "▁D",
+ "ub"
+ ],
+ [
+ "▁Du",
+ "b"
+ ],
+ [
+ "▁k",
+ "w"
+ ],
+ [
+ "▁",
+ "kw"
+ ],
+ [
+ "▁n",
+ "och"
+ ],
+ [
+ "▁no",
+ "ch"
+ ],
+ [
+ "▁char",
+ "ges"
+ ],
+ [
+ "▁charg",
+ "es"
+ ],
+ [
+ "▁charge",
+ "s"
+ ],
+ [
+ "ar",
+ "ios"
+ ],
+ [
+ "ari",
+ "os"
+ ],
+ [
+ "ario",
+ "s"
+ ],
+ [
+ "a",
+ "rios"
+ ],
+ [
+ "ua",
+ "r"
+ ],
+ [
+ "u",
+ "ar"
+ ],
+ [
+ ";",
+ "&"
+ ],
+ [
+ "▁hab",
+ "ía"
+ ],
+ [
+ "(",
+ "`"
+ ],
+ [
+ "▁t",
+ "x"
+ ],
+ [
+ "▁",
+ "tx"
+ ],
+ [
+ "el",
+ "ve"
+ ],
+ [
+ "▁a",
+ "ños"
+ ],
+ [
+ "▁año",
+ "s"
+ ],
+ [
+ "▁m",
+ "ath"
+ ],
+ [
+ "▁mat",
+ "h"
+ ],
+ [
+ "▁ma",
+ "th"
+ ],
+ [
+ "▁",
+ "math"
+ ],
+ [
+ "▁Al",
+ "f"
+ ],
+ [
+ "▁F",
+ "und"
+ ],
+ [
+ "▁Fun",
+ "d"
+ ],
+ [
+ "▁Fu",
+ "nd"
+ ],
+ [
+ "▁man",
+ "ifest"
+ ],
+ [
+ "▁manif",
+ "est"
+ ],
+ [
+ "▁att",
+ "ached"
+ ],
+ [
+ "▁attach",
+ "ed"
+ ],
+ [
+ "▁spirit",
+ "ual"
+ ],
+ [
+ "▁Alex",
+ "ander"
+ ],
+ [
+ "▁Alexand",
+ "er"
+ ],
+ [
+ "un",
+ "es"
+ ],
+ [
+ "une",
+ "s"
+ ],
+ [
+ "u",
+ "nes"
+ ],
+ [
+ "▁s",
+ "eed"
+ ],
+ [
+ "▁se",
+ "ed"
+ ],
+ [
+ "▁see",
+ "d"
+ ],
+ [
+ "▁",
+ "seed"
+ ],
+ [
+ "▁Н",
+ "о"
+ ],
+ [
+ "▁mag",
+ "azine"
+ ],
+ [
+ "▁magaz",
+ "ine"
+ ],
+ [
+ "▁e",
+ "igen"
+ ],
+ [
+ "▁о",
+ "бра"
+ ],
+ [
+ "▁об",
+ "ра"
+ ],
+ [
+ "▁",
+ "обра"
+ ],
+ [
+ "e",
+ "a"
+ ],
+ [
+ "▁P",
+ "H"
+ ],
+ [
+ "▁",
+ "PH"
+ ],
+ [
+ "sw",
+ "ing"
+ ],
+ [
+ "s",
+ "wing"
+ ],
+ [
+ "▁As",
+ "ia"
+ ],
+ [
+ "ј",
+ "у"
+ ],
+ [
+ "▁K",
+ "IND"
+ ],
+ [
+ "Ident",
+ "ifier"
+ ],
+ [
+ "on",
+ "ce"
+ ],
+ [
+ "▁al",
+ "cohol"
+ ],
+ [
+ "ці",
+ "ї"
+ ],
+ [
+ "st",
+ "yles"
+ ],
+ [
+ "style",
+ "s"
+ ],
+ [
+ "sty",
+ "les"
+ ],
+ [
+ "assert",
+ "Equal"
+ ],
+ [
+ "▁R",
+ "a"
+ ],
+ [
+ "гра",
+ "фи"
+ ],
+ [
+ "▁mill",
+ "ions"
+ ],
+ [
+ "▁million",
+ "s"
+ ],
+ [
+ "▁ch",
+ "unk"
+ ],
+ [
+ "▁",
+ "chunk"
+ ],
+ [
+ "де",
+ "р"
+ ],
+ [
+ "д",
+ "ер"
+ ],
+ [
+ "Pack",
+ "age"
+ ],
+ [
+ "US",
+ "T"
+ ],
+ [
+ "U",
+ "ST"
+ ],
+ [
+ "▁N",
+ "othing"
+ ],
+ [
+ "▁No",
+ "thing"
+ ],
+ [
+ "▁Not",
+ "hing"
+ ],
+ [
+ "▁",
+ "Nothing"
+ ],
+ [
+ "(\"",
+ "#"
+ ],
+ [
+ "▁M",
+ "id"
+ ],
+ [
+ "▁Mi",
+ "d"
+ ],
+ [
+ "▁на",
+ "ча"
+ ],
+ [
+ "▁",
+ "нача"
+ ],
+ [
+ "ł",
+ "y"
+ ],
+ [
+ "AA",
+ "AA"
+ ],
+ [
+ "▁la",
+ "unched"
+ ],
+ [
+ "▁launch",
+ "ed"
+ ],
+ [
+ "▁w",
+ "ake"
+ ],
+ [
+ "▁wa",
+ "ke"
+ ],
+ [
+ "▁",
+ "wake"
+ ],
+ [
+ "▁gu",
+ "ests"
+ ],
+ [
+ "▁guest",
+ "s"
+ ],
+ [
+ "▁dif",
+ "ferences"
+ ],
+ [
+ "▁differ",
+ "ences"
+ ],
+ [
+ "▁difference",
+ "s"
+ ],
+ [
+ "ud",
+ "i"
+ ],
+ [
+ "u",
+ "di"
+ ],
+ [
+ "▁a",
+ "id"
+ ],
+ [
+ "▁ai",
+ "d"
+ ],
+ [
+ "▁",
+ "aid"
+ ],
+ [
+ "▁S",
+ "port"
+ ],
+ [
+ "▁Sp",
+ "ort"
+ ],
+ [
+ "ul",
+ "ator"
+ ],
+ [
+ "ula",
+ "tor"
+ ],
+ [
+ "ex",
+ "ecute"
+ ],
+ [
+ "exec",
+ "ute"
+ ],
+ [
+ "execut",
+ "e"
+ ],
+ [
+ "pl",
+ "ot"
+ ],
+ [
+ "plo",
+ "t"
+ ],
+ [
+ "p",
+ "lot"
+ ],
+ [
+ "ch",
+ "ing"
+ ],
+ [
+ "chi",
+ "ng"
+ ],
+ [
+ "c",
+ "hing"
+ ],
+ [
+ "▁N",
+ "orm"
+ ],
+ [
+ "▁No",
+ "rm"
+ ],
+ [
+ "▁Nor",
+ "m"
+ ],
+ [
+ "▁",
+ "Norm"
+ ],
+ [
+ "t",
+ "m"
+ ],
+ [
+ "\\",
+ "+"
+ ],
+ [
+ "AR",
+ "D"
+ ],
+ [
+ "A",
+ "RD"
+ ],
+ [
+ "▁be",
+ "er"
+ ],
+ [
+ "▁п",
+ "ід"
+ ],
+ [
+ "▁пі",
+ "д"
+ ],
+ [
+ "IA",
+ "L"
+ ],
+ [
+ "I",
+ "AL"
+ ],
+ [
+ "st",
+ "orage"
+ ],
+ [
+ "sto",
+ "rage"
+ ],
+ [
+ "▁An",
+ "na"
+ ],
+ [
+ "▁Ann",
+ "a"
+ ],
+ [
+ "▁y",
+ "ards"
+ ],
+ [
+ "▁yard",
+ "s"
+ ],
+ [
+ "▁techn",
+ "ique"
+ ],
+ [
+ "▁o",
+ "ù"
+ ],
+ [
+ "at",
+ "ten"
+ ],
+ [
+ "att",
+ "en"
+ ],
+ [
+ "atte",
+ "n"
+ ],
+ [
+ "UN",
+ "T"
+ ],
+ [
+ "U",
+ "NT"
+ ],
+ [
+ "do",
+ "n"
+ ],
+ [
+ "d",
+ "on"
+ ],
+ [
+ "фо",
+ "р"
+ ],
+ [
+ "ф",
+ "ор"
+ ],
+ [
+ "▁h",
+ "oping"
+ ],
+ [
+ "▁hop",
+ "ing"
+ ],
+ [
+ "▁ho",
+ "ping"
+ ],
+ [
+ "▁vict",
+ "ory"
+ ],
+ [
+ "it",
+ "at"
+ ],
+ [
+ "ita",
+ "t"
+ ],
+ [
+ "i",
+ "tat"
+ ],
+ [
+ "▁signific",
+ "antly"
+ ],
+ [
+ "▁significant",
+ "ly"
+ ],
+ [
+ "▁pract",
+ "ical"
+ ],
+ [
+ "ij",
+ "e"
+ ],
+ [
+ "i",
+ "je"
+ ],
+ [
+ "▁exp",
+ "ansion"
+ ],
+ [
+ "▁expans",
+ "ion"
+ ],
+ [
+ "J",
+ "S"
+ ],
+ [
+ "ix",
+ "els"
+ ],
+ [
+ "ixel",
+ "s"
+ ],
+ [
+ "US",
+ "ER"
+ ],
+ [
+ "USE",
+ "R"
+ ],
+ [
+ "U",
+ "SER"
+ ],
+ [
+ "Sh",
+ "ape"
+ ],
+ [
+ "▁ext",
+ "ent"
+ ],
+ [
+ "li",
+ "o"
+ ],
+ [
+ "l",
+ "io"
+ ],
+ [
+ "▁p",
+ "ued"
+ ],
+ [
+ "▁pu",
+ "ed"
+ ],
+ [
+ "ol",
+ "id"
+ ],
+ [
+ "oli",
+ "d"
+ ],
+ [
+ "▁g",
+ "am"
+ ],
+ [
+ "▁ga",
+ "m"
+ ],
+ [
+ "▁s",
+ "event"
+ ],
+ [
+ "▁se",
+ "vent"
+ ],
+ [
+ "▁seven",
+ "t"
+ ],
+ [
+ "▁G",
+ "a"
+ ],
+ [
+ "▁",
+ "Ga"
+ ],
+ [
+ "angu",
+ "ages"
+ ],
+ [
+ "anguage",
+ "s"
+ ],
+ [
+ "((",
+ "("
+ ],
+ [
+ "(",
+ "(("
+ ],
+ [
+ "ъ",
+ "л"
+ ],
+ [
+ "▁Ex",
+ "per"
+ ],
+ [
+ "▁Exp",
+ "er"
+ ],
+ [
+ "▁",
+ "Exper"
+ ],
+ [
+ "as",
+ "ty"
+ ],
+ [
+ "ast",
+ "y"
+ ],
+ [
+ "a",
+ "sty"
+ ],
+ [
+ "ri",
+ "eg"
+ ],
+ [
+ "rie",
+ "g"
+ ],
+ [
+ "r",
+ "ieg"
+ ],
+ [
+ "gi",
+ "o"
+ ],
+ [
+ "g",
+ "io"
+ ],
+ [
+ "od",
+ "o"
+ ],
+ [
+ "o",
+ "do"
+ ],
+ [
+ "▁col",
+ "le"
+ ],
+ [
+ "▁co",
+ "lle"
+ ],
+ [
+ "▁coll",
+ "e"
+ ],
+ [
+ "▁st",
+ "ored"
+ ],
+ [
+ "▁store",
+ "d"
+ ],
+ [
+ "▁stor",
+ "ed"
+ ],
+ [
+ "▁sto",
+ "red"
+ ],
+ [
+ "▁S",
+ "che"
+ ],
+ [
+ "▁Sch",
+ "e"
+ ],
+ [
+ "▁Sc",
+ "he"
+ ],
+ [
+ "▁",
+ "Sche"
+ ],
+ [
+ "ist",
+ "ant"
+ ],
+ [
+ "ista",
+ "nt"
+ ],
+ [
+ "istan",
+ "t"
+ ],
+ [
+ "i",
+ "stant"
+ ],
+ [
+ "▁l",
+ "ip"
+ ],
+ [
+ "▁li",
+ "p"
+ ],
+ [
+ "B",
+ "R"
+ ],
+ [
+ "▁a",
+ "ug"
+ ],
+ [
+ "▁au",
+ "g"
+ ],
+ [
+ "▁",
+ "aug"
+ ],
+ [
+ "▁S",
+ "earch"
+ ],
+ [
+ "▁Se",
+ "arch"
+ ],
+ [
+ "▁",
+ "Search"
+ ],
+ [
+ ")=",
+ "\\"
+ ],
+ [
+ ")",
+ "=\\"
+ ],
+ [
+ "▁U",
+ "r"
+ ],
+ [
+ "▁s",
+ "ole"
+ ],
+ [
+ "▁so",
+ "le"
+ ],
+ [
+ "▁sol",
+ "e"
+ ],
+ [
+ "▁",
+ "sole"
+ ],
+ [
+ "il",
+ "lo"
+ ],
+ [
+ "ill",
+ "o"
+ ],
+ [
+ "▁me",
+ "hr"
+ ],
+ [
+ "ki",
+ "t"
+ ],
+ [
+ "k",
+ "it"
+ ],
+ [
+ "▁in",
+ "terior"
+ ],
+ [
+ "▁inter",
+ "ior"
+ ],
+ [
+ "▁inte",
+ "rior"
+ ],
+ [
+ "LI",
+ "ST"
+ ],
+ [
+ "L",
+ "IST"
+ ],
+ [
+ "ad",
+ "el"
+ ],
+ [
+ "ade",
+ "l"
+ ],
+ [
+ "a",
+ "del"
+ ],
+ [
+ "▁shop",
+ "ping"
+ ],
+ [
+ "▁s",
+ "lä"
+ ],
+ [
+ "▁sl",
+ "ä"
+ ],
+ [
+ "You",
+ "r"
+ ],
+ [
+ "Y",
+ "our"
+ ],
+ [
+ "DI",
+ "TION"
+ ],
+ [
+ "D",
+ "ITION"
+ ],
+ [
+ "▁H",
+ "ttp"
+ ],
+ [
+ "▁",
+ "Http"
+ ],
+ [
+ "ra",
+ "ham"
+ ],
+ [
+ "rah",
+ "am"
+ ],
+ [
+ "т",
+ "ри"
+ ],
+ [
+ "▁b",
+ "rings"
+ ],
+ [
+ "▁br",
+ "ings"
+ ],
+ [
+ "▁bring",
+ "s"
+ ],
+ [
+ "Re",
+ "v"
+ ],
+ [
+ "R",
+ "ev"
+ ],
+ [
+ "▁pro",
+ "pag"
+ ],
+ [
+ "▁prop",
+ "ag"
+ ],
+ [
+ "ity",
+ "Engine"
+ ],
+ [
+ "()",
+ "),"
+ ],
+ [
+ "())",
+ ","
+ ],
+ [
+ "(",
+ ")),"
+ ],
+ [
+ "▁ing",
+ "år"
+ ],
+ [
+ "▁Ir",
+ "eland"
+ ],
+ [
+ "▁Ire",
+ "land"
+ ],
+ [
+ "▁\"",
+ "./"
+ ],
+ [
+ "▁\".",
+ "/"
+ ],
+ [
+ "▁H",
+ "arr"
+ ],
+ [
+ "▁Har",
+ "r"
+ ],
+ [
+ "▁Ha",
+ "rr"
+ ],
+ [
+ "▁ad",
+ "min"
+ ],
+ [
+ "▁adm",
+ "in"
+ ],
+ [
+ "▁",
+ "admin"
+ ],
+ [
+ "en",
+ "o"
+ ],
+ [
+ "e",
+ "no"
+ ],
+ [
+ "▁k",
+ "r"
+ ],
+ [
+ "▁",
+ "kr"
+ ],
+ [
+ "▁est",
+ "á"
+ ],
+ [
+ "▁pro",
+ "ps"
+ ],
+ [
+ "▁pr",
+ "ops"
+ ],
+ [
+ "▁prop",
+ "s"
+ ],
+ [
+ "▁",
+ "props"
+ ],
+ [
+ "to",
+ "k"
+ ],
+ [
+ "t",
+ "ok"
+ ],
+ [
+ "om",
+ "orph"
+ ],
+ [
+ "▁affect",
+ "ed"
+ ],
+ [
+ "Ph",
+ "one"
+ ],
+ [
+ "▁deg",
+ "rees"
+ ],
+ [
+ "▁degree",
+ "s"
+ ],
+ [
+ "so",
+ "me"
+ ],
+ [
+ "som",
+ "e"
+ ],
+ [
+ "s",
+ "ome"
+ ],
+ [
+ "▁n",
+ "in"
+ ],
+ [
+ "▁ni",
+ "n"
+ ],
+ [
+ "EV",
+ "ENT"
+ ],
+ [
+ "▁inter",
+ "action"
+ ],
+ [
+ "▁inte",
+ "raction"
+ ],
+ [
+ "▁interact",
+ "ion"
+ ],
+ [
+ "▁T",
+ "uesday"
+ ],
+ [
+ "iter",
+ "ator"
+ ],
+ [
+ "▁N",
+ "ob"
+ ],
+ [
+ "▁No",
+ "b"
+ ],
+ [
+ "▁sc",
+ "atter"
+ ],
+ [
+ "uck",
+ "et"
+ ],
+ [
+ "uc",
+ "ket"
+ ],
+ [
+ "com",
+ "plete"
+ ],
+ [
+ "comp",
+ "lete"
+ ],
+ [
+ "▁d",
+ "uty"
+ ],
+ [
+ "▁du",
+ "ty"
+ ],
+ [
+ "▁dut",
+ "y"
+ ],
+ [
+ "▁answ",
+ "ers"
+ ],
+ [
+ "▁answer",
+ "s"
+ ],
+ [
+ "Pro",
+ "gress"
+ ],
+ [
+ "ee",
+ "d"
+ ],
+ [
+ "e",
+ "ed"
+ ],
+ [
+ "ро",
+ "н"
+ ],
+ [
+ "р",
+ "он"
+ ],
+ [
+ "▁v",
+ "ie"
+ ],
+ [
+ "▁vi",
+ "e"
+ ],
+ [
+ "▁de",
+ "pos"
+ ],
+ [
+ "▁dep",
+ "os"
+ ],
+ [
+ "▁p",
+ "acket"
+ ],
+ [
+ "▁pack",
+ "et"
+ ],
+ [
+ "▁pac",
+ "ket"
+ ],
+ [
+ "▁",
+ "packet"
+ ],
+ [
+ "▁t",
+ "ow"
+ ],
+ [
+ "▁to",
+ "w"
+ ],
+ [
+ "▁de",
+ "leg"
+ ],
+ [
+ "▁del",
+ "eg"
+ ],
+ [
+ "▁",
+ "deleg"
+ ],
+ [
+ "aud",
+ "io"
+ ],
+ [
+ "a",
+ "udio"
+ ],
+ [
+ "▁v",
+ "ary"
+ ],
+ [
+ "▁var",
+ "y"
+ ],
+ [
+ "▁va",
+ "ry"
+ ],
+ [
+ "▁m",
+ "igr"
+ ],
+ [
+ "▁mi",
+ "gr"
+ ],
+ [
+ "▁mig",
+ "r"
+ ],
+ [
+ "▁",
+ "migr"
+ ],
+ [
+ "ф",
+ "і"
+ ],
+ [
+ "es",
+ "a"
+ ],
+ [
+ "e",
+ "sa"
+ ],
+ [
+ "Event",
+ "s"
+ ],
+ [
+ "Ev",
+ "ents"
+ ],
+ [
+ "Even",
+ "ts"
+ ],
+ [
+ "ha",
+ "us"
+ ],
+ [
+ "h",
+ "aus"
+ ],
+ [
+ "▁S",
+ "av"
+ ],
+ [
+ "▁Sa",
+ "v"
+ ],
+ [
+ "▁Port",
+ "ug"
+ ],
+ [
+ "▁с",
+ "то"
+ ],
+ [
+ "▁ст",
+ "о"
+ ],
+ [
+ "▁",
+ "сто"
+ ],
+ [
+ "il",
+ "ation"
+ ],
+ [
+ "i",
+ "lation"
+ ],
+ [
+ "▁met",
+ "adata"
+ ],
+ [
+ "▁meta",
+ "data"
+ ],
+ [
+ "▁",
+ "metadata"
+ ],
+ [
+ "la",
+ "s"
+ ],
+ [
+ "l",
+ "as"
+ ],
+ [
+ "▁a",
+ "i"
+ ],
+ [
+ "▁",
+ "ai"
+ ],
+ [
+ "▁an",
+ "ger"
+ ],
+ [
+ "▁ang",
+ "er"
+ ],
+ [
+ "▁ange",
+ "r"
+ ],
+ [
+ "▁",
+ "anger"
+ ],
+ [
+ "▁h",
+ "am"
+ ],
+ [
+ "▁ha",
+ "m"
+ ],
+ [
+ "▁",
+ "ham"
+ ],
+ [
+ "▁A",
+ "nal"
+ ],
+ [
+ "▁An",
+ "al"
+ ],
+ [
+ "▁Ana",
+ "l"
+ ],
+ [
+ "▁",
+ "Anal"
+ ],
+ [
+ "▁frequ",
+ "ently"
+ ],
+ [
+ "▁frequent",
+ "ly"
+ ],
+ [
+ "▁F",
+ "ALSE"
+ ],
+ [
+ "▁",
+ "FALSE"
+ ],
+ [
+ "oc",
+ "he"
+ ],
+ [
+ "och",
+ "e"
+ ],
+ [
+ "o",
+ "che"
+ ],
+ [
+ "re",
+ "z"
+ ],
+ [
+ "r",
+ "ez"
+ ],
+ [
+ "▁V",
+ "iet"
+ ],
+ [
+ "▁Vi",
+ "et"
+ ],
+ [
+ "qu",
+ "is"
+ ],
+ [
+ "q",
+ "uis"
+ ],
+ [
+ "▁char",
+ "ged"
+ ],
+ [
+ "▁charg",
+ "ed"
+ ],
+ [
+ "▁charge",
+ "d"
+ ],
+ [
+ "ä",
+ "s"
+ ],
+ [
+ "▁P",
+ "ath"
+ ],
+ [
+ "▁Pat",
+ "h"
+ ],
+ [
+ "▁Pa",
+ "th"
+ ],
+ [
+ "▁",
+ "Path"
+ ],
+ [
+ "▁accur",
+ "ate"
+ ],
+ [
+ "▁Pl",
+ "us"
+ ],
+ [
+ "▁",
+ "Plus"
+ ],
+ [
+ "ke",
+ "it"
+ ],
+ [
+ "▁In",
+ "put"
+ ],
+ [
+ "▁",
+ "Input"
+ ],
+ [
+ "wh",
+ "en"
+ ],
+ [
+ "whe",
+ "n"
+ ],
+ [
+ "w",
+ "hen"
+ ],
+ [
+ "er",
+ "as"
+ ],
+ [
+ "era",
+ "s"
+ ],
+ [
+ "e",
+ "ras"
+ ],
+ [
+ "▁во",
+ "з"
+ ],
+ [
+ "▁de",
+ "rived"
+ ],
+ [
+ "▁der",
+ "ived"
+ ],
+ [
+ "▁deriv",
+ "ed"
+ ],
+ [
+ "▁derive",
+ "d"
+ ],
+ [
+ "aj",
+ "e"
+ ],
+ [
+ "a",
+ "je"
+ ],
+ [
+ "▁H",
+ "ad"
+ ],
+ [
+ "▁Ha",
+ "d"
+ ],
+ [
+ "ur",
+ "en"
+ ],
+ [
+ "ure",
+ "n"
+ ],
+ [
+ "u",
+ "ren"
+ ],
+ [
+ "ó",
+ "r"
+ ],
+ [
+ "}=",
+ "\\"
+ ],
+ [
+ "}",
+ "=\\"
+ ],
+ [
+ "ur",
+ "eau"
+ ],
+ [
+ "ure",
+ "au"
+ ],
+ [
+ "al",
+ "and"
+ ],
+ [
+ "ala",
+ "nd"
+ ],
+ [
+ "a",
+ "land"
+ ],
+ [
+ "Execut",
+ "ion"
+ ],
+ [
+ "Exec",
+ "ution"
+ ],
+ [
+ "ed",
+ "en"
+ ],
+ [
+ "ede",
+ "n"
+ ],
+ [
+ "e",
+ "den"
+ ],
+ [
+ "▁se",
+ "eking"
+ ],
+ [
+ "▁see",
+ "king"
+ ],
+ [
+ "▁seek",
+ "ing"
+ ],
+ [
+ "ch",
+ "anged"
+ ],
+ [
+ "change",
+ "d"
+ ],
+ [
+ "chan",
+ "ged"
+ ],
+ [
+ "▁t",
+ "rem"
+ ],
+ [
+ "▁tr",
+ "em"
+ ],
+ [
+ "▁tre",
+ "m"
+ ],
+ [
+ "ск",
+ "у"
+ ],
+ [
+ "с",
+ "ку"
+ ],
+ [
+ "▁G",
+ "eme"
+ ],
+ [
+ "▁Ge",
+ "me"
+ ],
+ [
+ "▁Gem",
+ "e"
+ ],
+ [
+ "in",
+ "ating"
+ ],
+ [
+ "ina",
+ "ting"
+ ],
+ [
+ "▁column",
+ "s"
+ ],
+ [
+ "▁",
+ "columns"
+ ],
+ [
+ "E",
+ "P"
+ ],
+ [
+ "▁inj",
+ "ury"
+ ],
+ [
+ "end",
+ "ent"
+ ],
+ [
+ "ende",
+ "nt"
+ ],
+ [
+ "enden",
+ "t"
+ ],
+ [
+ "▁he",
+ "aded"
+ ],
+ [
+ "▁head",
+ "ed"
+ ],
+ [
+ "▁",
+ "headed"
+ ],
+ [
+ "AS",
+ "E"
+ ],
+ [
+ "A",
+ "SE"
+ ],
+ [
+ "▁Mus",
+ "lim"
+ ],
+ [
+ "▁cl",
+ "imate"
+ ],
+ [
+ "▁clim",
+ "ate"
+ ],
+ [
+ "▁f",
+ "ake"
+ ],
+ [
+ "▁fa",
+ "ke"
+ ],
+ [
+ "▁",
+ "fake"
+ ],
+ [
+ "CM",
+ "D"
+ ],
+ [
+ "C",
+ "MD"
+ ],
+ [
+ "ј",
+ "и"
+ ],
+ [
+ "▁Ar",
+ "ts"
+ ],
+ [
+ "▁Art",
+ "s"
+ ],
+ [
+ "fe",
+ "ction"
+ ],
+ [
+ "fect",
+ "ion"
+ ],
+ [
+ "f",
+ "ection"
+ ],
+ [
+ "▁p",
+ "it"
+ ],
+ [
+ "▁pi",
+ "t"
+ ],
+ [
+ "▁",
+ "pit"
+ ],
+ [
+ ">",
+ "\\"
+ ],
+ [
+ "an",
+ "al"
+ ],
+ [
+ "ana",
+ "l"
+ ],
+ [
+ "a",
+ "nal"
+ ],
+ [
+ "Se",
+ "ction"
+ ],
+ [
+ "S",
+ "ection"
+ ],
+ [
+ "pl",
+ "us"
+ ],
+ [
+ "ü",
+ "t"
+ ],
+ [
+ "▁em",
+ "bed"
+ ],
+ [
+ "▁emb",
+ "ed"
+ ],
+ [
+ "▁",
+ "embed"
+ ],
+ [
+ "▁st",
+ "rings"
+ ],
+ [
+ "▁str",
+ "ings"
+ ],
+ [
+ "▁string",
+ "s"
+ ],
+ [
+ "▁",
+ "strings"
+ ],
+ [
+ "Be",
+ "fore"
+ ],
+ [
+ "B",
+ "efore"
+ ],
+ [
+ "pro",
+ "c"
+ ],
+ [
+ "pr",
+ "oc"
+ ],
+ [
+ "p",
+ "roc"
+ ],
+ [
+ "▁с",
+ "по"
+ ],
+ [
+ "▁сп",
+ "о"
+ ],
+ [
+ "▁",
+ "спо"
+ ],
+ [
+ "tr",
+ "l"
+ ],
+ [
+ "t",
+ "rl"
+ ],
+ [
+ "v",
+ "r"
+ ],
+ [
+ "Back",
+ "ground"
+ ],
+ [
+ "log",
+ "ger"
+ ],
+ [
+ "ag",
+ "raph"
+ ],
+ [
+ "agr",
+ "aph"
+ ],
+ [
+ "agra",
+ "ph"
+ ],
+ [
+ "a",
+ "graph"
+ ],
+ [
+ "ie",
+ "st"
+ ],
+ [
+ "ies",
+ "t"
+ ],
+ [
+ "i",
+ "est"
+ ],
+ [
+ "▁good",
+ "s"
+ ],
+ [
+ "bat",
+ "ch"
+ ],
+ [
+ "b",
+ "atch"
+ ],
+ [
+ "▁opt",
+ "ional"
+ ],
+ [
+ "▁option",
+ "al"
+ ],
+ [
+ "▁",
+ "optional"
+ ],
+ [
+ "▁Tay",
+ "lor"
+ ],
+ [
+ "▁recogn",
+ "ize"
+ ],
+ [
+ "wal",
+ "k"
+ ],
+ [
+ "w",
+ "alk"
+ ],
+ [
+ "▁H",
+ "it"
+ ],
+ [
+ "▁Hi",
+ "t"
+ ],
+ [
+ "▁",
+ "Hit"
+ ],
+ [
+ "▁Eliz",
+ "abeth"
+ ],
+ [
+ "}",
+ ":"
+ ],
+ [
+ "▁care",
+ "ful"
+ ],
+ [
+ "кра",
+ "ї"
+ ],
+ [
+ "▁loc",
+ "ations"
+ ],
+ [
+ "▁location",
+ "s"
+ ],
+ [
+ "▁struct",
+ "ures"
+ ],
+ [
+ "▁structure",
+ "s"
+ ],
+ [
+ "▁d",
+ "isk"
+ ],
+ [
+ "▁dis",
+ "k"
+ ],
+ [
+ "▁di",
+ "sk"
+ ],
+ [
+ "▁",
+ "disk"
+ ],
+ [
+ "▁sh",
+ "ips"
+ ],
+ [
+ "▁ship",
+ "s"
+ ],
+ [
+ "▁",
+ "ships"
+ ],
+ [
+ "▁s",
+ "uo"
+ ],
+ [
+ "▁su",
+ "o"
+ ],
+ [
+ "▁s",
+ "owie"
+ ],
+ [
+ "▁so",
+ "wie"
+ ],
+ [
+ "▁sow",
+ "ie"
+ ],
+ [
+ "▁E",
+ "ss"
+ ],
+ [
+ "▁Es",
+ "s"
+ ],
+ [
+ "▁H",
+ "ash"
+ ],
+ [
+ "▁Ha",
+ "sh"
+ ],
+ [
+ "▁Has",
+ "h"
+ ],
+ [
+ "▁",
+ "Hash"
+ ],
+ [
+ "▁reason",
+ "able"
+ ],
+ [
+ "▁More",
+ "over"
+ ],
+ [
+ "▁form",
+ "ula"
+ ],
+ [
+ "▁C",
+ "entre"
+ ],
+ [
+ "▁Cent",
+ "re"
+ ],
+ [
+ "▁res",
+ "idents"
+ ],
+ [
+ "▁resident",
+ "s"
+ ],
+ [
+ "▁resid",
+ "ents"
+ ],
+ [
+ "R",
+ "S"
+ ],
+ [
+ "Id",
+ "s"
+ ],
+ [
+ "I",
+ "ds"
+ ],
+ [
+ "▁K",
+ "now"
+ ],
+ [
+ "▁Kn",
+ "ow"
+ ],
+ [
+ "▁t",
+ "rib"
+ ],
+ [
+ "▁tr",
+ "ib"
+ ],
+ [
+ "▁tri",
+ "b"
+ ],
+ [
+ "▁r",
+ "és"
+ ],
+ [
+ "▁ré",
+ "s"
+ ],
+ [
+ "▁s",
+ "table"
+ ],
+ [
+ "▁st",
+ "able"
+ ],
+ [
+ "▁sta",
+ "ble"
+ ],
+ [
+ "▁stab",
+ "le"
+ ],
+ [
+ "▁",
+ "stable"
+ ],
+ [
+ "▁W",
+ "ould"
+ ],
+ [
+ "▁Wo",
+ "uld"
+ ],
+ [
+ "▁",
+ "Would"
+ ],
+ [
+ "▁break",
+ "ing"
+ ],
+ [
+ "▁bre",
+ "aking"
+ ],
+ [
+ "▁",
+ "breaking"
+ ],
+ [
+ "▁me",
+ "al"
+ ],
+ [
+ "▁p",
+ "hen"
+ ],
+ [
+ "▁ph",
+ "en"
+ ],
+ [
+ "▁f",
+ "el"
+ ],
+ [
+ "▁fe",
+ "l"
+ ],
+ [
+ "▁",
+ "fel"
+ ],
+ [
+ "▁F",
+ "red"
+ ],
+ [
+ "▁Fr",
+ "ed"
+ ],
+ [
+ "▁Fre",
+ "d"
+ ],
+ [
+ "Aut",
+ "hor"
+ ],
+ [
+ "Auth",
+ "or"
+ ],
+ [
+ "▁c",
+ "apture"
+ ],
+ [
+ "▁capt",
+ "ure"
+ ],
+ [
+ "▁",
+ "capture"
+ ],
+ [
+ "op",
+ "ts"
+ ],
+ [
+ "opt",
+ "s"
+ ],
+ [
+ "o",
+ "pts"
+ ],
+ [
+ "▁every",
+ "where"
+ ],
+ [
+ "▁s",
+ "que"
+ ],
+ [
+ "▁squ",
+ "e"
+ ],
+ [
+ "▁sq",
+ "ue"
+ ],
+ [
+ "▁m",
+ "oder"
+ ],
+ [
+ "▁mod",
+ "er"
+ ],
+ [
+ "▁mo",
+ "der"
+ ],
+ [
+ "▁mode",
+ "r"
+ ],
+ [
+ "set",
+ "up"
+ ],
+ [
+ "▁S",
+ "upp"
+ ],
+ [
+ "▁Su",
+ "pp"
+ ],
+ [
+ "▁Sup",
+ "p"
+ ],
+ [
+ "▁",
+ "Supp"
+ ],
+ [
+ "▁when",
+ "ever"
+ ],
+ [
+ "▁whe",
+ "never"
+ ],
+ [
+ "{",
+ "("
+ ],
+ [
+ "wa",
+ "rt"
+ ],
+ [
+ "war",
+ "t"
+ ],
+ [
+ "w",
+ "art"
+ ],
+ [
+ "▁t",
+ "oe"
+ ],
+ [
+ "▁to",
+ "e"
+ ],
+ [
+ "Pre",
+ "fix"
+ ],
+ [
+ "Pref",
+ "ix"
+ ],
+ [
+ "P",
+ "refix"
+ ],
+ [
+ "ho",
+ "u"
+ ],
+ [
+ "h",
+ "ou"
+ ],
+ [
+ "ga",
+ "ge"
+ ],
+ [
+ "g",
+ "age"
+ ],
+ [
+ ">",
+ "\""
+ ],
+ [
+ "▁f",
+ "rag"
+ ],
+ [
+ "▁fr",
+ "ag"
+ ],
+ [
+ "▁fra",
+ "g"
+ ],
+ [
+ "▁",
+ "frag"
+ ],
+ [
+ "▁The",
+ "orem"
+ ],
+ [
+ "mem",
+ "ory"
+ ],
+ [
+ "▁cont",
+ "ents"
+ ],
+ [
+ "▁content",
+ "s"
+ ],
+ [
+ "▁conten",
+ "ts"
+ ],
+ [
+ "▁",
+ "contents"
+ ],
+ [
+ "do",
+ "cs"
+ ],
+ [
+ "doc",
+ "s"
+ ],
+ [
+ "}",
+ "'"
+ ],
+ [
+ "▁Ir",
+ "ish"
+ ],
+ [
+ "The",
+ "n"
+ ],
+ [
+ "Th",
+ "en"
+ ],
+ [
+ "T",
+ "hen"
+ ],
+ [
+ "aa",
+ "ts"
+ ],
+ [
+ "aat",
+ "s"
+ ],
+ [
+ "a",
+ "ats"
+ ],
+ [
+ "Sa",
+ "ve"
+ ],
+ [
+ "S",
+ "ave"
+ ],
+ [
+ "▁a",
+ "gency"
+ ],
+ [
+ "▁ag",
+ "ency"
+ ],
+ [
+ "▁и",
+ "ме"
+ ],
+ [
+ "▁им",
+ "е"
+ ],
+ [
+ "до",
+ "ва"
+ ],
+ [
+ "дов",
+ "а"
+ ],
+ [
+ "▁F",
+ "unction"
+ ],
+ [
+ "▁Fun",
+ "ction"
+ ],
+ [
+ "▁",
+ "Function"
+ ],
+ [
+ "N",
+ "N"
+ ],
+ [
+ "dest",
+ "roy"
+ ],
+ [
+ "▁M",
+ "essage"
+ ],
+ [
+ "▁Mess",
+ "age"
+ ],
+ [
+ "▁",
+ "Message"
+ ],
+ [
+ "▁c",
+ "ancel"
+ ],
+ [
+ "▁can",
+ "cel"
+ ],
+ [
+ "▁",
+ "cancel"
+ ],
+ [
+ "▁super",
+ "ior"
+ ],
+ [
+ "▁e",
+ "c"
+ ],
+ [
+ "▁",
+ "ec"
+ ],
+ [
+ "▁liter",
+ "ature"
+ ],
+ [
+ "▁P",
+ "ART"
+ ],
+ [
+ "▁PA",
+ "RT"
+ ],
+ [
+ "▁PAR",
+ "T"
+ ],
+ [
+ "▁",
+ "PART"
+ ],
+ [
+ "I",
+ "l"
+ ],
+ [
+ "▁C",
+ "ab"
+ ],
+ [
+ "▁Ca",
+ "b"
+ ],
+ [
+ "eng",
+ "ine"
+ ],
+ [
+ "▁b",
+ "asket"
+ ],
+ [
+ "▁bas",
+ "ket"
+ ],
+ [
+ "wor",
+ "th"
+ ],
+ [
+ "wort",
+ "h"
+ ],
+ [
+ "w",
+ "orth"
+ ],
+ [
+ "▁S",
+ "el"
+ ],
+ [
+ "▁Se",
+ "l"
+ ],
+ [
+ "f",
+ "etch"
+ ],
+ [
+ "▁St",
+ "adt"
+ ],
+ [
+ "▁Stad",
+ "t"
+ ],
+ [
+ "▁Sta",
+ "dt"
+ ],
+ [
+ "▁К",
+ "и"
+ ],
+ [
+ "▁con",
+ "j"
+ ],
+ [
+ "▁se",
+ "iner"
+ ],
+ [
+ "▁sein",
+ "er"
+ ],
+ [
+ "▁seine",
+ "r"
+ ],
+ [
+ "▁sei",
+ "ner"
+ ],
+ [
+ "▁conf",
+ "irmed"
+ ],
+ [
+ "▁confirm",
+ "ed"
+ ],
+ [
+ "▁Ar",
+ "gent"
+ ],
+ [
+ "▁Arg",
+ "ent"
+ ],
+ [
+ "am",
+ "ar"
+ ],
+ [
+ "ama",
+ "r"
+ ],
+ [
+ "a",
+ "mar"
+ ],
+ [
+ "pgf",
+ "path"
+ ],
+ [
+ "▁strugg",
+ "le"
+ ],
+ [
+ "Pat",
+ "tern"
+ ],
+ [
+ "▁M",
+ "iddle"
+ ],
+ [
+ "it",
+ "an"
+ ],
+ [
+ "ita",
+ "n"
+ ],
+ [
+ "i",
+ "tan"
+ ],
+ [
+ "▁m",
+ "oon"
+ ],
+ [
+ "▁mo",
+ "on"
+ ],
+ [
+ "or",
+ "ough"
+ ],
+ [
+ "oro",
+ "ugh"
+ ],
+ [
+ "o",
+ "rough"
+ ],
+ [
+ "▁Cath",
+ "olic"
+ ],
+ [
+ "▁str",
+ "uck"
+ ],
+ [
+ "▁stru",
+ "ck"
+ ],
+ [
+ "]",
+ "->"
+ ],
+ [
+ "▁we",
+ "apon"
+ ],
+ [
+ "▁weap",
+ "on"
+ ],
+ [
+ "▁su",
+ "bst"
+ ],
+ [
+ "▁sub",
+ "st"
+ ],
+ [
+ "▁subs",
+ "t"
+ ],
+ [
+ "▁inst",
+ "ructions"
+ ],
+ [
+ "▁instruct",
+ "ions"
+ ],
+ [
+ "▁instruction",
+ "s"
+ ],
+ [
+ "▁occ",
+ "as"
+ ],
+ [
+ "▁oc",
+ "cas"
+ ],
+ [
+ "prote",
+ "cted"
+ ],
+ [
+ "▁L",
+ "ess"
+ ],
+ [
+ "▁Le",
+ "ss"
+ ],
+ [
+ "▁Les",
+ "s"
+ ],
+ [
+ "▁",
+ "Less"
+ ],
+ [
+ "▁b",
+ "atch"
+ ],
+ [
+ "▁bat",
+ "ch"
+ ],
+ [
+ "▁",
+ "batch"
+ ],
+ [
+ "▁con",
+ "tra"
+ ],
+ [
+ "▁cont",
+ "ra"
+ ],
+ [
+ "▁contr",
+ "a"
+ ],
+ [
+ "▁de",
+ "ck"
+ ],
+ [
+ "▁dec",
+ "k"
+ ],
+ [
+ "▁",
+ "deck"
+ ],
+ [
+ "▁ign",
+ "ored"
+ ],
+ [
+ "▁ignore",
+ "d"
+ ],
+ [
+ "▁ignor",
+ "ed"
+ ],
+ [
+ "▁ref",
+ "used"
+ ],
+ [
+ "▁refuse",
+ "d"
+ ],
+ [
+ "tr",
+ "igger"
+ ],
+ [
+ "▁crim",
+ "inal"
+ ],
+ [
+ "G",
+ "A"
+ ],
+ [
+ "ol",
+ "ly"
+ ],
+ [
+ "oll",
+ "y"
+ ],
+ [
+ "▁B",
+ "ell"
+ ],
+ [
+ "▁Be",
+ "ll"
+ ],
+ [
+ "▁Bel",
+ "l"
+ ],
+ [
+ "▁",
+ "Ю"
+ ],
+ [
+ "for",
+ "ward"
+ ],
+ [
+ "▁p",
+ "refix"
+ ],
+ [
+ "▁pre",
+ "fix"
+ ],
+ [
+ "▁pref",
+ "ix"
+ ],
+ [
+ "▁",
+ "prefix"
+ ],
+ [
+ "▁im",
+ "mediate"
+ ],
+ [
+ "▁immedi",
+ "ate"
+ ],
+ [
+ "▁as",
+ "signed"
+ ],
+ [
+ "▁ass",
+ "igned"
+ ],
+ [
+ "▁assign",
+ "ed"
+ ],
+ [
+ "▁e",
+ "lected"
+ ],
+ [
+ "▁elect",
+ "ed"
+ ],
+ [
+ "▁ele",
+ "cted"
+ ],
+ [
+ "▁to",
+ "night"
+ ],
+ [
+ "▁ton",
+ "ight"
+ ],
+ [
+ "▁D",
+ "ies"
+ ],
+ [
+ "▁Die",
+ "s"
+ ],
+ [
+ "▁Di",
+ "es"
+ ],
+ [
+ "▁B",
+ "each"
+ ],
+ [
+ "▁Be",
+ "ach"
+ ],
+ [
+ "▁pre",
+ "ced"
+ ],
+ [
+ "▁prec",
+ "ed"
+ ],
+ [
+ "ow",
+ "ał"
+ ],
+ [
+ "owa",
+ "ł"
+ ],
+ [
+ "▁gal",
+ "ax"
+ ],
+ [
+ "▁log",
+ "ic"
+ ],
+ [
+ "en",
+ "za"
+ ],
+ [
+ "enz",
+ "a"
+ ],
+ [
+ "▁Cap",
+ "tain"
+ ],
+ [
+ "▁Capt",
+ "ain"
+ ],
+ [
+ "▁H",
+ "ay"
+ ],
+ [
+ "▁Ha",
+ "y"
+ ],
+ [
+ "▁f",
+ "acts"
+ ],
+ [
+ "▁fact",
+ "s"
+ ],
+ [
+ "▁fac",
+ "ts"
+ ],
+ [
+ "▁н",
+ "и"
+ ],
+ [
+ "▁",
+ "ни"
+ ],
+ [
+ "t",
+ "é"
+ ],
+ [
+ "▁s",
+ "b"
+ ],
+ [
+ "▁",
+ "sb"
+ ],
+ [
+ "op",
+ "ed"
+ ],
+ [
+ "ope",
+ "d"
+ ],
+ [
+ "o",
+ "ped"
+ ],
+ [
+ "▁com",
+ "bat"
+ ],
+ [
+ "▁comb",
+ "at"
+ ],
+ [
+ "▁expl",
+ "ore"
+ ],
+ [
+ "▁explo",
+ "re"
+ ],
+ [
+ "▁(",
+ "-"
+ ],
+ [
+ "▁",
+ "(-"
+ ],
+ [
+ "Load",
+ "er"
+ ],
+ [
+ "Lo",
+ "ader"
+ ],
+ [
+ "▁Wil",
+ "son"
+ ],
+ [
+ "▁l",
+ "ocked"
+ ],
+ [
+ "▁loc",
+ "ked"
+ ],
+ [
+ "▁lock",
+ "ed"
+ ],
+ [
+ "▁",
+ "locked"
+ ],
+ [
+ ":",
+ ""
+ ],
+ [
+ "▁O",
+ "d"
+ ],
+ [
+ "▁P",
+ "rote"
+ ],
+ [
+ "▁Pro",
+ "te"
+ ],
+ [
+ "▁Pr",
+ "ote"
+ ],
+ [
+ "▁",
+ "Prote"
+ ],
+ [
+ "▁dis",
+ "abled"
+ ],
+ [
+ "▁disable",
+ "d"
+ ],
+ [
+ "▁",
+ "disabled"
+ ],
+ [
+ "▁h",
+ "atte"
+ ],
+ [
+ "▁hat",
+ "te"
+ ],
+ [
+ "▁sh",
+ "out"
+ ],
+ [
+ "▁con",
+ "structor"
+ ],
+ [
+ "▁construct",
+ "or"
+ ],
+ [
+ "▁constru",
+ "ctor"
+ ],
+ [
+ "▁",
+ "constructor"
+ ],
+ [
+ "б",
+ "і"
+ ],
+ [
+ "▁t",
+ "ras"
+ ],
+ [
+ "▁tr",
+ "as"
+ ],
+ [
+ "▁tra",
+ "s"
+ ],
+ [
+ "▁",
+ "tras"
+ ],
+ [
+ "▁F",
+ "ather"
+ ],
+ [
+ "▁Fa",
+ "ther"
+ ],
+ [
+ "▁Fat",
+ "her"
+ ],
+ [
+ "▁ad",
+ "j"
+ ],
+ [
+ "▁",
+ "adj"
+ ],
+ [
+ "▁Carol",
+ "ina"
+ ],
+ [
+ "▁F",
+ "ood"
+ ],
+ [
+ "▁Fo",
+ "od"
+ ],
+ [
+ "ba",
+ "d"
+ ],
+ [
+ "b",
+ "ad"
+ ],
+ [
+ "at",
+ "ore"
+ ],
+ [
+ "ator",
+ "e"
+ ],
+ [
+ "ato",
+ "re"
+ ],
+ [
+ "param",
+ "eters"
+ ],
+ [
+ "parameter",
+ "s"
+ ],
+ [
+ "▁F",
+ "ull"
+ ],
+ [
+ "▁Fu",
+ "ll"
+ ],
+ [
+ "▁",
+ "Full"
+ ],
+ [
+ "[",
+ "-"
+ ],
+ [
+ "▁\"",
+ "#"
+ ],
+ [
+ "▁T",
+ "ry"
+ ],
+ [
+ "▁Tr",
+ "y"
+ ],
+ [
+ "▁",
+ "Try"
+ ],
+ [
+ "сь",
+ "кої"
+ ],
+ [
+ "сько",
+ "ї"
+ ],
+ [
+ "▁ex",
+ "haust"
+ ],
+ [
+ "▁sc",
+ "roll"
+ ],
+ [
+ "▁scr",
+ "oll"
+ ],
+ [
+ "▁",
+ "scroll"
+ ],
+ [
+ "_",
+ ";"
+ ],
+ [
+ "Wh",
+ "o"
+ ],
+ [
+ "W",
+ "ho"
+ ],
+ [
+ "▁deliver",
+ "ed"
+ ],
+ [
+ "▁re",
+ "ferred"
+ ],
+ [
+ "▁refer",
+ "red"
+ ],
+ [
+ "▁pro",
+ "spect"
+ ],
+ [
+ "▁pros",
+ "pect"
+ ],
+ [
+ "sc",
+ "an"
+ ],
+ [
+ "s",
+ "can"
+ ],
+ [
+ "▁mod",
+ "ified"
+ ],
+ [
+ "▁",
+ "modified"
+ ],
+ [
+ "Gener",
+ "ator"
+ ],
+ [
+ "▁ex",
+ "cess"
+ ],
+ [
+ "▁exc",
+ "ess"
+ ],
+ [
+ "▁k",
+ "g"
+ ],
+ [
+ "▁",
+ "kg"
+ ],
+ [
+ "ze",
+ "t"
+ ],
+ [
+ "z",
+ "et"
+ ],
+ [
+ "ic",
+ "z"
+ ],
+ [
+ "i",
+ "cz"
+ ],
+ [
+ "clip",
+ "se"
+ ],
+ [
+ "cli",
+ "pse"
+ ],
+ [
+ "▁t",
+ "ank"
+ ],
+ [
+ "▁tan",
+ "k"
+ ],
+ [
+ "▁g",
+ "uns"
+ ],
+ [
+ "▁gu",
+ "ns"
+ ],
+ [
+ "▁gun",
+ "s"
+ ],
+ [
+ "▁G",
+ "es"
+ ],
+ [
+ "▁Ge",
+ "s"
+ ],
+ [
+ "in",
+ "ton"
+ ],
+ [
+ "int",
+ "on"
+ ],
+ [
+ "into",
+ "n"
+ ],
+ [
+ "▁Wed",
+ "nesday"
+ ],
+ [
+ "▁main",
+ "ly"
+ ],
+ [
+ "par",
+ "ser"
+ ],
+ [
+ "parse",
+ "r"
+ ],
+ [
+ "pars",
+ "er"
+ ],
+ [
+ "▁effect",
+ "ively"
+ ],
+ [
+ "▁effective",
+ "ly"
+ ],
+ [
+ "▁К",
+ "у"
+ ],
+ [
+ "▁res",
+ "ident"
+ ],
+ [
+ "▁resid",
+ "ent"
+ ],
+ [
+ "▁L",
+ "i"
+ ],
+ [
+ "▁",
+ "Li"
+ ],
+ [
+ "▁f",
+ "lying"
+ ],
+ [
+ "▁fl",
+ "ying"
+ ],
+ [
+ "▁fly",
+ "ing"
+ ],
+ [
+ "▁may",
+ "or"
+ ],
+ [
+ "▁mayo",
+ "r"
+ ],
+ [
+ "ü",
+ "h"
+ ],
+ [
+ "ut",
+ "a"
+ ],
+ [
+ "u",
+ "ta"
+ ],
+ [
+ "▁col",
+ "our"
+ ],
+ [
+ "▁air",
+ "craft"
+ ],
+ [
+ "ter",
+ "ior"
+ ],
+ [
+ "te",
+ "rior"
+ ],
+ [
+ "n",
+ "r"
+ ],
+ [
+ "▁ke",
+ "eps"
+ ],
+ [
+ "▁keep",
+ "s"
+ ],
+ [
+ "fa",
+ "n"
+ ],
+ [
+ "f",
+ "an"
+ ],
+ [
+ "▁sh",
+ "irt"
+ ],
+ [
+ "▁",
+ "shirt"
+ ],
+ [
+ "Com",
+ "par"
+ ],
+ [
+ "Comp",
+ "ar"
+ ],
+ [
+ "▁E",
+ "th"
+ ],
+ [
+ "▁Et",
+ "h"
+ ],
+ [
+ "Ma",
+ "c"
+ ],
+ [
+ "M",
+ "ac"
+ ],
+ [
+ "cle",
+ "an"
+ ],
+ [
+ "c",
+ "lean"
+ ],
+ [
+ "sl",
+ "ice"
+ ],
+ [
+ "cz",
+ "y"
+ ],
+ [
+ "c",
+ "zy"
+ ],
+ [
+ "▁g",
+ "ender"
+ ],
+ [
+ "▁gen",
+ "der"
+ ],
+ [
+ "▁ge",
+ "nder"
+ ],
+ [
+ "▁",
+ "gender"
+ ],
+ [
+ "▁b",
+ "utter"
+ ],
+ [
+ "▁but",
+ "ter"
+ ],
+ [
+ "▁butt",
+ "er"
+ ],
+ [
+ "AU",
+ "T"
+ ],
+ [
+ "A",
+ "UT"
+ ],
+ [
+ "▁E",
+ "lement"
+ ],
+ [
+ "▁El",
+ "ement"
+ ],
+ [
+ "▁Ele",
+ "ment"
+ ],
+ [
+ "▁",
+ "Element"
+ ],
+ [
+ "Fi",
+ "n"
+ ],
+ [
+ "F",
+ "in"
+ ],
+ [
+ "dm",
+ "a"
+ ],
+ [
+ "d",
+ "ma"
+ ],
+ [
+ "sam",
+ "ple"
+ ],
+ [
+ "s",
+ "ample"
+ ],
+ [
+ "Reg",
+ "istry"
+ ],
+ [
+ "▁class",
+ "ic"
+ ],
+ [
+ "▁dr",
+ "ove"
+ ],
+ [
+ "▁dro",
+ "ve"
+ ],
+ [
+ "p",
+ "b"
+ ],
+ [
+ "def",
+ "ined"
+ ],
+ [
+ "define",
+ "d"
+ ],
+ [
+ "d",
+ "efined"
+ ],
+ [
+ "▁re",
+ "ward"
+ ],
+ [
+ "▁r",
+ "eward"
+ ],
+ [
+ "ya",
+ "l"
+ ],
+ [
+ "y",
+ "al"
+ ],
+ [
+ "])",
+ ","
+ ],
+ [
+ "]",
+ "),"
+ ],
+ [
+ "▁B",
+ "AS"
+ ],
+ [
+ "▁BA",
+ "S"
+ ],
+ [
+ "▁hy",
+ "per"
+ ],
+ [
+ "▁hyp",
+ "er"
+ ],
+ [
+ "▁",
+ "hyper"
+ ],
+ [
+ "▁Н",
+ "и"
+ ],
+ [
+ "▁)",
+ "."
+ ],
+ [
+ "▁",
+ ")."
+ ],
+ [
+ "Ps",
+ "i"
+ ],
+ [
+ "P",
+ "si"
+ ],
+ [
+ "▁ent",
+ "ries"
+ ],
+ [
+ "▁entr",
+ "ies"
+ ],
+ [
+ "▁",
+ "entries"
+ ],
+ [
+ "▁King",
+ "dom"
+ ],
+ [
+ "▁S",
+ "ong"
+ ],
+ [
+ "▁So",
+ "ng"
+ ],
+ [
+ "▁Son",
+ "g"
+ ],
+ [
+ "▁prom",
+ "pt"
+ ],
+ [
+ "cent",
+ "ering"
+ ],
+ [
+ "center",
+ "ing"
+ ],
+ [
+ "▁H",
+ "olly"
+ ],
+ [
+ "▁Hol",
+ "ly"
+ ],
+ [
+ "▁Holl",
+ "y"
+ ],
+ [
+ "em",
+ "an"
+ ],
+ [
+ "ema",
+ "n"
+ ],
+ [
+ "e",
+ "man"
+ ],
+ [
+ "▁pain",
+ "ting"
+ ],
+ [
+ "▁paint",
+ "ing"
+ ],
+ [
+ "▁form",
+ "ation"
+ ],
+ [
+ "▁format",
+ "ion"
+ ],
+ [
+ "▁",
+ "formation"
+ ],
+ [
+ "▁Re",
+ "quest"
+ ],
+ [
+ "▁Requ",
+ "est"
+ ],
+ [
+ "▁",
+ "Request"
+ ],
+ [
+ "cont",
+ "roller"
+ ],
+ [
+ "control",
+ "ler"
+ ],
+ [
+ "Reg",
+ "ion"
+ ],
+ [
+ "P",
+ "Y"
+ ],
+ [
+ "id",
+ "ades"
+ ],
+ [
+ "ida",
+ "des"
+ ],
+ [
+ "idad",
+ "es"
+ ],
+ [
+ "idade",
+ "s"
+ ],
+ [
+ "T",
+ "L"
+ ],
+ [
+ "▁dis",
+ "able"
+ ],
+ [
+ "▁",
+ "disable"
+ ],
+ [
+ "▁re",
+ "in"
+ ],
+ [
+ "ri",
+ "cal"
+ ],
+ [
+ "ric",
+ "al"
+ ],
+ [
+ "r",
+ "ical"
+ ],
+ [
+ "\"",
+ "\r"
+ ],
+ [
+ "%",
+ ")"
+ ],
+ [
+ "▁S",
+ "ab"
+ ],
+ [
+ "▁Sa",
+ "b"
+ ],
+ [
+ "▁With",
+ "out"
+ ],
+ [
+ "▁",
+ "Without"
+ ],
+ [
+ "Se",
+ "rv"
+ ],
+ [
+ "Ser",
+ "v"
+ ],
+ [
+ "S",
+ "erv"
+ ],
+ [
+ "▁Sh",
+ "ort"
+ ],
+ [
+ "▁",
+ "Short"
+ ],
+ [
+ "▁",
+ "ю"
+ ],
+ [
+ "▁re",
+ "sc"
+ ],
+ [
+ "▁r",
+ "esc"
+ ],
+ [
+ "▁res",
+ "c"
+ ],
+ [
+ "▁",
+ "resc"
+ ],
+ [
+ "▁pattern",
+ "s"
+ ],
+ [
+ "▁Array",
+ "List"
+ ],
+ [
+ "▁",
+ "ArrayList"
+ ],
+ [
+ "sym",
+ "bol"
+ ],
+ [
+ "s",
+ "ymbol"
+ ],
+ [
+ "ac",
+ "o"
+ ],
+ [
+ "a",
+ "co"
+ ],
+ [
+ "▁H",
+ "om"
+ ],
+ [
+ "▁Ho",
+ "m"
+ ],
+ [
+ "▁",
+ "Hom"
+ ],
+ [
+ "he",
+ "lp"
+ ],
+ [
+ "hel",
+ "p"
+ ],
+ [
+ "▁h",
+ "asta"
+ ],
+ [
+ "▁has",
+ "ta"
+ ],
+ [
+ "▁ha",
+ "sta"
+ ],
+ [
+ "▁hast",
+ "a"
+ ],
+ [
+ "▁inst",
+ "alled"
+ ],
+ [
+ "▁install",
+ "ed"
+ ],
+ [
+ "at",
+ "ie"
+ ],
+ [
+ "ati",
+ "e"
+ ],
+ [
+ "▁vis",
+ "ited"
+ ],
+ [
+ "▁visit",
+ "ed"
+ ],
+ [
+ "▁Б",
+ "е"
+ ],
+ [
+ "){",
+ "\\"
+ ],
+ [
+ ")",
+ "{\\"
+ ],
+ [
+ "▁des",
+ "de"
+ ],
+ [
+ "J",
+ "ECT"
+ ],
+ [
+ "▁d",
+ "rew"
+ ],
+ [
+ "▁dr",
+ "ew"
+ ],
+ [
+ "▁dre",
+ "w"
+ ],
+ [
+ "▁St",
+ "ock"
+ ],
+ [
+ "▁Sto",
+ "ck"
+ ],
+ [
+ "▁C",
+ "ru"
+ ],
+ [
+ "▁Cr",
+ "u"
+ ],
+ [
+ "DE",
+ "F"
+ ],
+ [
+ "D",
+ "EF"
+ ],
+ [
+ "ob",
+ "by"
+ ],
+ [
+ "obb",
+ "y"
+ ],
+ [
+ "iz",
+ "able"
+ ],
+ [
+ "iza",
+ "ble"
+ ],
+ [
+ "og",
+ "ether"
+ ],
+ [
+ "oge",
+ "ther"
+ ],
+ [
+ "▁a",
+ "ber"
+ ],
+ [
+ "▁ab",
+ "er"
+ ],
+ [
+ "▁d",
+ "an"
+ ],
+ [
+ "▁da",
+ "n"
+ ],
+ [
+ "▁",
+ "dan"
+ ],
+ [
+ "al",
+ "is"
+ ],
+ [
+ "ali",
+ "s"
+ ],
+ [
+ "ta",
+ "il"
+ ],
+ [
+ "t",
+ "ail"
+ ],
+ [
+ "▁ex",
+ "pressed"
+ ],
+ [
+ "▁exp",
+ "ressed"
+ ],
+ [
+ "▁express",
+ "ed"
+ ],
+ [
+ "▁expr",
+ "essed"
+ ],
+ [
+ "▁A",
+ "ccess"
+ ],
+ [
+ "▁Acc",
+ "ess"
+ ],
+ [
+ "▁Ac",
+ "cess"
+ ],
+ [
+ "▁",
+ "Access"
+ ],
+ [
+ "Se",
+ "g"
+ ],
+ [
+ "S",
+ "eg"
+ ],
+ [
+ "▁L",
+ "ib"
+ ],
+ [
+ "▁Li",
+ "b"
+ ],
+ [
+ "▁",
+ "Lib"
+ ],
+ [
+ "▁sup",
+ "ports"
+ ],
+ [
+ "▁support",
+ "s"
+ ],
+ [
+ "▁supp",
+ "orts"
+ ],
+ [
+ "back",
+ "ground"
+ ],
+ [
+ "▁comm",
+ "une"
+ ],
+ [
+ "▁commun",
+ "e"
+ ],
+ [
+ "cal",
+ "led"
+ ],
+ [
+ "call",
+ "ed"
+ ],
+ [
+ "c",
+ "alled"
+ ],
+ [
+ "▁print",
+ "f"
+ ],
+ [
+ "▁prin",
+ "tf"
+ ],
+ [
+ "▁",
+ "printf"
+ ],
+ [
+ "▁Pr",
+ "ince"
+ ],
+ [
+ "▁Prin",
+ "ce"
+ ],
+ [
+ "ни",
+ "те"
+ ],
+ [
+ "de",
+ "pend"
+ ],
+ [
+ "dep",
+ "end"
+ ],
+ [
+ "▁d",
+ "els"
+ ],
+ [
+ "▁de",
+ "ls"
+ ],
+ [
+ "▁del",
+ "s"
+ ],
+ [
+ "ne",
+ "ur"
+ ],
+ [
+ "n",
+ "eur"
+ ],
+ [
+ "▁recomm",
+ "ended"
+ ],
+ [
+ "▁recommend",
+ "ed"
+ ],
+ [
+ "▁found",
+ "ed"
+ ],
+ [
+ "▁mark",
+ "ets"
+ ],
+ [
+ "▁market",
+ "s"
+ ],
+ [
+ "▁destroy",
+ "ed"
+ ],
+ [
+ "▁ab",
+ "stract"
+ ],
+ [
+ "▁abs",
+ "tract"
+ ],
+ [
+ "▁",
+ "abstract"
+ ],
+ [
+ "▁s",
+ "erie"
+ ],
+ [
+ "▁se",
+ "rie"
+ ],
+ [
+ "▁ser",
+ "ie"
+ ],
+ [
+ "▁",
+ "serie"
+ ],
+ [
+ "▁D",
+ "un"
+ ],
+ [
+ "▁Du",
+ "n"
+ ],
+ [
+ "Te",
+ "rm"
+ ],
+ [
+ "T",
+ "erm"
+ ],
+ [
+ "▁p",
+ "ortion"
+ ],
+ [
+ "▁port",
+ "ion"
+ ],
+ [
+ "ad",
+ "apter"
+ ],
+ [
+ "is",
+ "set"
+ ],
+ [
+ "iss",
+ "et"
+ ],
+ [
+ "isse",
+ "t"
+ ],
+ [
+ "че",
+ "ски"
+ ],
+ [
+ "▁in",
+ "teger"
+ ],
+ [
+ "▁inte",
+ "ger"
+ ],
+ [
+ "▁",
+ "integer"
+ ],
+ [
+ "▁return",
+ "ing"
+ ],
+ [
+ "en",
+ "ties"
+ ],
+ [
+ "ent",
+ "ies"
+ ],
+ [
+ "enti",
+ "es"
+ ],
+ [
+ "▁F",
+ "air"
+ ],
+ [
+ "▁Fa",
+ "ir"
+ ],
+ [
+ "▁U",
+ "SB"
+ ],
+ [
+ "▁US",
+ "B"
+ ],
+ [
+ "▁",
+ "USB"
+ ],
+ [
+ "▁P",
+ "rice"
+ ],
+ [
+ "▁Pr",
+ "ice"
+ ],
+ [
+ "▁Pri",
+ "ce"
+ ],
+ [
+ "▁",
+ "Price"
+ ],
+ [
+ "ig",
+ "ate"
+ ],
+ [
+ "iga",
+ "te"
+ ],
+ [
+ "i",
+ "gate"
+ ],
+ [
+ "▁sett",
+ "led"
+ ],
+ [
+ "▁settle",
+ "d"
+ ],
+ [
+ "({",
+ "\\"
+ ],
+ [
+ "(",
+ "{\\"
+ ],
+ [
+ "ne",
+ "k"
+ ],
+ [
+ "n",
+ "ek"
+ ],
+ [
+ "▁the",
+ "rm"
+ ],
+ [
+ "▁th",
+ "erm"
+ ],
+ [
+ "▁ther",
+ "m"
+ ],
+ [
+ "▁c",
+ "ig"
+ ],
+ [
+ "▁ci",
+ "g"
+ ],
+ [
+ "án",
+ "y"
+ ],
+ [
+ "á",
+ "ny"
+ ],
+ [
+ "▁invest",
+ "igation"
+ ],
+ [
+ "▁investig",
+ "ation"
+ ],
+ [
+ "om",
+ "eter"
+ ],
+ [
+ "ome",
+ "ter"
+ ],
+ [
+ "omet",
+ "er"
+ ],
+ [
+ "SU",
+ "P"
+ ],
+ [
+ "S",
+ "UP"
+ ],
+ [
+ "So",
+ "me"
+ ],
+ [
+ "Som",
+ "e"
+ ],
+ [
+ "S",
+ "ome"
+ ],
+ [
+ "si",
+ "ng"
+ ],
+ [
+ "sin",
+ "g"
+ ],
+ [
+ "s",
+ "ing"
+ ],
+ [
+ "Con",
+ "stant"
+ ],
+ [
+ "Const",
+ "ant"
+ ],
+ [
+ "▁re",
+ "tail"
+ ],
+ [
+ "▁ret",
+ "ail"
+ ],
+ [
+ "ż",
+ "y"
+ ],
+ [
+ "▁dr",
+ "inking"
+ ],
+ [
+ "▁drink",
+ "ing"
+ ],
+ [
+ "▁In",
+ "vest"
+ ],
+ [
+ "▁Inv",
+ "est"
+ ],
+ [
+ "S",
+ "V"
+ ],
+ [
+ "ig",
+ "inal"
+ ],
+ [
+ "igin",
+ "al"
+ ],
+ [
+ "igi",
+ "nal"
+ ],
+ [
+ "▁B",
+ "ow"
+ ],
+ [
+ "▁Bo",
+ "w"
+ ],
+ [
+ "{{",
+ "\\"
+ ],
+ [
+ "{",
+ "{\\"
+ ],
+ [
+ "▁ass",
+ "istance"
+ ],
+ [
+ "▁assist",
+ "ance"
+ ],
+ [
+ "▁intel",
+ "lect"
+ ],
+ [
+ "IN",
+ "IT"
+ ],
+ [
+ "au",
+ "g"
+ ],
+ [
+ "a",
+ "ug"
+ ],
+ [
+ "▁Le",
+ "on"
+ ],
+ [
+ "▁Leo",
+ "n"
+ ],
+ [
+ "Su",
+ "r"
+ ],
+ [
+ "S",
+ "ur"
+ ],
+ [
+ "▁ad",
+ "mit"
+ ],
+ [
+ "▁adm",
+ "it"
+ ],
+ [
+ "▁Com",
+ "mand"
+ ],
+ [
+ "▁Comm",
+ "and"
+ ],
+ [
+ "▁",
+ "Command"
+ ],
+ [
+ "il",
+ "les"
+ ],
+ [
+ "ill",
+ "es"
+ ],
+ [
+ "ille",
+ "s"
+ ],
+ [
+ "ro",
+ "v"
+ ],
+ [
+ "r",
+ "ov"
+ ],
+ [
+ "▁o",
+ "h"
+ ],
+ [
+ "▁",
+ "oh"
+ ],
+ [
+ "▁n",
+ "ão"
+ ],
+ [
+ "▁mat",
+ "ching"
+ ],
+ [
+ "▁match",
+ "ing"
+ ],
+ [
+ "▁g",
+ "enu"
+ ],
+ [
+ "▁gen",
+ "u"
+ ],
+ [
+ "▁ge",
+ "nu"
+ ],
+ [
+ "▁O",
+ "x"
+ ],
+ [
+ "т",
+ "ся"
+ ],
+ [
+ "not",
+ "ation"
+ ],
+ [
+ "G",
+ "O"
+ ],
+ [
+ "▁N",
+ "ap"
+ ],
+ [
+ "▁Na",
+ "p"
+ ],
+ [
+ "▁ver",
+ "ify"
+ ],
+ [
+ "▁",
+ "verify"
+ ],
+ [
+ "▁aus",
+ "si"
+ ],
+ [
+ "▁auss",
+ "i"
+ ],
+ [
+ "Date",
+ "Time"
+ ],
+ [
+ "▁su",
+ "itable"
+ ],
+ [
+ "▁suit",
+ "able"
+ ],
+ [
+ "▁ind",
+ "icate"
+ ],
+ [
+ "▁indic",
+ "ate"
+ ],
+ [
+ "▁L",
+ "ive"
+ ],
+ [
+ "▁Li",
+ "ve"
+ ],
+ [
+ "▁Liv",
+ "e"
+ ],
+ [
+ "▁",
+ "Live"
+ ],
+ [
+ "Fe",
+ "ature"
+ ],
+ [
+ "▁tr",
+ "acks"
+ ],
+ [
+ "▁track",
+ "s"
+ ],
+ [
+ "▁tra",
+ "cks"
+ ],
+ [
+ "▁has",
+ "n"
+ ],
+ [
+ "▁ha",
+ "sn"
+ ],
+ [
+ "▁J",
+ "ava"
+ ],
+ [
+ "▁Ja",
+ "va"
+ ],
+ [
+ "▁",
+ "Java"
+ ],
+ [
+ "▁close",
+ "ly"
+ ],
+ [
+ "▁clos",
+ "ely"
+ ],
+ [
+ "▁D",
+ "ad"
+ ],
+ [
+ "▁Da",
+ "d"
+ ],
+ [
+ "ce",
+ "ive"
+ ],
+ [
+ "▁Mar",
+ "ket"
+ ],
+ [
+ "▁Mark",
+ "et"
+ ],
+ [
+ "ag",
+ "y"
+ ],
+ [
+ "a",
+ "gy"
+ ],
+ [
+ "▁\"",
+ "-"
+ ],
+ [
+ "aw",
+ "n"
+ ],
+ [
+ "a",
+ "wn"
+ ],
+ [
+ "st",
+ "ell"
+ ],
+ [
+ "ste",
+ "ll"
+ ],
+ [
+ "pt",
+ "on"
+ ],
+ [
+ "pto",
+ "n"
+ ],
+ [
+ "p",
+ "ton"
+ ],
+ [
+ "ze",
+ "it"
+ ],
+ [
+ "▁V",
+ "ector"
+ ],
+ [
+ "▁Ve",
+ "ctor"
+ ],
+ [
+ "▁Vec",
+ "tor"
+ ],
+ [
+ "▁",
+ "Vector"
+ ],
+ [
+ "▁M",
+ "AX"
+ ],
+ [
+ "▁MA",
+ "X"
+ ],
+ [
+ "▁",
+ "MAX"
+ ],
+ [
+ "▁F",
+ "ederal"
+ ],
+ [
+ "▁Feder",
+ "al"
+ ],
+ [
+ "▁Fed",
+ "eral"
+ ],
+ [
+ "wa",
+ "ll"
+ ],
+ [
+ "wal",
+ "l"
+ ],
+ [
+ "w",
+ "all"
+ ],
+ [
+ "▁J",
+ "en"
+ ],
+ [
+ "▁Je",
+ "n"
+ ],
+ [
+ "de",
+ "lay"
+ ],
+ [
+ "del",
+ "ay"
+ ],
+ [
+ "▁lim",
+ "its"
+ ],
+ [
+ "▁limit",
+ "s"
+ ],
+ [
+ "▁",
+ "limits"
+ ],
+ [
+ "▁Q",
+ "uest"
+ ],
+ [
+ "▁Qu",
+ "est"
+ ],
+ [
+ "▁Que",
+ "st"
+ ],
+ [
+ "▁",
+ "Quest"
+ ],
+ [
+ "C",
+ "am"
+ ],
+ [
+ "▁F",
+ "el"
+ ],
+ [
+ "▁Fe",
+ "l"
+ ],
+ [
+ "write",
+ "r"
+ ],
+ [
+ "wr",
+ "iter"
+ ],
+ [
+ "writ",
+ "er"
+ ],
+ [
+ "w",
+ "riter"
+ ],
+ [
+ "L",
+ "P"
+ ],
+ [
+ "▁m",
+ "oves"
+ ],
+ [
+ "▁mov",
+ "es"
+ ],
+ [
+ "▁move",
+ "s"
+ ],
+ [
+ "▁mo",
+ "ves"
+ ],
+ [
+ "▁Ex",
+ "ecut"
+ ],
+ [
+ "▁",
+ "Execut"
+ ],
+ [
+ "▁D",
+ "B"
+ ],
+ [
+ "▁",
+ "DB"
+ ],
+ [
+ "ok",
+ "er"
+ ],
+ [
+ "oke",
+ "r"
+ ],
+ [
+ "o",
+ "ker"
+ ],
+ [
+ "sc",
+ "ribe"
+ ],
+ [
+ "scri",
+ "be"
+ ],
+ [
+ "scr",
+ "ibe"
+ ],
+ [
+ "scrib",
+ "e"
+ ],
+ [
+ "el",
+ "ijk"
+ ],
+ [
+ "elij",
+ "k"
+ ],
+ [
+ "eli",
+ "jk"
+ ],
+ [
+ "Const",
+ "ants"
+ ],
+ [
+ "Constant",
+ "s"
+ ],
+ [
+ "Add",
+ "r"
+ ],
+ [
+ "Ad",
+ "dr"
+ ],
+ [
+ "▁}",
+ "}"
+ ],
+ [
+ "▁",
+ "}}"
+ ],
+ [
+ "▁ch",
+ "annels"
+ ],
+ [
+ "▁channel",
+ "s"
+ ],
+ [
+ "▁",
+ "channels"
+ ],
+ [
+ "i",
+ "y"
+ ],
+ [
+ "rior",
+ "ity"
+ ],
+ [
+ "▁tr",
+ "ading"
+ ],
+ [
+ "▁trad",
+ "ing"
+ ],
+ [
+ "▁tra",
+ "ding"
+ ],
+ [
+ "▁fac",
+ "ilities"
+ ],
+ [
+ "▁facil",
+ "ities"
+ ],
+ [
+ "▁P",
+ "ack"
+ ],
+ [
+ "▁Pa",
+ "ck"
+ ],
+ [
+ "▁Pac",
+ "k"
+ ],
+ [
+ "▁",
+ "Pack"
+ ],
+ [
+ "▁s",
+ "ys"
+ ],
+ [
+ "▁sy",
+ "s"
+ ],
+ [
+ "▁",
+ "sys"
+ ],
+ [
+ "▁m",
+ "eta"
+ ],
+ [
+ "▁me",
+ "ta"
+ ],
+ [
+ "▁met",
+ "a"
+ ],
+ [
+ "▁",
+ "meta"
+ ],
+ [
+ "▁est",
+ "imate"
+ ],
+ [
+ "▁estim",
+ "ate"
+ ],
+ [
+ "▁L",
+ "ater"
+ ],
+ [
+ "▁La",
+ "ter"
+ ],
+ [
+ "▁Lat",
+ "er"
+ ],
+ [
+ "▁Late",
+ "r"
+ ],
+ [
+ "iss",
+ "ue"
+ ],
+ [
+ "▁H",
+ "aving"
+ ],
+ [
+ "▁Ha",
+ "ving"
+ ],
+ [
+ "▁Hav",
+ "ing"
+ ],
+ [
+ "▁g",
+ "uest"
+ ],
+ [
+ "▁gu",
+ "est"
+ ],
+ [
+ "▁no",
+ "body"
+ ],
+ [
+ "▁nob",
+ "ody"
+ ],
+ [
+ "dep",
+ "th"
+ ],
+ [
+ "▁z",
+ "ostał"
+ ],
+ [
+ "пе",
+ "ра"
+ ],
+ [
+ "пер",
+ "а"
+ ],
+ [
+ ")}",
+ "\\"
+ ],
+ [
+ ")",
+ "}\\"
+ ],
+ [
+ "b",
+ "g"
+ ],
+ [
+ "▁Tw",
+ "itter"
+ ],
+ [
+ "▁dark",
+ "ness"
+ ],
+ [
+ "j",
+ "pg"
+ ],
+ [
+ "con",
+ "tr"
+ ],
+ [
+ "cont",
+ "r"
+ ],
+ [
+ "ker",
+ "nel"
+ ],
+ [
+ "kern",
+ "el"
+ ],
+ [
+ "k",
+ "ernel"
+ ],
+ [
+ "]",
+ "\\"
+ ],
+ [
+ "▁ext",
+ "end"
+ ],
+ [
+ "▁",
+ "extend"
+ ],
+ [
+ "ro",
+ "c"
+ ],
+ [
+ "r",
+ "oc"
+ ],
+ [
+ "NE",
+ "T"
+ ],
+ [
+ "N",
+ "ET"
+ ],
+ [
+ "MS",
+ "G"
+ ],
+ [
+ "M",
+ "SG"
+ ],
+ [
+ "▁b",
+ "urst"
+ ],
+ [
+ "▁bur",
+ "st"
+ ],
+ [
+ "▁re",
+ "pair"
+ ],
+ [
+ "▁rep",
+ "air"
+ ],
+ [
+ "▁f",
+ "etch"
+ ],
+ [
+ "▁fet",
+ "ch"
+ ],
+ [
+ "▁",
+ "fetch"
+ ],
+ [
+ "ie",
+ "g"
+ ],
+ [
+ "i",
+ "eg"
+ ],
+ [
+ "ú",
+ "s"
+ ],
+ [
+ "Sc",
+ "reen"
+ ],
+ [
+ "S",
+ "creen"
+ ],
+ [
+ "ble",
+ "m"
+ ],
+ [
+ "bl",
+ "em"
+ ],
+ [
+ "b",
+ "lem"
+ ],
+ [
+ "App",
+ "Compat"
+ ],
+ [
+ "▁ch",
+ "ap"
+ ],
+ [
+ "▁cha",
+ "p"
+ ],
+ [
+ "▁",
+ "chap"
+ ],
+ [
+ "EL",
+ "D"
+ ],
+ [
+ "E",
+ "LD"
+ ],
+ [
+ "▁P",
+ "enn"
+ ],
+ [
+ "▁Pe",
+ "nn"
+ ],
+ [
+ "▁Pen",
+ "n"
+ ],
+ [
+ "▁prom",
+ "ote"
+ ],
+ [
+ "▁promot",
+ "e"
+ ],
+ [
+ "▁U",
+ "kr"
+ ],
+ [
+ "ar",
+ "est"
+ ],
+ [
+ "are",
+ "st"
+ ],
+ [
+ "ares",
+ "t"
+ ],
+ [
+ "a",
+ "rest"
+ ],
+ [
+ "▁s",
+ "amples"
+ ],
+ [
+ "▁sam",
+ "ples"
+ ],
+ [
+ "▁sample",
+ "s"
+ ],
+ [
+ "▁",
+ "samples"
+ ],
+ [
+ "▁G",
+ "reek"
+ ],
+ [
+ "▁Gre",
+ "ek"
+ ],
+ [
+ "▁Gree",
+ "k"
+ ],
+ [
+ "▁con",
+ "stru"
+ ],
+ [
+ "▁const",
+ "ru"
+ ],
+ [
+ "▁constr",
+ "u"
+ ],
+ [
+ "▁un",
+ "iverse"
+ ],
+ [
+ "▁univers",
+ "e"
+ ],
+ [
+ "elij",
+ "ke"
+ ],
+ [
+ "elijk",
+ "e"
+ ],
+ [
+ "▁pre",
+ "ferred"
+ ],
+ [
+ "▁prefer",
+ "red"
+ ],
+ [
+ "▁Д",
+ "е"
+ ],
+ [
+ "▁I",
+ "ra"
+ ],
+ [
+ "▁Ir",
+ "a"
+ ],
+ [
+ "▁d",
+ "ow"
+ ],
+ [
+ "▁do",
+ "w"
+ ],
+ [
+ "ag",
+ "ues"
+ ],
+ [
+ "ague",
+ "s"
+ ],
+ [
+ "agu",
+ "es"
+ ],
+ [
+ "HE",
+ "RE"
+ ],
+ [
+ "HER",
+ "E"
+ ],
+ [
+ "H",
+ "ERE"
+ ],
+ [
+ "▁exper",
+ "ts"
+ ],
+ [
+ "▁exp",
+ "erts"
+ ],
+ [
+ "▁expert",
+ "s"
+ ],
+ [
+ "Pro",
+ "tocol"
+ ],
+ [
+ "Proto",
+ "col"
+ ],
+ [
+ "PI",
+ "O"
+ ],
+ [
+ "P",
+ "IO"
+ ],
+ [
+ "▁n",
+ "az"
+ ],
+ [
+ "▁na",
+ "z"
+ ],
+ [
+ "▁K",
+ "h"
+ ],
+ [
+ "hö",
+ "r"
+ ],
+ [
+ "h",
+ "ör"
+ ],
+ [
+ "▁dist",
+ "ingu"
+ ],
+ [
+ "▁B",
+ "Y"
+ ],
+ [
+ "▁",
+ "BY"
+ ],
+ [
+ "▁se",
+ "ine"
+ ],
+ [
+ "▁sein",
+ "e"
+ ],
+ [
+ "▁sei",
+ "ne"
+ ],
+ [
+ "ep",
+ "ing"
+ ],
+ [
+ "e",
+ "ping"
+ ],
+ [
+ "▁fair",
+ "ly"
+ ],
+ [
+ "▁Me",
+ "an"
+ ],
+ [
+ "ix",
+ "er"
+ ],
+ [
+ "in",
+ "si"
+ ],
+ [
+ "ins",
+ "i"
+ ],
+ [
+ "▁author",
+ "s"
+ ],
+ [
+ "▁auth",
+ "ors"
+ ],
+ [
+ "**",
+ "."
+ ],
+ [
+ "*",
+ "*."
+ ],
+ [
+ "A",
+ "I"
+ ],
+ [
+ "▁ed",
+ "ges"
+ ],
+ [
+ "▁edge",
+ "s"
+ ],
+ [
+ "▁",
+ "edges"
+ ],
+ [
+ "▁shoot",
+ "ing"
+ ],
+ [
+ "Ad",
+ "min"
+ ],
+ [
+ "▁m",
+ "aps"
+ ],
+ [
+ "▁map",
+ "s"
+ ],
+ [
+ "▁ma",
+ "ps"
+ ],
+ [
+ "▁",
+ "maps"
+ ],
+ [
+ "ch",
+ "ant"
+ ],
+ [
+ "chan",
+ "t"
+ ],
+ [
+ "cha",
+ "nt"
+ ],
+ [
+ "▁CO",
+ "VID"
+ ],
+ [
+ "▁link",
+ "ed"
+ ],
+ [
+ "▁lin",
+ "ked"
+ ],
+ [
+ "▁",
+ "linked"
+ ],
+ [
+ "▁s",
+ "ke"
+ ],
+ [
+ "▁sk",
+ "e"
+ ],
+ [
+ "▁",
+ "ske"
+ ],
+ [
+ "▁power",
+ "s"
+ ],
+ [
+ "▁pow",
+ "ers"
+ ],
+ [
+ "á",
+ "d"
+ ],
+ [
+ "▁stom",
+ "ach"
+ ],
+ [
+ "▁us",
+ "age"
+ ],
+ [
+ "▁",
+ "usage"
+ ],
+ [
+ "▁def",
+ "end"
+ ],
+ [
+ "▁defe",
+ "nd"
+ ],
+ [
+ "▁s",
+ "ustain"
+ ],
+ [
+ "▁sus",
+ "tain"
+ ],
+ [
+ "▁sust",
+ "ain"
+ ],
+ [
+ "▁up",
+ "dates"
+ ],
+ [
+ "▁update",
+ "s"
+ ],
+ [
+ "▁as",
+ "sign"
+ ],
+ [
+ "▁ass",
+ "ign"
+ ],
+ [
+ "▁",
+ "assign"
+ ],
+ [
+ "H",
+ "L"
+ ],
+ [
+ "▁S",
+ "ea"
+ ],
+ [
+ "▁Se",
+ "a"
+ ],
+ [
+ "▁dis",
+ "cipl"
+ ],
+ [
+ "V",
+ "ideo"
+ ],
+ [
+ "▁Ch",
+ "ief"
+ ],
+ [
+ "▁Chi",
+ "ef"
+ ],
+ [
+ "▁b",
+ "unch"
+ ],
+ [
+ "▁Ob",
+ "ama"
+ ],
+ [
+ "ni",
+ "s"
+ ],
+ [
+ "n",
+ "is"
+ ],
+ [
+ "vo",
+ "r"
+ ],
+ [
+ "v",
+ "or"
+ ],
+ [
+ "▁ag",
+ "ents"
+ ],
+ [
+ "▁agent",
+ "s"
+ ],
+ [
+ "ca",
+ "s"
+ ],
+ [
+ "c",
+ "as"
+ ],
+ [
+ "ch",
+ "ter"
+ ],
+ [
+ "cht",
+ "er"
+ ],
+ [
+ "chte",
+ "r"
+ ],
+ [
+ "▁gl",
+ "anced"
+ ],
+ [
+ "▁glance",
+ "d"
+ ],
+ [
+ "support",
+ "ed"
+ ],
+ [
+ "supp",
+ "orted"
+ ],
+ [
+ "▁Cons",
+ "ider"
+ ],
+ [
+ "▁Every",
+ "one"
+ ],
+ [
+ "▁l",
+ "ect"
+ ],
+ [
+ "▁le",
+ "ct"
+ ],
+ [
+ "▁",
+ "lect"
+ ],
+ [
+ "▁St",
+ "one"
+ ],
+ [
+ "▁Sto",
+ "ne"
+ ],
+ [
+ "▁J",
+ "am"
+ ],
+ [
+ "▁Ja",
+ "m"
+ ],
+ [
+ "og",
+ "ram"
+ ],
+ [
+ "o",
+ "gram"
+ ],
+ [
+ "form",
+ "ance"
+ ],
+ [
+ "▁\\",
+ "\""
+ ],
+ [
+ "▁",
+ "\\\""
+ ],
+ [
+ "▁p",
+ "atch"
+ ],
+ [
+ "▁pat",
+ "ch"
+ ],
+ [
+ "▁",
+ "patch"
+ ],
+ [
+ "▁v",
+ "it"
+ ],
+ [
+ "▁vi",
+ "t"
+ ],
+ [
+ "Po",
+ "wer"
+ ],
+ [
+ "P",
+ "ower"
+ ],
+ [
+ "▁hard",
+ "er"
+ ],
+ [
+ "▁har",
+ "der"
+ ],
+ [
+ "An",
+ "al"
+ ],
+ [
+ "A",
+ "nal"
+ ],
+ [
+ "▁des",
+ "ired"
+ ],
+ [
+ "▁desire",
+ "d"
+ ],
+ [
+ "▁j",
+ "ug"
+ ],
+ [
+ "▁ju",
+ "g"
+ ],
+ [
+ "▁support",
+ "ing"
+ ],
+ [
+ "D",
+ "U"
+ ],
+ [
+ "]]",
+ ","
+ ],
+ [
+ "]",
+ "],"
+ ],
+ [
+ "▁Ad",
+ "ministr"
+ ],
+ [
+ "▁Admin",
+ "istr"
+ ],
+ [
+ "uck",
+ "y"
+ ],
+ [
+ "uc",
+ "ky"
+ ],
+ [
+ "▁cont",
+ "roller"
+ ],
+ [
+ "▁control",
+ "ler"
+ ],
+ [
+ "▁",
+ "controller"
+ ],
+ [
+ "▁iss",
+ "ued"
+ ],
+ [
+ "▁issue",
+ "d"
+ ],
+ [
+ "▁S",
+ "in"
+ ],
+ [
+ "▁Si",
+ "n"
+ ],
+ [
+ "▁aff",
+ "ili"
+ ],
+ [
+ "▁part",
+ "ners"
+ ],
+ [
+ "▁partner",
+ "s"
+ ],
+ [
+ "cd",
+ "ots"
+ ],
+ [
+ "cdot",
+ "s"
+ ],
+ [
+ "c",
+ "dots"
+ ],
+ [
+ "ct",
+ "ic"
+ ],
+ [
+ "C",
+ "ar"
+ ],
+ [
+ "▁N",
+ "Y"
+ ],
+ [
+ "▁",
+ "NY"
+ ],
+ [
+ "▁p",
+ "riority"
+ ],
+ [
+ "▁prior",
+ "ity"
+ ],
+ [
+ "▁",
+ "priority"
+ ],
+ [
+ "or",
+ "iginal"
+ ],
+ [
+ "orig",
+ "inal"
+ ],
+ [
+ "origin",
+ "al"
+ ],
+ [
+ "S",
+ "ql"
+ ],
+ [
+ "▁decl",
+ "ared"
+ ],
+ [
+ "▁declare",
+ "d"
+ ],
+ [
+ "▁declar",
+ "ed"
+ ],
+ [
+ "▁Hot",
+ "el"
+ ],
+ [
+ "▁b",
+ "rowser"
+ ],
+ [
+ "▁brow",
+ "ser"
+ ],
+ [
+ "▁brows",
+ "er"
+ ],
+ [
+ "▁",
+ "browser"
+ ],
+ [
+ "▁gr",
+ "ande"
+ ],
+ [
+ "▁grand",
+ "e"
+ ],
+ [
+ "▁gran",
+ "de"
+ ],
+ [
+ "▁gra",
+ "nde"
+ ],
+ [
+ "}^",
+ "\\"
+ ],
+ [
+ "}",
+ "^\\"
+ ],
+ [
+ "bo",
+ "w"
+ ],
+ [
+ "b",
+ "ow"
+ ],
+ [
+ "▁accom",
+ "mod"
+ ],
+ [
+ "Direct",
+ "ory"
+ ],
+ [
+ "▁suff",
+ "ering"
+ ],
+ [
+ "▁suffer",
+ "ing"
+ ],
+ [
+ "▁log",
+ "ger"
+ ],
+ [
+ "▁",
+ "logger"
+ ],
+ [
+ "▁break",
+ "fast"
+ ],
+ [
+ "ul",
+ "i"
+ ],
+ [
+ "u",
+ "li"
+ ],
+ [
+ "▁b",
+ "oot"
+ ],
+ [
+ "▁bo",
+ "ot"
+ ],
+ [
+ "▁",
+ "boot"
+ ],
+ [
+ "▁contribut",
+ "ion"
+ ],
+ [
+ "NE",
+ "SS"
+ ],
+ [
+ "▁T",
+ "en"
+ ],
+ [
+ "▁Te",
+ "n"
+ ],
+ [
+ "▁",
+ "Ten"
+ ],
+ [
+ "sem",
+ "ble"
+ ],
+ [
+ "semb",
+ "le"
+ ],
+ [
+ "sembl",
+ "e"
+ ],
+ [
+ "▁h",
+ "ousing"
+ ],
+ [
+ "▁hous",
+ "ing"
+ ],
+ [
+ "▁ho",
+ "using"
+ ],
+ [
+ "R",
+ "aw"
+ ],
+ [
+ "AN",
+ "CE"
+ ],
+ [
+ "▁П",
+ "ри"
+ ],
+ [
+ "▁b",
+ "rit"
+ ],
+ [
+ "▁br",
+ "it"
+ ],
+ [
+ "▁",
+ "brit"
+ ],
+ [
+ "es",
+ "sa"
+ ],
+ [
+ "ess",
+ "a"
+ ],
+ [
+ "in",
+ "son"
+ ],
+ [
+ "ins",
+ "on"
+ ],
+ [
+ "▁B",
+ "all"
+ ],
+ [
+ "▁Ba",
+ "ll"
+ ],
+ [
+ "▁Bal",
+ "l"
+ ],
+ [
+ "en",
+ "tes"
+ ],
+ [
+ "ent",
+ "es"
+ ],
+ [
+ "ente",
+ "s"
+ ],
+ [
+ "▁B",
+ "ra"
+ ],
+ [
+ "▁Br",
+ "a"
+ ],
+ [
+ "sc",
+ "ore"
+ ],
+ [
+ "s",
+ "core"
+ ],
+ [
+ "GE",
+ "R"
+ ],
+ [
+ "G",
+ "ER"
+ ],
+ [
+ "ro",
+ "ute"
+ ],
+ [
+ "rou",
+ "te"
+ ],
+ [
+ "r",
+ "oute"
+ ],
+ [
+ "ap",
+ "sed"
+ ],
+ [
+ "aps",
+ "ed"
+ ],
+ [
+ "apse",
+ "d"
+ ],
+ [
+ "ро",
+ "й"
+ ],
+ [
+ "di",
+ "ff"
+ ],
+ [
+ "d",
+ "iff"
+ ],
+ [
+ "▁broad",
+ "cast"
+ ],
+ [
+ "▁t",
+ "ar"
+ ],
+ [
+ "▁ta",
+ "r"
+ ],
+ [
+ "▁",
+ "tar"
+ ],
+ [
+ "▁de",
+ "light"
+ ],
+ [
+ "▁del",
+ "ight"
+ ],
+ [
+ ")",
+ "?"
+ ],
+ [
+ "ch",
+ "ester"
+ ],
+ [
+ "che",
+ "ster"
+ ],
+ [
+ "ches",
+ "ter"
+ ],
+ [
+ "Pl",
+ "atform"
+ ],
+ [
+ "▁emer",
+ "gency"
+ ],
+ [
+ "▁c",
+ "es"
+ ],
+ [
+ "▁ce",
+ "s"
+ ],
+ [
+ "▁",
+ "ces"
+ ],
+ [
+ "ner",
+ "ship"
+ ],
+ [
+ "ners",
+ "hip"
+ ],
+ [
+ "n",
+ "ership"
+ ],
+ [
+ "▁sit",
+ "uations"
+ ],
+ [
+ "▁situ",
+ "ations"
+ ],
+ [
+ "▁situation",
+ "s"
+ ],
+ [
+ "▁famil",
+ "jen"
+ ],
+ [
+ "▁G",
+ "eb"
+ ],
+ [
+ "▁Ge",
+ "b"
+ ],
+ [
+ "en",
+ "ta"
+ ],
+ [
+ "ent",
+ "a"
+ ],
+ [
+ "ú",
+ "blic"
+ ],
+ [
+ "▁P",
+ "lace"
+ ],
+ [
+ "▁Pl",
+ "ace"
+ ],
+ [
+ "▁",
+ "Place"
+ ],
+ [
+ "IL",
+ "L"
+ ],
+ [
+ "I",
+ "LL"
+ ],
+ [
+ "▁m",
+ "arch"
+ ],
+ [
+ "▁mar",
+ "ch"
+ ],
+ [
+ "▁fundament",
+ "al"
+ ],
+ [
+ "att",
+ "ributes"
+ ],
+ [
+ "attribute",
+ "s"
+ ],
+ [
+ "кт",
+ "и"
+ ],
+ [
+ "к",
+ "ти"
+ ],
+ [
+ "▁F",
+ "u"
+ ],
+ [
+ "F",
+ "D"
+ ],
+ [
+ "▁ра",
+ "с"
+ ],
+ [
+ "▁academ",
+ "ic"
+ ],
+ [
+ "pr",
+ "es"
+ ],
+ [
+ "pre",
+ "s"
+ ],
+ [
+ "p",
+ "res"
+ ],
+ [
+ "▁r",
+ "ising"
+ ],
+ [
+ "▁ri",
+ "sing"
+ ],
+ [
+ "▁ris",
+ "ing"
+ ],
+ [
+ "▁B",
+ "raz"
+ ],
+ [
+ "▁Br",
+ "az"
+ ],
+ [
+ "▁Bra",
+ "z"
+ ],
+ [
+ "▁rece",
+ "iving"
+ ],
+ [
+ "WAR",
+ "N"
+ ],
+ [
+ "▁jud",
+ "g"
+ ],
+ [
+ "▁necess",
+ "arily"
+ ],
+ [
+ "]",
+ "="
+ ],
+ [
+ "▁deep",
+ "ly"
+ ],
+ [
+ "▁g",
+ "ray"
+ ],
+ [
+ "▁gr",
+ "ay"
+ ],
+ [
+ "▁gra",
+ "y"
+ ],
+ [
+ "▁",
+ "gray"
+ ],
+ [
+ "He",
+ "aders"
+ ],
+ [
+ "Head",
+ "ers"
+ ],
+ [
+ "Header",
+ "s"
+ ],
+ [
+ "▁co",
+ "al"
+ ],
+ [
+ "\\",
+ "{"
+ ],
+ [
+ "Mu",
+ "t"
+ ],
+ [
+ "M",
+ "ut"
+ ],
+ [
+ "ba",
+ "ch"
+ ],
+ [
+ "b",
+ "ach"
+ ],
+ [
+ "▁pro",
+ "fit"
+ ],
+ [
+ "▁prof",
+ "it"
+ ],
+ [
+ "▁",
+ "profit"
+ ],
+ [
+ "во",
+ "го"
+ ],
+ [
+ "в",
+ "ого"
+ ],
+ [
+ "ig",
+ "s"
+ ],
+ [
+ "i",
+ "gs"
+ ],
+ [
+ "og",
+ "rap"
+ ],
+ [
+ "\";",
+ "\r"
+ ],
+ [
+ "\"",
+ ";\r"
+ ],
+ [
+ "▁adv",
+ "oc"
+ ],
+ [
+ "Gener",
+ "ated"
+ ],
+ [
+ "Generate",
+ "d"
+ ],
+ [
+ "ме",
+ "ри"
+ ],
+ [
+ "мер",
+ "и"
+ ],
+ [
+ "▁C",
+ "ond"
+ ],
+ [
+ "▁Con",
+ "d"
+ ],
+ [
+ "▁Co",
+ "nd"
+ ],
+ [
+ "▁",
+ "Cond"
+ ],
+ [
+ "▁ag",
+ "ric"
+ ],
+ [
+ "BA",
+ "SE"
+ ],
+ [
+ "B",
+ "ASE"
+ ],
+ [
+ "▁arr",
+ "ang"
+ ],
+ [
+ "▁flow",
+ "ers"
+ ],
+ [
+ "▁flower",
+ "s"
+ ],
+ [
+ "i",
+ "w"
+ ],
+ [
+ "▁]",
+ ";"
+ ],
+ [
+ "▁",
+ "];"
+ ],
+ [
+ "▁во",
+ "й"
+ ],
+ [
+ "▁",
+ "вой"
+ ],
+ [
+ "ume",
+ "rate"
+ ],
+ [
+ "umer",
+ "ate"
+ ],
+ [
+ "▁i",
+ "hr"
+ ],
+ [
+ "▁ih",
+ "r"
+ ],
+ [
+ "▁п",
+ "ар"
+ ],
+ [
+ "▁па",
+ "р"
+ ],
+ [
+ "▁",
+ "пар"
+ ],
+ [
+ "▁m",
+ "ont"
+ ],
+ [
+ "▁mon",
+ "t"
+ ],
+ [
+ "▁mo",
+ "nt"
+ ],
+ [
+ "▁",
+ "mont"
+ ],
+ [
+ "wide",
+ "hat"
+ ],
+ [
+ "m",
+ "g"
+ ],
+ [
+ "▁b",
+ "tn"
+ ],
+ [
+ "▁bt",
+ "n"
+ ],
+ [
+ "▁",
+ "btn"
+ ],
+ [
+ "▁b",
+ "esk"
+ ],
+ [
+ "▁be",
+ "sk"
+ ],
+ [
+ "▁bes",
+ "k"
+ ],
+ [
+ "▁act",
+ "s"
+ ],
+ [
+ "▁ac",
+ "ts"
+ ],
+ [
+ "▁",
+ "acts"
+ ],
+ [
+ "ó",
+ "s"
+ ],
+ [
+ "~~",
+ "~~"
+ ],
+ [
+ "▁cur",
+ "ve"
+ ],
+ [
+ "▁curv",
+ "e"
+ ],
+ [
+ "l",
+ "anguage"
+ ],
+ [
+ "▁TR",
+ "UE"
+ ],
+ [
+ "▁",
+ "TRUE"
+ ],
+ [
+ "▁cle",
+ "aning"
+ ],
+ [
+ "▁clean",
+ "ing"
+ ],
+ [
+ "Mat",
+ "h"
+ ],
+ [
+ "Ma",
+ "th"
+ ],
+ [
+ "M",
+ "ath"
+ ],
+ [
+ "▁reg",
+ "ional"
+ ],
+ [
+ "▁region",
+ "al"
+ ],
+ [
+ "▁est",
+ "imated"
+ ],
+ [
+ "▁estim",
+ "ated"
+ ],
+ [
+ "▁estimate",
+ "d"
+ ],
+ [
+ "ar",
+ "ity"
+ ],
+ [
+ "ari",
+ "ty"
+ ],
+ [
+ "ier",
+ "ung"
+ ],
+ [
+ "/",
+ "{"
+ ],
+ [
+ "jan",
+ "go"
+ ],
+ [
+ "j",
+ "ango"
+ ],
+ [
+ "$",
+ "_"
+ ],
+ [
+ "▁th",
+ "rew"
+ ],
+ [
+ "▁thr",
+ "ew"
+ ],
+ [
+ "r",
+ "q"
+ ],
+ [
+ "co",
+ "p"
+ ],
+ [
+ "c",
+ "op"
+ ],
+ [
+ "ner",
+ "gy"
+ ],
+ [
+ "▁Acc",
+ "ount"
+ ],
+ [
+ "▁Ac",
+ "count"
+ ],
+ [
+ "▁",
+ "Account"
+ ],
+ [
+ "pa",
+ "l"
+ ],
+ [
+ "p",
+ "al"
+ ],
+ [
+ "▁N",
+ "ic"
+ ],
+ [
+ "▁Ni",
+ "c"
+ ],
+ [
+ "])",
+ ")"
+ ],
+ [
+ "]",
+ "))"
+ ],
+ [
+ "▁aw",
+ "esome"
+ ],
+ [
+ "▁L",
+ "oad"
+ ],
+ [
+ "▁Lo",
+ "ad"
+ ],
+ [
+ "▁",
+ "Load"
+ ],
+ [
+ "un",
+ "nel"
+ ],
+ [
+ "unn",
+ "el"
+ ],
+ [
+ "▁r",
+ "ows"
+ ],
+ [
+ "▁ro",
+ "ws"
+ ],
+ [
+ "▁row",
+ "s"
+ ],
+ [
+ "▁",
+ "rows"
+ ],
+ [
+ "▁for",
+ "each"
+ ],
+ [
+ "▁fore",
+ "ach"
+ ],
+ [
+ "▁fo",
+ "reach"
+ ],
+ [
+ "▁",
+ "foreach"
+ ],
+ [
+ "▁P",
+ "od"
+ ],
+ [
+ "▁Po",
+ "d"
+ ],
+ [
+ "▁",
+ "Pod"
+ ],
+ [
+ "▁E",
+ "N"
+ ],
+ [
+ "▁",
+ "EN"
+ ],
+ [
+ "▁.",
+ "="
+ ],
+ [
+ "ua",
+ "te"
+ ],
+ [
+ "u",
+ "ate"
+ ],
+ [
+ "frastr",
+ "ucture"
+ ],
+ [
+ "▁W",
+ "atch"
+ ],
+ [
+ "▁Wat",
+ "ch"
+ ],
+ [
+ "▁",
+ "Watch"
+ ],
+ [
+ "St",
+ "and"
+ ],
+ [
+ "▁r",
+ "outine"
+ ],
+ [
+ "▁rout",
+ "ine"
+ ],
+ [
+ "▁p",
+ "ic"
+ ],
+ [
+ "▁pi",
+ "c"
+ ],
+ [
+ "▁",
+ "pic"
+ ],
+ [
+ "hel",
+ "per"
+ ],
+ [
+ "help",
+ "er"
+ ],
+ [
+ "▁hor",
+ "ses"
+ ],
+ [
+ "▁horse",
+ "s"
+ ],
+ [
+ "▁hors",
+ "es"
+ ],
+ [
+ "▁requ",
+ "ested"
+ ],
+ [
+ "▁request",
+ "ed"
+ ],
+ [
+ "▁-",
+ "--"
+ ],
+ [
+ "▁--",
+ "-"
+ ],
+ [
+ "▁",
+ "---"
+ ],
+ [
+ "bor",
+ "der"
+ ],
+ [
+ "b",
+ "order"
+ ],
+ [
+ "▁lif",
+ "ted"
+ ],
+ [
+ "▁lift",
+ "ed"
+ ],
+ [
+ "▁P",
+ "ed"
+ ],
+ [
+ "▁Pe",
+ "d"
+ ],
+ [
+ "Im",
+ "port"
+ ],
+ [
+ "Imp",
+ "ort"
+ ],
+ [
+ "љ",
+ "е"
+ ],
+ [
+ "▁Л",
+ "и"
+ ],
+ [
+ "▁m",
+ "yst"
+ ],
+ [
+ "▁my",
+ "st"
+ ],
+ [
+ "TH",
+ "ER"
+ ],
+ [
+ "THE",
+ "R"
+ ],
+ [
+ "T",
+ "HER"
+ ],
+ [
+ "▁A",
+ "C"
+ ],
+ [
+ "▁",
+ "AC"
+ ],
+ [
+ "Pro",
+ "xy"
+ ],
+ [
+ "Pr",
+ "oxy"
+ ],
+ [
+ "pro",
+ "v"
+ ],
+ [
+ "pr",
+ "ov"
+ ],
+ [
+ "p",
+ "rov"
+ ],
+ [
+ "▁N",
+ "ik"
+ ],
+ [
+ "▁Ni",
+ "k"
+ ],
+ [
+ "he",
+ "mat"
+ ],
+ [
+ "hem",
+ "at"
+ ],
+ [
+ "h",
+ "emat"
+ ],
+ [
+ "он",
+ "аль"
+ ],
+ [
+ "она",
+ "ль"
+ ],
+ [
+ "о",
+ "наль"
+ ],
+ [
+ "▁\"",
+ "."
+ ],
+ [
+ "▁",
+ "\"."
+ ],
+ [
+ "ul",
+ "ui"
+ ],
+ [
+ "ulu",
+ "i"
+ ],
+ [
+ "▁impro",
+ "ved"
+ ],
+ [
+ "▁improve",
+ "d"
+ ],
+ [
+ "ie",
+ "ren"
+ ],
+ [
+ "ier",
+ "en"
+ ],
+ [
+ "iere",
+ "n"
+ ],
+ [
+ "i",
+ "eren"
+ ],
+ [
+ "oc",
+ "olate"
+ ],
+ [
+ "ocol",
+ "ate"
+ ],
+ [
+ "oco",
+ "late"
+ ],
+ [
+ "Sc",
+ "he"
+ ],
+ [
+ "Sch",
+ "e"
+ ],
+ [
+ "S",
+ "che"
+ ],
+ [
+ "un",
+ "ic"
+ ],
+ [
+ "uni",
+ "c"
+ ],
+ [
+ "u",
+ "nic"
+ ],
+ [
+ "▁Profess",
+ "or"
+ ],
+ [
+ "ie",
+ "ler"
+ ],
+ [
+ "iel",
+ "er"
+ ],
+ [
+ "iele",
+ "r"
+ ],
+ [
+ "i",
+ "eler"
+ ],
+ [
+ "▁d",
+ "uration"
+ ],
+ [
+ "▁dur",
+ "ation"
+ ],
+ [
+ "▁",
+ "duration"
+ ],
+ [
+ "▁time",
+ "out"
+ ],
+ [
+ "▁",
+ "timeout"
+ ],
+ [
+ "ho",
+ "m"
+ ],
+ [
+ "h",
+ "om"
+ ],
+ [
+ "▁l",
+ "ux"
+ ],
+ [
+ "▁lu",
+ "x"
+ ],
+ [
+ "▁t",
+ "rab"
+ ],
+ [
+ "▁tr",
+ "ab"
+ ],
+ [
+ "▁tra",
+ "b"
+ ],
+ [
+ "it",
+ "ary"
+ ],
+ [
+ "ita",
+ "ry"
+ ],
+ [
+ "itar",
+ "y"
+ ],
+ [
+ "њ",
+ "е"
+ ],
+ [
+ "▁insp",
+ "ired"
+ ],
+ [
+ "▁inspir",
+ "ed"
+ ],
+ [
+ "▁inspire",
+ "d"
+ ],
+ [
+ "})",
+ "\\"
+ ],
+ [
+ "}",
+ ")\\"
+ ],
+ [
+ "is",
+ "ely"
+ ],
+ [
+ "ise",
+ "ly"
+ ],
+ [
+ "ial",
+ "s"
+ ],
+ [
+ "ia",
+ "ls"
+ ],
+ [
+ "i",
+ "als"
+ ],
+ [
+ "▁V",
+ "or"
+ ],
+ [
+ "▁Vo",
+ "r"
+ ],
+ [
+ "▁enh",
+ "ance"
+ ],
+ [
+ "▁l",
+ "ucky"
+ ],
+ [
+ "▁luck",
+ "y"
+ ],
+ [
+ "▁luc",
+ "ky"
+ ],
+ [
+ "W",
+ "orld"
+ ],
+ [
+ "el",
+ "o"
+ ],
+ [
+ "e",
+ "lo"
+ ],
+ [
+ "if",
+ "iers"
+ ],
+ [
+ "ifier",
+ "s"
+ ],
+ [
+ "ifi",
+ "ers"
+ ],
+ [
+ "▁f",
+ "acing"
+ ],
+ [
+ "▁fac",
+ "ing"
+ ],
+ [
+ "▁fa",
+ "cing"
+ ],
+ [
+ "▁appreci",
+ "ate"
+ ],
+ [
+ "▁",
+ "être"
+ ],
+ [
+ "▁ben",
+ "ch"
+ ],
+ [
+ "▁",
+ "bench"
+ ],
+ [
+ "at",
+ "ted"
+ ],
+ [
+ "att",
+ "ed"
+ ],
+ [
+ "atte",
+ "d"
+ ],
+ [
+ "gen",
+ "ce"
+ ],
+ [
+ "g",
+ "ence"
+ ],
+ [
+ "c",
+ "ourse"
+ ],
+ [
+ "▁t",
+ "ub"
+ ],
+ [
+ "▁tu",
+ "b"
+ ],
+ [
+ "▁l",
+ "ors"
+ ],
+ [
+ "▁lo",
+ "rs"
+ ],
+ [
+ "▁mis",
+ "take"
+ ],
+ [
+ "▁mist",
+ "ake"
+ ],
+ [
+ "no",
+ "m"
+ ],
+ [
+ "n",
+ "om"
+ ],
+ [
+ "▁p",
+ "aus"
+ ],
+ [
+ "▁pa",
+ "us"
+ ],
+ [
+ "▁\"",
+ "\";"
+ ],
+ [
+ "▁\"\"",
+ ";"
+ ],
+ [
+ "▁su",
+ "bs"
+ ],
+ [
+ "▁sub",
+ "s"
+ ],
+ [
+ "▁st",
+ "ato"
+ ],
+ [
+ "▁stat",
+ "o"
+ ],
+ [
+ "▁sta",
+ "to"
+ ],
+ [
+ "$",
+ ")"
+ ],
+ [
+ "▁g",
+ "ay"
+ ],
+ [
+ "▁ga",
+ "y"
+ ],
+ [
+ "or",
+ "ry"
+ ],
+ [
+ "orr",
+ "y"
+ ],
+ [
+ "▁veh",
+ "icles"
+ ],
+ [
+ "▁vehicle",
+ "s"
+ ],
+ [
+ "▁br",
+ "ill"
+ ],
+ [
+ "ma",
+ "y"
+ ],
+ [
+ "m",
+ "ay"
+ ],
+ [
+ "re",
+ "sp"
+ ],
+ [
+ "res",
+ "p"
+ ],
+ [
+ "r",
+ "esp"
+ ],
+ [
+ "▁w",
+ "ore"
+ ],
+ [
+ "▁wor",
+ "e"
+ ],
+ [
+ "▁wo",
+ "re"
+ ],
+ [
+ "j",
+ "ą"
+ ],
+ [
+ "b",
+ "p"
+ ],
+ [
+ "on",
+ "el"
+ ],
+ [
+ "one",
+ "l"
+ ],
+ [
+ "o",
+ "nel"
+ ],
+ [
+ "▁C",
+ "R"
+ ],
+ [
+ "▁",
+ "CR"
+ ],
+ [
+ "▁di",
+ "agn"
+ ],
+ [
+ "▁dia",
+ "gn"
+ ],
+ [
+ "math",
+ "sf"
+ ],
+ [
+ "▁hol",
+ "iday"
+ ],
+ [
+ "▁achie",
+ "ved"
+ ],
+ [
+ "▁achieve",
+ "d"
+ ],
+ [
+ "▁{",
+ "'"
+ ],
+ [
+ "▁",
+ "{'"
+ ],
+ [
+ "▁Re",
+ "source"
+ ],
+ [
+ "▁Res",
+ "ource"
+ ],
+ [
+ "▁",
+ "Resource"
+ ],
+ [
+ "▁h",
+ "i"
+ ],
+ [
+ "▁",
+ "hi"
+ ],
+ [
+ "▁b",
+ "ra"
+ ],
+ [
+ "▁br",
+ "a"
+ ],
+ [
+ "▁",
+ "bra"
+ ],
+ [
+ "▁CON",
+ "DITION"
+ ],
+ [
+ "ct",
+ "r"
+ ],
+ [
+ "c",
+ "tr"
+ ],
+ [
+ "▁W",
+ "rite"
+ ],
+ [
+ "▁Writ",
+ "e"
+ ],
+ [
+ "▁Wr",
+ "ite"
+ ],
+ [
+ "▁",
+ "Write"
+ ],
+ [
+ "is",
+ "hop"
+ ],
+ [
+ "ish",
+ "op"
+ ],
+ [
+ "i",
+ "shop"
+ ],
+ [
+ "OL",
+ "D"
+ ],
+ [
+ "O",
+ "LD"
+ ],
+ [
+ "▁c",
+ "pu"
+ ],
+ [
+ "▁cp",
+ "u"
+ ],
+ [
+ "▁",
+ "cpu"
+ ],
+ [
+ "▁occ",
+ "urs"
+ ],
+ [
+ "▁occur",
+ "s"
+ ],
+ [
+ "▁oc",
+ "curs"
+ ],
+ [
+ "ó",
+ "ł"
+ ],
+ [
+ "str",
+ "aint"
+ ],
+ [
+ "stra",
+ "int"
+ ],
+ [
+ "▁nu",
+ "clear"
+ ],
+ [
+ "▁nuc",
+ "lear"
+ ],
+ [
+ "▁nucle",
+ "ar"
+ ],
+ [
+ "Ar",
+ "ea"
+ ],
+ [
+ "Are",
+ "a"
+ ],
+ [
+ "A",
+ "rea"
+ ],
+ [
+ "cl",
+ "uster"
+ ],
+ [
+ "▁surround",
+ "ing"
+ ],
+ [
+ "▁J",
+ "uan"
+ ],
+ [
+ "▁Ju",
+ "an"
+ ],
+ [
+ "▁pr",
+ "ima"
+ ],
+ [
+ "▁prim",
+ "a"
+ ],
+ [
+ "▁pri",
+ "ma"
+ ],
+ [
+ "▁South",
+ "ern"
+ ],
+ [
+ "▁Sou",
+ "thern"
+ ],
+ [
+ "it",
+ "ty"
+ ],
+ [
+ "itt",
+ "y"
+ ],
+ [
+ "i",
+ "tty"
+ ],
+ [
+ "▁As",
+ "sembly"
+ ],
+ [
+ "▁",
+ "Assembly"
+ ],
+ [
+ "el",
+ "em"
+ ],
+ [
+ "ele",
+ "m"
+ ],
+ [
+ "e",
+ "lem"
+ ],
+ [
+ "ad",
+ "i"
+ ],
+ [
+ "a",
+ "di"
+ ],
+ [
+ "ér",
+ "al"
+ ],
+ [
+ "éra",
+ "l"
+ ],
+ [
+ "é",
+ "ral"
+ ],
+ [
+ "▁W",
+ "at"
+ ],
+ [
+ "▁Wa",
+ "t"
+ ],
+ [
+ "▁R",
+ "adio"
+ ],
+ [
+ "▁Rad",
+ "io"
+ ],
+ [
+ "▁",
+ "Radio"
+ ],
+ [
+ "▁g",
+ "egen"
+ ],
+ [
+ "▁ge",
+ "gen"
+ ],
+ [
+ "▁T",
+ "ony"
+ ],
+ [
+ "▁To",
+ "ny"
+ ],
+ [
+ "▁Ton",
+ "y"
+ ],
+ [
+ "pr",
+ "essed"
+ ],
+ [
+ "press",
+ "ed"
+ ],
+ [
+ "pres",
+ "sed"
+ ],
+ [
+ "p",
+ "ressed"
+ ],
+ [
+ "▁An",
+ "ne"
+ ],
+ [
+ "▁Ann",
+ "e"
+ ],
+ [
+ "▁N",
+ "S"
+ ],
+ [
+ "▁",
+ "NS"
+ ],
+ [
+ "▁P",
+ "ak"
+ ],
+ [
+ "▁Pa",
+ "k"
+ ],
+ [
+ "▁C",
+ "ivil"
+ ],
+ [
+ "▁Ci",
+ "vil"
+ ],
+ [
+ "▁th",
+ "rown"
+ ],
+ [
+ "▁throw",
+ "n"
+ ],
+ [
+ "▁thr",
+ "own"
+ ],
+ [
+ "▁thro",
+ "wn"
+ ],
+ [
+ "NO",
+ "NE"
+ ],
+ [
+ "NON",
+ "E"
+ ],
+ [
+ "N",
+ "ONE"
+ ],
+ [
+ "▁p",
+ "ump"
+ ],
+ [
+ "▁pu",
+ "mp"
+ ],
+ [
+ "▁s",
+ "olve"
+ ],
+ [
+ "▁sol",
+ "ve"
+ ],
+ [
+ "EN",
+ "ABLE"
+ ],
+ [
+ "▁Ph",
+ "ys"
+ ],
+ [
+ "▁",
+ "Phys"
+ ],
+ [
+ "▁]",
+ ","
+ ],
+ [
+ "▁",
+ "],"
+ ],
+ [
+ "PO",
+ "SE"
+ ],
+ [
+ "POS",
+ "E"
+ ],
+ [
+ "kt",
+ "et"
+ ],
+ [
+ "kte",
+ "t"
+ ],
+ [
+ "▁F",
+ "ab"
+ ],
+ [
+ "▁Fa",
+ "b"
+ ],
+ [
+ "valid",
+ "ate"
+ ],
+ [
+ "Iter",
+ "ator"
+ ],
+ [
+ "cond",
+ "ition"
+ ],
+ [
+ "re",
+ "du"
+ ],
+ [
+ "red",
+ "u"
+ ],
+ [
+ "r",
+ "edu"
+ ],
+ [
+ "▁neg",
+ "oti"
+ ],
+ [
+ "an",
+ "no"
+ ],
+ [
+ "ann",
+ "o"
+ ],
+ [
+ "▁s",
+ "ans"
+ ],
+ [
+ "▁sa",
+ "ns"
+ ],
+ [
+ "▁san",
+ "s"
+ ],
+ [
+ "▁U",
+ "l"
+ ],
+ [
+ "CH",
+ "AR"
+ ],
+ [
+ "▁ed",
+ "ition"
+ ],
+ [
+ "▁edit",
+ "ion"
+ ],
+ [
+ "▁spect",
+ "rum"
+ ],
+ [
+ "or",
+ "ie"
+ ],
+ [
+ "ori",
+ "e"
+ ],
+ [
+ "o",
+ "rie"
+ ],
+ [
+ "▁execut",
+ "ion"
+ ],
+ [
+ "▁exec",
+ "ution"
+ ],
+ [
+ "P",
+ "lease"
+ ],
+ [
+ "▁B",
+ "O"
+ ],
+ [
+ "▁",
+ "BO"
+ ],
+ [
+ "UR",
+ "N"
+ ],
+ [
+ "▁c",
+ "ow"
+ ],
+ [
+ "▁co",
+ "w"
+ ],
+ [
+ "▁",
+ "cow"
+ ],
+ [
+ "ст",
+ "ан"
+ ],
+ [
+ "ста",
+ "н"
+ ],
+ [
+ "с",
+ "тан"
+ ],
+ [
+ "istribut",
+ "ion"
+ ],
+ [
+ "Do",
+ "main"
+ ],
+ [
+ "Dom",
+ "ain"
+ ],
+ [
+ "▁re",
+ "aders"
+ ],
+ [
+ "▁read",
+ "ers"
+ ],
+ [
+ "▁reader",
+ "s"
+ ],
+ [
+ "▁cons",
+ "umer"
+ ],
+ [
+ "▁consum",
+ "er"
+ ],
+ [
+ "▁consume",
+ "r"
+ ],
+ [
+ "▁st",
+ "yles"
+ ],
+ [
+ "▁style",
+ "s"
+ ],
+ [
+ "▁sty",
+ "les"
+ ],
+ [
+ "▁",
+ "styles"
+ ],
+ [
+ "en",
+ "code"
+ ],
+ [
+ "enc",
+ "ode"
+ ],
+ [
+ "▁C",
+ "y"
+ ],
+ [
+ "Com",
+ "mon"
+ ],
+ [
+ "Comm",
+ "on"
+ ],
+ [
+ "▁P",
+ "rop"
+ ],
+ [
+ "▁Pro",
+ "p"
+ ],
+ [
+ "▁Pr",
+ "op"
+ ],
+ [
+ "▁",
+ "Prop"
+ ],
+ [
+ "▁ex",
+ "ecute"
+ ],
+ [
+ "▁execut",
+ "e"
+ ],
+ [
+ "▁exec",
+ "ute"
+ ],
+ [
+ "▁",
+ "execute"
+ ],
+ [
+ "▁e",
+ "q"
+ ],
+ [
+ "▁",
+ "eq"
+ ],
+ [
+ "▁vis",
+ "itors"
+ ],
+ [
+ "▁visit",
+ "ors"
+ ],
+ [
+ "▁visitor",
+ "s"
+ ],
+ [
+ "▁A",
+ "mb"
+ ],
+ [
+ "▁Am",
+ "b"
+ ],
+ [
+ "ud",
+ "ad"
+ ],
+ [
+ "uda",
+ "d"
+ ],
+ [
+ "q",
+ "quad"
+ ],
+ [
+ "▁C",
+ "ert"
+ ],
+ [
+ "▁Ce",
+ "rt"
+ ],
+ [
+ "▁Cer",
+ "t"
+ ],
+ [
+ "▁",
+ "Cert"
+ ],
+ [
+ "▁t",
+ "rop"
+ ],
+ [
+ "▁tr",
+ "op"
+ ],
+ [
+ "▁tro",
+ "p"
+ ],
+ [
+ "▁yes",
+ "terday"
+ ],
+ [
+ "ta",
+ "in"
+ ],
+ [
+ "t",
+ "ain"
+ ],
+ [
+ "L",
+ "D"
+ ],
+ [
+ "at",
+ "ro"
+ ],
+ [
+ "atr",
+ "o"
+ ],
+ [
+ "▁incre",
+ "ases"
+ ],
+ [
+ "▁increase",
+ "s"
+ ],
+ [
+ "▁W",
+ "ars"
+ ],
+ [
+ "▁War",
+ "s"
+ ],
+ [
+ "▁Wa",
+ "rs"
+ ],
+ [
+ "ne",
+ "d"
+ ],
+ [
+ "n",
+ "ed"
+ ],
+ [
+ "be",
+ "fore"
+ ],
+ [
+ "b",
+ "efore"
+ ],
+ [
+ "au",
+ "pt"
+ ],
+ [
+ "a",
+ "upt"
+ ],
+ [
+ "▁E",
+ "RR"
+ ],
+ [
+ "▁ER",
+ "R"
+ ],
+ [
+ "▁",
+ "ERR"
+ ],
+ [
+ "▁F",
+ "ord"
+ ],
+ [
+ "▁For",
+ "d"
+ ],
+ [
+ "▁Fo",
+ "rd"
+ ],
+ [
+ "▁d",
+ "alla"
+ ],
+ [
+ "▁da",
+ "lla"
+ ],
+ [
+ "▁dal",
+ "la"
+ ],
+ [
+ "▁dall",
+ "a"
+ ],
+ [
+ "UL",
+ "AR"
+ ],
+ [
+ "▁st",
+ "rike"
+ ],
+ [
+ "▁str",
+ "ike"
+ ],
+ [
+ "▁stri",
+ "ke"
+ ],
+ [
+ "Ar",
+ "r"
+ ],
+ [
+ "A",
+ "rr"
+ ],
+ [
+ "▁re",
+ "covery"
+ ],
+ [
+ "▁rec",
+ "overy"
+ ],
+ [
+ "▁recover",
+ "y"
+ ],
+ [
+ "▁Res",
+ "ponse"
+ ],
+ [
+ "▁",
+ "Response"
+ ],
+ [
+ "▁strateg",
+ "ies"
+ ],
+ [
+ "▁і",
+ "н"
+ ],
+ [
+ "▁",
+ "ін"
+ ],
+ [
+ "▁re",
+ "ar"
+ ],
+ [
+ "▁r",
+ "ear"
+ ],
+ [
+ "▁adult",
+ "s"
+ ],
+ [
+ "▁Н",
+ "е"
+ ],
+ [
+ "window",
+ "s"
+ ],
+ [
+ "wind",
+ "ows"
+ ],
+ [
+ "de",
+ "cl"
+ ],
+ [
+ "dec",
+ "l"
+ ],
+ [
+ "ol",
+ "en"
+ ],
+ [
+ "ole",
+ "n"
+ ],
+ [
+ "o",
+ "len"
+ ],
+ [
+ "▁J",
+ "ord"
+ ],
+ [
+ "▁Jo",
+ "rd"
+ ],
+ [
+ "▁K",
+ "al"
+ ],
+ [
+ "▁Ka",
+ "l"
+ ],
+ [
+ "▁c",
+ "ui"
+ ],
+ [
+ "▁cu",
+ "i"
+ ],
+ [
+ "▁П",
+ "ро"
+ ],
+ [
+ "▁S",
+ "ever"
+ ],
+ [
+ "▁Se",
+ "ver"
+ ],
+ [
+ "▁Sev",
+ "er"
+ ],
+ [
+ "▁a",
+ "le"
+ ],
+ [
+ "▁al",
+ "e"
+ ],
+ [
+ "▁",
+ "ale"
+ ],
+ [
+ "▁pe",
+ "ut"
+ ],
+ [
+ "▁peu",
+ "t"
+ ],
+ [
+ "St",
+ "ats"
+ ],
+ [
+ "Stat",
+ "s"
+ ],
+ [
+ "▁R",
+ "oss"
+ ],
+ [
+ "▁Ro",
+ "ss"
+ ],
+ [
+ "▁Ros",
+ "s"
+ ],
+ [
+ "ar",
+ "ten"
+ ],
+ [
+ "art",
+ "en"
+ ],
+ [
+ "arte",
+ "n"
+ ],
+ [
+ "sh",
+ "all"
+ ],
+ [
+ "shal",
+ "l"
+ ],
+ [
+ "sha",
+ "ll"
+ ],
+ [
+ "s",
+ "hall"
+ ],
+ [
+ "▁ent",
+ "ertain"
+ ],
+ [
+ "▁enter",
+ "tain"
+ ],
+ [
+ "▁entert",
+ "ain"
+ ],
+ [
+ "▁par",
+ "king"
+ ],
+ [
+ "▁park",
+ "ing"
+ ],
+ [
+ "но",
+ "ви"
+ ],
+ [
+ "нов",
+ "и"
+ ],
+ [
+ "er",
+ "re"
+ ],
+ [
+ "err",
+ "e"
+ ],
+ [
+ "▁fun",
+ "ding"
+ ],
+ [
+ "▁fund",
+ "ing"
+ ],
+ [
+ "▁C",
+ "le"
+ ],
+ [
+ "▁Cl",
+ "e"
+ ],
+ [
+ "▁O",
+ "t"
+ ],
+ [
+ "un",
+ "st"
+ ],
+ [
+ "uns",
+ "t"
+ ],
+ [
+ "assert",
+ "Equals"
+ ],
+ [
+ "assertEqual",
+ "s"
+ ],
+ [
+ "▁c",
+ "ancell"
+ ],
+ [
+ "▁can",
+ "cell"
+ ],
+ [
+ "▁cancel",
+ "l"
+ ],
+ [
+ "TA",
+ "G"
+ ],
+ [
+ "T",
+ "AG"
+ ],
+ [
+ "▁E",
+ "arly"
+ ],
+ [
+ "▁Earl",
+ "y"
+ ],
+ [
+ "▁feed",
+ "back"
+ ],
+ [
+ "▁p",
+ "and"
+ ],
+ [
+ "▁pan",
+ "d"
+ ],
+ [
+ "▁pa",
+ "nd"
+ ],
+ [
+ "y",
+ "o"
+ ],
+ [
+ "▁mir",
+ "ror"
+ ],
+ [
+ "▁ver",
+ "b"
+ ],
+ [
+ "▁ve",
+ "rb"
+ ],
+ [
+ "▁",
+ "verb"
+ ],
+ [
+ "▁high",
+ "light"
+ ],
+ [
+ "er",
+ "ialize"
+ ],
+ [
+ "erial",
+ "ize"
+ ],
+ [
+ "▁g",
+ "rade"
+ ],
+ [
+ "▁gr",
+ "ade"
+ ],
+ [
+ "▁grad",
+ "e"
+ ],
+ [
+ "▁gra",
+ "de"
+ ],
+ [
+ "▁",
+ "grade"
+ ],
+ [
+ "ла",
+ "сь"
+ ],
+ [
+ "▁Br",
+ "ook"
+ ],
+ [
+ "▁Bro",
+ "ok"
+ ],
+ [
+ "▁L",
+ "I"
+ ],
+ [
+ "▁",
+ "LI"
+ ],
+ [
+ "▁im",
+ "plies"
+ ],
+ [
+ "▁impl",
+ "ies"
+ ],
+ [
+ "▁e",
+ "norm"
+ ],
+ [
+ "▁en",
+ "orm"
+ ],
+ [
+ "aj",
+ "ą"
+ ],
+ [
+ "a",
+ "ją"
+ ],
+ [
+ "▁W",
+ "er"
+ ],
+ [
+ "▁We",
+ "r"
+ ],
+ [
+ "aw",
+ "ay"
+ ],
+ [
+ "awa",
+ "y"
+ ],
+ [
+ "a",
+ "way"
+ ],
+ [
+ "▁machine",
+ "s"
+ ],
+ [
+ "▁mach",
+ "ines"
+ ],
+ [
+ "▁d",
+ "ent"
+ ],
+ [
+ "▁de",
+ "nt"
+ ],
+ [
+ "▁den",
+ "t"
+ ],
+ [
+ "Id",
+ "x"
+ ],
+ [
+ "I",
+ "dx"
+ ],
+ [
+ "▁t",
+ "id"
+ ],
+ [
+ "▁ti",
+ "d"
+ ],
+ [
+ "▁",
+ "tid"
+ ],
+ [
+ ")",
+ "\""
+ ],
+ [
+ "▁m",
+ "ole"
+ ],
+ [
+ "▁mo",
+ "le"
+ ],
+ [
+ "▁mol",
+ "e"
+ ],
+ [
+ "bo",
+ "ld"
+ ],
+ [
+ "bol",
+ "d"
+ ],
+ [
+ "b",
+ "old"
+ ],
+ [
+ "CO",
+ "NT"
+ ],
+ [
+ "CON",
+ "T"
+ ],
+ [
+ "C",
+ "ONT"
+ ],
+ [
+ "▁é",
+ "p"
+ ],
+ [
+ "▁",
+ "ép"
+ ],
+ [
+ "▁cut",
+ "ting"
+ ],
+ [
+ "▁N",
+ "eg"
+ ],
+ [
+ "▁Ne",
+ "g"
+ ],
+ [
+ "▁",
+ "Neg"
+ ],
+ [
+ "▁t",
+ "ong"
+ ],
+ [
+ "▁to",
+ "ng"
+ ],
+ [
+ "▁ton",
+ "g"
+ ],
+ [
+ "▁net",
+ "works"
+ ],
+ [
+ "▁network",
+ "s"
+ ],
+ [
+ "▁F",
+ "all"
+ ],
+ [
+ "▁Fa",
+ "ll"
+ ],
+ [
+ "▁Fal",
+ "l"
+ ],
+ [
+ "▁",
+ "Fall"
+ ],
+ [
+ "gener",
+ "ated"
+ ],
+ [
+ "generate",
+ "d"
+ ],
+ [
+ "▁P",
+ "ri"
+ ],
+ [
+ "▁Pr",
+ "i"
+ ],
+ [
+ "UE",
+ "ST"
+ ],
+ [
+ "UES",
+ "T"
+ ],
+ [
+ "U",
+ "EST"
+ ],
+ [
+ "▁Be",
+ "lg"
+ ],
+ [
+ "▁Bel",
+ "g"
+ ],
+ [
+ "▁s",
+ "heet"
+ ],
+ [
+ "▁she",
+ "et"
+ ],
+ [
+ "▁",
+ "sheet"
+ ],
+ [
+ "кс",
+ "и"
+ ],
+ [
+ "к",
+ "си"
+ ],
+ [
+ "▁",
+ "†"
+ ],
+ [
+ "▁y",
+ "eah"
+ ],
+ [
+ "▁ye",
+ "ah"
+ ],
+ [
+ "▁Vict",
+ "or"
+ ],
+ [
+ "▁Vi",
+ "ctor"
+ ],
+ [
+ "▁Vic",
+ "tor"
+ ],
+ [
+ "▁R",
+ "ub"
+ ],
+ [
+ "▁Ru",
+ "b"
+ ],
+ [
+ "▁candid",
+ "ates"
+ ],
+ [
+ "▁candidate",
+ "s"
+ ],
+ [
+ "pr",
+ "és"
+ ],
+ [
+ "▁E",
+ "U"
+ ],
+ [
+ "et",
+ "r"
+ ],
+ [
+ "e",
+ "tr"
+ ],
+ [
+ "▁roll",
+ "ed"
+ ],
+ [
+ "▁",
+ "rolled"
+ ],
+ [
+ "▁P",
+ "as"
+ ],
+ [
+ "▁Pa",
+ "s"
+ ],
+ [
+ "▁Ar",
+ "thur"
+ ],
+ [
+ "Ar",
+ "ch"
+ ],
+ [
+ "Arc",
+ "h"
+ ],
+ [
+ "▁M",
+ "ann"
+ ],
+ [
+ "▁Man",
+ "n"
+ ],
+ [
+ "▁Ma",
+ "nn"
+ ],
+ [
+ "Amer",
+ "ican"
+ ],
+ [
+ "America",
+ "n"
+ ],
+ [
+ "ze",
+ "s"
+ ],
+ [
+ "z",
+ "es"
+ ],
+ [
+ "in",
+ "ners"
+ ],
+ [
+ "inn",
+ "ers"
+ ],
+ [
+ "inner",
+ "s"
+ ],
+ [
+ "▁A",
+ "uto"
+ ],
+ [
+ "▁Aut",
+ "o"
+ ],
+ [
+ "▁Au",
+ "to"
+ ],
+ [
+ "▁",
+ "Auto"
+ ],
+ [
+ "▁profess",
+ "or"
+ ],
+ [
+ "▁profes",
+ "sor"
+ ],
+ [
+ "▁)",
+ ";\r"
+ ],
+ [
+ "▁);",
+ "\r"
+ ],
+ [
+ "▁",
+ ");\r"
+ ],
+ [
+ "▁ad",
+ "dr"
+ ],
+ [
+ "▁add",
+ "r"
+ ],
+ [
+ "▁",
+ "addr"
+ ],
+ [
+ "▁Med",
+ "ical"
+ ],
+ [
+ "▁Medic",
+ "al"
+ ],
+ [
+ "▁f",
+ "ired"
+ ],
+ [
+ "▁fire",
+ "d"
+ ],
+ [
+ "▁fi",
+ "red"
+ ],
+ [
+ "▁fir",
+ "ed"
+ ],
+ [
+ "▁C",
+ "ore"
+ ],
+ [
+ "▁Co",
+ "re"
+ ],
+ [
+ "▁Cor",
+ "e"
+ ],
+ [
+ "▁",
+ "Core"
+ ],
+ [
+ "▁CON",
+ "FIG"
+ ],
+ [
+ "▁",
+ "CONFIG"
+ ],
+ [
+ "▁s",
+ "ql"
+ ],
+ [
+ "▁sq",
+ "l"
+ ],
+ [
+ "▁",
+ "sql"
+ ],
+ [
+ "▁Con",
+ "serv"
+ ],
+ [
+ "▁Cons",
+ "erv"
+ ],
+ [
+ "▁Conse",
+ "rv"
+ ],
+ [
+ "ic",
+ "hen"
+ ],
+ [
+ "ich",
+ "en"
+ ],
+ [
+ "iche",
+ "n"
+ ],
+ [
+ "i",
+ "chen"
+ ],
+ [
+ "Ver",
+ "tex"
+ ],
+ [
+ "Vert",
+ "ex"
+ ],
+ [
+ "▁H",
+ "O"
+ ],
+ [
+ "▁",
+ "HO"
+ ],
+ [
+ "Y",
+ "eah"
+ ],
+ [
+ "No",
+ "te"
+ ],
+ [
+ "Not",
+ "e"
+ ],
+ [
+ "N",
+ "ote"
+ ],
+ [
+ "▁O",
+ "K"
+ ],
+ [
+ "▁",
+ "OK"
+ ],
+ [
+ "mu",
+ "s"
+ ],
+ [
+ "m",
+ "us"
+ ],
+ [
+ "f",
+ "ocus"
+ ],
+ [
+ "aj",
+ "a"
+ ],
+ [
+ "a",
+ "ja"
+ ],
+ [
+ "r",
+ "á"
+ ],
+ [
+ "▁h",
+ "ence"
+ ],
+ [
+ "▁hen",
+ "ce"
+ ],
+ [
+ "▁execut",
+ "ive"
+ ],
+ [
+ "▁liqu",
+ "id"
+ ],
+ [
+ "uj",
+ "e"
+ ],
+ [
+ "u",
+ "je"
+ ],
+ [
+ "▁d",
+ "riven"
+ ],
+ [
+ "▁dr",
+ "iven"
+ ],
+ [
+ "▁dri",
+ "ven"
+ ],
+ [
+ "▁driv",
+ "en"
+ ],
+ [
+ "▁drive",
+ "n"
+ ],
+ [
+ "▁",
+ "driven"
+ ],
+ [
+ "ig",
+ "ue"
+ ],
+ [
+ "igu",
+ "e"
+ ],
+ [
+ "i",
+ "gue"
+ ],
+ [
+ "▁W",
+ "ik"
+ ],
+ [
+ "▁Wi",
+ "k"
+ ],
+ [
+ "R",
+ "ate"
+ ],
+ [
+ "ra",
+ "nd"
+ ],
+ [
+ "ran",
+ "d"
+ ],
+ [
+ "r",
+ "and"
+ ],
+ [
+ "Result",
+ "s"
+ ],
+ [
+ "▁cop",
+ "ies"
+ ],
+ [
+ "▁t",
+ "an"
+ ],
+ [
+ "▁ta",
+ "n"
+ ],
+ [
+ "▁",
+ "tan"
+ ],
+ [
+ "rit",
+ "eria"
+ ],
+ [
+ "rite",
+ "ria"
+ ],
+ [
+ "riter",
+ "ia"
+ ],
+ [
+ "en",
+ "en"
+ ],
+ [
+ "ene",
+ "n"
+ ],
+ [
+ "e",
+ "nen"
+ ],
+ [
+ "}_",
+ "\\"
+ ],
+ [
+ "}",
+ "_\\"
+ ],
+ [
+ "▁po",
+ "bl"
+ ],
+ [
+ "▁pob",
+ "l"
+ ],
+ [
+ "▁sou",
+ "thern"
+ ],
+ [
+ "▁south",
+ "ern"
+ ],
+ [
+ "el",
+ "n"
+ ],
+ [
+ "e",
+ "ln"
+ ],
+ [
+ "▁z",
+ "wei"
+ ],
+ [
+ "▁zwe",
+ "i"
+ ],
+ [
+ "▁zw",
+ "ei"
+ ],
+ [
+ "▁con",
+ "crete"
+ ],
+ [
+ "▁CONDITION",
+ "S"
+ ],
+ [
+ "▁dream",
+ "s"
+ ],
+ [
+ "▁dre",
+ "ams"
+ ],
+ [
+ "▁min",
+ "im"
+ ],
+ [
+ "▁mi",
+ "nim"
+ ],
+ [
+ "▁mini",
+ "m"
+ ],
+ [
+ "▁em",
+ "ployee"
+ ],
+ [
+ "▁employ",
+ "ee"
+ ],
+ [
+ "▁n",
+ "ap"
+ ],
+ [
+ "▁na",
+ "p"
+ ],
+ [
+ "▁su",
+ "spect"
+ ],
+ [
+ "▁sus",
+ "pect"
+ ],
+ [
+ "▁susp",
+ "ect"
+ ],
+ [
+ "Mo",
+ "use"
+ ],
+ [
+ "M",
+ "ouse"
+ ],
+ [
+ "▁ther",
+ "apy"
+ ],
+ [
+ "▁therap",
+ "y"
+ ],
+ [
+ "av",
+ "al"
+ ],
+ [
+ "ava",
+ "l"
+ ],
+ [
+ "a",
+ "val"
+ ],
+ [
+ "▁An",
+ "th"
+ ],
+ [
+ "▁Ant",
+ "h"
+ ],
+ [
+ "ST",
+ "ART"
+ ],
+ [
+ "st",
+ "ers"
+ ],
+ [
+ "ster",
+ "s"
+ ],
+ [
+ "ste",
+ "rs"
+ ],
+ [
+ "s",
+ "ters"
+ ],
+ [
+ "ish",
+ "ment"
+ ],
+ [
+ "fin",
+ "ite"
+ ],
+ [
+ "W",
+ "A"
+ ],
+ [
+ "v",
+ "y"
+ ],
+ [
+ "▁m",
+ "ood"
+ ],
+ [
+ "▁mo",
+ "od"
+ ],
+ [
+ "com",
+ "fort"
+ ],
+ [
+ "▁s",
+ "hr"
+ ],
+ [
+ "▁sh",
+ "r"
+ ],
+ [
+ "▁dec",
+ "ade"
+ ],
+ [
+ "я",
+ "бря"
+ ],
+ [
+ "▁'",
+ "#"
+ ],
+ [
+ "▁d",
+ "ot"
+ ],
+ [
+ "▁do",
+ "t"
+ ],
+ [
+ "▁",
+ "dot"
+ ],
+ [
+ "▁h",
+ "ill"
+ ],
+ [
+ "▁hi",
+ "ll"
+ ],
+ [
+ "▁",
+ "hill"
+ ],
+ [
+ "ar",
+ "ry"
+ ],
+ [
+ "arr",
+ "y"
+ ],
+ [
+ "cat",
+ "ch"
+ ],
+ [
+ "c",
+ "atch"
+ ],
+ [
+ "▁j",
+ "Query"
+ ],
+ [
+ "▁",
+ "jQuery"
+ ],
+ [
+ "▁corpor",
+ "ate"
+ ],
+ [
+ "▁BAS",
+ "IS"
+ ],
+ [
+ "▁appoint",
+ "ed"
+ ],
+ [
+ "▁em",
+ "bar"
+ ],
+ [
+ "▁emb",
+ "ar"
+ ],
+ [
+ "ograph",
+ "ie"
+ ],
+ [
+ "▁p",
+ "ressed"
+ ],
+ [
+ "▁pr",
+ "essed"
+ ],
+ [
+ "▁pres",
+ "sed"
+ ],
+ [
+ "▁press",
+ "ed"
+ ],
+ [
+ "▁",
+ "pressed"
+ ],
+ [
+ "▁ch",
+ "ampion"
+ ],
+ [
+ "▁champ",
+ "ion"
+ ],
+ [
+ "em",
+ "it"
+ ],
+ [
+ "emi",
+ "t"
+ ],
+ [
+ "e",
+ "mit"
+ ],
+ [
+ "▁B",
+ "ed"
+ ],
+ [
+ "▁Be",
+ "d"
+ ],
+ [
+ "ва",
+ "ння"
+ ],
+ [
+ "ван",
+ "ня"
+ ],
+ [
+ "Gu",
+ "i"
+ ],
+ [
+ "G",
+ "ui"
+ ],
+ [
+ "▁P",
+ "UR"
+ ],
+ [
+ "▁ur",
+ "ban"
+ ],
+ [
+ "▁urb",
+ "an"
+ ],
+ [
+ "▁sent",
+ "ence"
+ ],
+ [
+ "bu",
+ "ry"
+ ],
+ [
+ "bur",
+ "y"
+ ],
+ [
+ "b",
+ "ury"
+ ],
+ [
+ "▁V",
+ "ideo"
+ ],
+ [
+ "▁",
+ "Video"
+ ],
+ [
+ "▁regular",
+ "ly"
+ ],
+ [
+ "▁regul",
+ "arly"
+ ],
+ [
+ "v",
+ "l"
+ ],
+ [
+ "▁с",
+ "лу"
+ ],
+ [
+ "▁",
+ "слу"
+ ],
+ [
+ "oc",
+ "key"
+ ],
+ [
+ "ock",
+ "ey"
+ ],
+ [
+ "ev",
+ "in"
+ ],
+ [
+ "e",
+ "vin"
+ ],
+ [
+ "ult",
+ "ural"
+ ],
+ [
+ "ultur",
+ "al"
+ ],
+ [
+ "▁pass",
+ "age"
+ ],
+ [
+ "▁со",
+ "став"
+ ],
+ [
+ "▁соста",
+ "в"
+ ],
+ [
+ "▁large",
+ "ly"
+ ],
+ [
+ "▁larg",
+ "ely"
+ ],
+ [
+ "or",
+ "ters"
+ ],
+ [
+ "ort",
+ "ers"
+ ],
+ [
+ "orter",
+ "s"
+ ],
+ [
+ "orte",
+ "rs"
+ ],
+ [
+ "▁conne",
+ "ctions"
+ ],
+ [
+ "▁connection",
+ "s"
+ ],
+ [
+ "▁connect",
+ "ions"
+ ],
+ [
+ "▁surpr",
+ "ising"
+ ],
+ [
+ "b",
+ "c"
+ ],
+ [
+ "▁strong",
+ "ly"
+ ],
+ [
+ "ans",
+ "as"
+ ],
+ [
+ "▁s",
+ "ist"
+ ],
+ [
+ "▁si",
+ "st"
+ ],
+ [
+ "▁ext",
+ "reme"
+ ],
+ [
+ "▁extrem",
+ "e"
+ ],
+ [
+ "▁extr",
+ "eme"
+ ],
+ [
+ "wh",
+ "el"
+ ],
+ [
+ "whe",
+ "l"
+ ],
+ [
+ "w",
+ "hel"
+ ],
+ [
+ "▁de",
+ "aling"
+ ],
+ [
+ "▁deal",
+ "ing"
+ ],
+ [
+ "ograph",
+ "ic"
+ ],
+ [
+ "▁Republic",
+ "an"
+ ],
+ [
+ "▁gr",
+ "anted"
+ ],
+ [
+ "▁gran",
+ "ted"
+ ],
+ [
+ "▁grant",
+ "ed"
+ ],
+ [
+ "▁C",
+ "L"
+ ],
+ [
+ "▁",
+ "CL"
+ ],
+ [
+ "▁H",
+ "ope"
+ ],
+ [
+ "▁Ho",
+ "pe"
+ ],
+ [
+ "▁Hop",
+ "e"
+ ],
+ [
+ "less",
+ "ly"
+ ],
+ [
+ "▁u",
+ "pload"
+ ],
+ [
+ "▁up",
+ "load"
+ ],
+ [
+ "▁",
+ "upload"
+ ],
+ [
+ "▁-",
+ "\\"
+ ],
+ [
+ "▁",
+ "-\\"
+ ],
+ [
+ "ни",
+ "ю"
+ ],
+ [
+ "▁val",
+ "uable"
+ ],
+ [
+ "=",
+ "["
+ ],
+ [
+ "Pr",
+ "ice"
+ ],
+ [
+ "P",
+ "rice"
+ ],
+ [
+ "iss",
+ "ance"
+ ],
+ [
+ "ie",
+ "ns"
+ ],
+ [
+ "ien",
+ "s"
+ ],
+ [
+ "i",
+ "ens"
+ ],
+ [
+ "he",
+ "it"
+ ],
+ [
+ "▁sugg",
+ "ests"
+ ],
+ [
+ "▁suggest",
+ "s"
+ ],
+ [
+ "с",
+ "ло"
+ ],
+ [
+ "▁j",
+ "ur"
+ ],
+ [
+ "▁ju",
+ "r"
+ ],
+ [
+ "}",
+ "|"
+ ],
+ [
+ "l",
+ "p"
+ ],
+ [
+ "▁inv",
+ "ited"
+ ],
+ [
+ "▁invite",
+ "d"
+ ],
+ [
+ "▁de",
+ "riv"
+ ],
+ [
+ "▁der",
+ "iv"
+ ],
+ [
+ "IM",
+ "IT"
+ ],
+ [
+ "I",
+ "MIT"
+ ],
+ [
+ "ra",
+ "ss"
+ ],
+ [
+ "ras",
+ "s"
+ ],
+ [
+ "r",
+ "ass"
+ ],
+ [
+ "▁in",
+ "struct"
+ ],
+ [
+ "▁inst",
+ "ruct"
+ ],
+ [
+ "▁instr",
+ "uct"
+ ],
+ [
+ "▁c",
+ "ourses"
+ ],
+ [
+ "▁cour",
+ "ses"
+ ],
+ [
+ "▁course",
+ "s"
+ ],
+ [
+ "▁cours",
+ "es"
+ ],
+ [
+ "ä",
+ "ch"
+ ],
+ [
+ "▁fif",
+ "ty"
+ ],
+ [
+ "▁fi",
+ "fty"
+ ],
+ [
+ "DE",
+ "VICE"
+ ],
+ [
+ "DEV",
+ "ICE"
+ ],
+ [
+ "AS",
+ "H"
+ ],
+ [
+ "A",
+ "SH"
+ ],
+ [
+ "▁h",
+ "ip"
+ ],
+ [
+ "▁hi",
+ "p"
+ ],
+ [
+ "▁",
+ "hip"
+ ],
+ [
+ "Un",
+ "known"
+ ],
+ [
+ "▁C",
+ "atalogue"
+ ],
+ [
+ "▁Catal",
+ "ogue"
+ ],
+ [
+ "▁R",
+ "oll"
+ ],
+ [
+ "▁Ro",
+ "ll"
+ ],
+ [
+ "▁Rol",
+ "l"
+ ],
+ [
+ "▁",
+ "Roll"
+ ],
+ [
+ "▁t",
+ "ensor"
+ ],
+ [
+ "▁ten",
+ "sor"
+ ],
+ [
+ "▁tens",
+ "or"
+ ],
+ [
+ "▁",
+ "tensor"
+ ],
+ [
+ "be",
+ "c"
+ ],
+ [
+ "b",
+ "ec"
+ ],
+ [
+ "ét",
+ "é"
+ ],
+ [
+ "é",
+ "té"
+ ],
+ [
+ "Id",
+ "entity"
+ ],
+ [
+ "Ident",
+ "ity"
+ ],
+ [
+ "&",
+ "\\"
+ ],
+ [
+ "▁Step",
+ "hen"
+ ],
+ [
+ "▁Steph",
+ "en"
+ ],
+ [
+ "no",
+ "des"
+ ],
+ [
+ "node",
+ "s"
+ ],
+ [
+ "nod",
+ "es"
+ ],
+ [
+ "n",
+ "odes"
+ ],
+ [
+ "Di",
+ "m"
+ ],
+ [
+ "D",
+ "im"
+ ],
+ [
+ "▁cons",
+ "ists"
+ ],
+ [
+ "▁consist",
+ "s"
+ ],
+ [
+ "▁normal",
+ "ly"
+ ],
+ [
+ "▁norm",
+ "ally"
+ ],
+ [
+ "ub",
+ "l"
+ ],
+ [
+ "u",
+ "bl"
+ ],
+ [
+ "▁Pol",
+ "ice"
+ ],
+ [
+ "▁G",
+ "ames"
+ ],
+ [
+ "▁Game",
+ "s"
+ ],
+ [
+ "▁Ga",
+ "mes"
+ ],
+ [
+ "▁Gam",
+ "es"
+ ],
+ [
+ "fi",
+ "ve"
+ ],
+ [
+ "f",
+ "ive"
+ ],
+ [
+ "Ha",
+ "ve"
+ ],
+ [
+ "H",
+ "ave"
+ ],
+ [
+ "▁p",
+ "adding"
+ ],
+ [
+ "▁pad",
+ "ding"
+ ],
+ [
+ "▁",
+ "padding"
+ ],
+ [
+ "er",
+ "es"
+ ],
+ [
+ "ere",
+ "s"
+ ],
+ [
+ "e",
+ "res"
+ ],
+ [
+ "an",
+ "th"
+ ],
+ [
+ "ant",
+ "h"
+ ],
+ [
+ "▁p",
+ "uts"
+ ],
+ [
+ "▁put",
+ "s"
+ ],
+ [
+ "▁pu",
+ "ts"
+ ],
+ [
+ "um",
+ "inate"
+ ],
+ [
+ "umin",
+ "ate"
+ ],
+ [
+ "umi",
+ "nate"
+ ],
+ [
+ "ov",
+ "ie"
+ ],
+ [
+ "ovi",
+ "e"
+ ],
+ [
+ "▁In",
+ "dex"
+ ],
+ [
+ "▁Ind",
+ "ex"
+ ],
+ [
+ "▁",
+ "Index"
+ ],
+ [
+ "bl",
+ "ue"
+ ],
+ [
+ "Sc",
+ "al"
+ ],
+ [
+ "S",
+ "cal"
+ ],
+ [
+ "▁g",
+ "iant"
+ ],
+ [
+ "▁gi",
+ "ant"
+ ],
+ [
+ "T",
+ "F"
+ ],
+ [
+ "ps",
+ "on"
+ ],
+ [
+ "p",
+ "son"
+ ],
+ [
+ "▁vict",
+ "im"
+ ],
+ [
+ "▁vic",
+ "tim"
+ ],
+ [
+ "se",
+ "rial"
+ ],
+ [
+ "ser",
+ "ial"
+ ],
+ [
+ "s",
+ "erial"
+ ],
+ [
+ "▁S",
+ "ym"
+ ],
+ [
+ "▁Sy",
+ "m"
+ ],
+ [
+ "▁",
+ "Sym"
+ ],
+ [
+ "Sing",
+ "le"
+ ],
+ [
+ "S",
+ "ingle"
+ ],
+ [
+ "▁m",
+ "d"
+ ],
+ [
+ "▁",
+ "md"
+ ],
+ [
+ "▁att",
+ "ended"
+ ],
+ [
+ "▁attend",
+ "ed"
+ ],
+ [
+ "▁S",
+ "tra"
+ ],
+ [
+ "▁St",
+ "ra"
+ ],
+ [
+ "▁Str",
+ "a"
+ ],
+ [
+ "▁D",
+ "ark"
+ ],
+ [
+ "▁Dar",
+ "k"
+ ],
+ [
+ "▁",
+ "Dark"
+ ],
+ [
+ ")",
+ "|"
+ ],
+ [
+ "▁s",
+ "pan"
+ ],
+ [
+ "▁sp",
+ "an"
+ ],
+ [
+ "▁",
+ "span"
+ ],
+ [
+ "▁main",
+ "tenance"
+ ],
+ [
+ "▁b",
+ "ind"
+ ],
+ [
+ "▁bi",
+ "nd"
+ ],
+ [
+ "▁bin",
+ "d"
+ ],
+ [
+ "▁",
+ "bind"
+ ],
+ [
+ "Be",
+ "an"
+ ],
+ [
+ "il",
+ "arly"
+ ],
+ [
+ "ilar",
+ "ly"
+ ],
+ [
+ "▁con",
+ "vent"
+ ],
+ [
+ "▁conv",
+ "ent"
+ ],
+ [
+ "▁conven",
+ "t"
+ ],
+ [
+ "▁conve",
+ "nt"
+ ],
+ [
+ "▁Jos",
+ "é"
+ ],
+ [
+ "ud",
+ "d"
+ ],
+ [
+ "u",
+ "dd"
+ ],
+ [
+ "▁p",
+ "oly"
+ ],
+ [
+ "▁pol",
+ "y"
+ ],
+ [
+ "▁po",
+ "ly"
+ ],
+ [
+ "▁",
+ "poly"
+ ],
+ [
+ "▁i",
+ "dx"
+ ],
+ [
+ "▁id",
+ "x"
+ ],
+ [
+ "▁",
+ "idx"
+ ],
+ [
+ "▁as",
+ "ks"
+ ],
+ [
+ "▁ask",
+ "s"
+ ],
+ [
+ "▁ent",
+ "hus"
+ ],
+ [
+ "▁s",
+ "uck"
+ ],
+ [
+ "▁su",
+ "ck"
+ ],
+ [
+ "▁suc",
+ "k"
+ ],
+ [
+ "▁C",
+ "ou"
+ ],
+ [
+ "▁Co",
+ "u"
+ ],
+ [
+ "▁Corpor",
+ "ation"
+ ],
+ [
+ "us",
+ "ions"
+ ],
+ [
+ "usion",
+ "s"
+ ],
+ [
+ "op",
+ "her"
+ ],
+ [
+ "oph",
+ "er"
+ ],
+ [
+ "o",
+ "pher"
+ ],
+ [
+ "▁sympt",
+ "oms"
+ ],
+ [
+ "▁Joh",
+ "ann"
+ ],
+ [
+ "▁п",
+ "у"
+ ],
+ [
+ "▁",
+ "пу"
+ ],
+ [
+ "▁h",
+ "tml"
+ ],
+ [
+ "▁",
+ "html"
+ ],
+ [
+ "▁p",
+ "s"
+ ],
+ [
+ "▁",
+ "ps"
+ ],
+ [
+ "ear",
+ "ing"
+ ],
+ [
+ "ea",
+ "ring"
+ ],
+ [
+ "e",
+ "aring"
+ ],
+ [
+ "ge",
+ "sch"
+ ],
+ [
+ "ges",
+ "ch"
+ ],
+ [
+ "g",
+ "esch"
+ ],
+ [
+ "▁M",
+ "other"
+ ],
+ [
+ "▁Mo",
+ "ther"
+ ],
+ [
+ "▁Mot",
+ "her"
+ ],
+ [
+ "RE",
+ "T"
+ ],
+ [
+ "R",
+ "ET"
+ ],
+ [
+ "▁furn",
+ "iture"
+ ],
+ [
+ "P",
+ "F"
+ ],
+ [
+ "▁Gu",
+ "ard"
+ ],
+ [
+ "▁",
+ "Guard"
+ ],
+ [
+ "pat",
+ "tern"
+ ],
+ [
+ "▁love",
+ "ly"
+ ],
+ [
+ "▁lov",
+ "ely"
+ ],
+ [
+ "al",
+ "g"
+ ],
+ [
+ "a",
+ "lg"
+ ],
+ [
+ "ed",
+ "ly"
+ ],
+ [
+ "se",
+ "x"
+ ],
+ [
+ "s",
+ "ex"
+ ],
+ [
+ "▁fin",
+ "ds"
+ ],
+ [
+ "▁find",
+ "s"
+ ],
+ [
+ "Bu",
+ "f"
+ ],
+ [
+ "B",
+ "uf"
+ ],
+ [
+ "▁на",
+ "д"
+ ],
+ [
+ "▁",
+ "над"
+ ],
+ [
+ "▁к",
+ "м"
+ ],
+ [
+ "▁P",
+ "or"
+ ],
+ [
+ "▁Po",
+ "r"
+ ],
+ [
+ "С",
+ "Р"
+ ],
+ [
+ "En",
+ "ter"
+ ],
+ [
+ "Ent",
+ "er"
+ ],
+ [
+ "▁e",
+ "sta"
+ ],
+ [
+ "▁est",
+ "a"
+ ],
+ [
+ "▁es",
+ "ta"
+ ],
+ [
+ "▁",
+ "esta"
+ ],
+ [
+ "▁т",
+ "ре"
+ ],
+ [
+ "▁",
+ "тре"
+ ],
+ [
+ "▁\"",
+ "*"
+ ],
+ [
+ "▁F",
+ "ox"
+ ],
+ [
+ "▁Fo",
+ "x"
+ ],
+ [
+ "▁c",
+ "ock"
+ ],
+ [
+ "▁co",
+ "ck"
+ ],
+ [
+ "▁coc",
+ "k"
+ ],
+ [
+ "▁",
+ "cock"
+ ],
+ [
+ "B",
+ "undle"
+ ],
+ [
+ "▁p",
+ "uis"
+ ],
+ [
+ "▁pu",
+ "is"
+ ],
+ [
+ "▁",
+ "puis"
+ ],
+ [
+ "▁ann",
+ "ounce"
+ ],
+ [
+ "▁announ",
+ "ce"
+ ],
+ [
+ "▁g",
+ "uid"
+ ],
+ [
+ "▁gu",
+ "id"
+ ],
+ [
+ "▁",
+ "guid"
+ ],
+ [
+ "check",
+ "ed"
+ ],
+ [
+ "ic",
+ "ide"
+ ],
+ [
+ "ici",
+ "de"
+ ],
+ [
+ "ne",
+ "g"
+ ],
+ [
+ "n",
+ "eg"
+ ],
+ [
+ "▁G",
+ "il"
+ ],
+ [
+ "▁Gi",
+ "l"
+ ],
+ [
+ "sc",
+ "hen"
+ ],
+ [
+ "sch",
+ "en"
+ ],
+ [
+ "sche",
+ "n"
+ ],
+ [
+ "s",
+ "chen"
+ ],
+ [
+ "olog",
+ "ist"
+ ],
+ [
+ "is",
+ "o"
+ ],
+ [
+ "i",
+ "so"
+ ],
+ [
+ "group",
+ "s"
+ ],
+ [
+ "gro",
+ "ups"
+ ],
+ [
+ "g",
+ "roups"
+ ],
+ [
+ "▁some",
+ "body"
+ ],
+ [
+ "Da",
+ "y"
+ ],
+ [
+ "D",
+ "ay"
+ ],
+ [
+ "tr",
+ "as"
+ ],
+ [
+ "tra",
+ "s"
+ ],
+ [
+ "t",
+ "ras"
+ ],
+ [
+ "▁comp",
+ "act"
+ ],
+ [
+ "▁organ",
+ "ized"
+ ],
+ [
+ "▁organiz",
+ "ed"
+ ],
+ [
+ "▁organize",
+ "d"
+ ],
+ [
+ "▁r",
+ "oles"
+ ],
+ [
+ "▁ro",
+ "les"
+ ],
+ [
+ "▁role",
+ "s"
+ ],
+ [
+ "▁h",
+ "int"
+ ],
+ [
+ "▁hi",
+ "nt"
+ ],
+ [
+ "▁",
+ "hint"
+ ],
+ [
+ "▁s",
+ "å"
+ ],
+ [
+ "▁p",
+ "ays"
+ ],
+ [
+ "▁pay",
+ "s"
+ ],
+ [
+ "▁pa",
+ "ys"
+ ],
+ [
+ "▁С",
+ "и"
+ ],
+ [
+ "▁h",
+ "oped"
+ ],
+ [
+ "▁hope",
+ "d"
+ ],
+ [
+ "▁hop",
+ "ed"
+ ],
+ [
+ "▁ho",
+ "ped"
+ ],
+ [
+ "▁s",
+ "ail"
+ ],
+ [
+ "▁sa",
+ "il"
+ ],
+ [
+ "▁V",
+ "ers"
+ ],
+ [
+ "▁Ver",
+ "s"
+ ],
+ [
+ "▁Ve",
+ "rs"
+ ],
+ [
+ "▁",
+ "Vers"
+ ],
+ [
+ "▁em",
+ "br"
+ ],
+ [
+ "▁emb",
+ "r"
+ ],
+ [
+ "▁b",
+ "ot"
+ ],
+ [
+ "▁bo",
+ "t"
+ ],
+ [
+ "▁",
+ "bot"
+ ],
+ [
+ "▁ex",
+ "ceed"
+ ],
+ [
+ "▁exc",
+ "eed"
+ ],
+ [
+ "BA",
+ "CK"
+ ],
+ [
+ "B",
+ "ACK"
+ ],
+ [
+ "▁g",
+ "aze"
+ ],
+ [
+ "▁gaz",
+ "e"
+ ],
+ [
+ "▁ga",
+ "ze"
+ ],
+ [
+ "▁s",
+ "pons"
+ ],
+ [
+ "▁sp",
+ "ons"
+ ],
+ [
+ "▁spo",
+ "ns"
+ ],
+ [
+ "AS",
+ "T"
+ ],
+ [
+ "A",
+ "ST"
+ ],
+ [
+ "▁tor",
+ "ch"
+ ],
+ [
+ "▁",
+ "torch"
+ ],
+ [
+ "▁news",
+ "paper"
+ ],
+ [
+ "▁newsp",
+ "aper"
+ ],
+ [
+ "▁D",
+ "ist"
+ ],
+ [
+ "▁Dis",
+ "t"
+ ],
+ [
+ "▁Di",
+ "st"
+ ],
+ [
+ "▁",
+ "Dist"
+ ],
+ [
+ "▁b",
+ "ass"
+ ],
+ [
+ "▁bas",
+ "s"
+ ],
+ [
+ "▁ba",
+ "ss"
+ ],
+ [
+ "▁h",
+ "anging"
+ ],
+ [
+ "▁han",
+ "ging"
+ ],
+ [
+ "▁hang",
+ "ing"
+ ],
+ [
+ "▁e",
+ "ars"
+ ],
+ [
+ "▁ear",
+ "s"
+ ],
+ [
+ "▁",
+ "ears"
+ ],
+ [
+ "ń",
+ "sk"
+ ],
+ [
+ "get",
+ "Value"
+ ],
+ [
+ "▁un",
+ "us"
+ ],
+ [
+ "▁E",
+ "le"
+ ],
+ [
+ "▁El",
+ "e"
+ ],
+ [
+ "serv",
+ "ices"
+ ],
+ [
+ "service",
+ "s"
+ ],
+ [
+ "s",
+ "ervices"
+ ],
+ [
+ "▁d",
+ "ressed"
+ ],
+ [
+ "▁dr",
+ "essed"
+ ],
+ [
+ "▁dress",
+ "ed"
+ ],
+ [
+ "la",
+ "v"
+ ],
+ [
+ "l",
+ "av"
+ ],
+ [
+ "▁п",
+ "ла"
+ ],
+ [
+ "▁",
+ "пла"
+ ],
+ [
+ "Priv",
+ "ate"
+ ],
+ [
+ "P",
+ "rivate"
+ ],
+ [
+ "mi",
+ "c"
+ ],
+ [
+ "m",
+ "ic"
+ ],
+ [
+ "▁par",
+ "ser"
+ ],
+ [
+ "▁parse",
+ "r"
+ ],
+ [
+ "▁",
+ "parser"
+ ],
+ [
+ "▁se",
+ "ctions"
+ ],
+ [
+ "▁section",
+ "s"
+ ],
+ [
+ "▁sect",
+ "ions"
+ ],
+ [
+ "▁",
+ "sections"
+ ],
+ [
+ "▁f",
+ "o"
+ ],
+ [
+ "▁",
+ "fo"
+ ],
+ [
+ "Err",
+ "orf"
+ ],
+ [
+ "Error",
+ "f"
+ ],
+ [
+ "in",
+ "z"
+ ],
+ [
+ "ör",
+ "d"
+ ],
+ [
+ "ö",
+ "rd"
+ ],
+ [
+ "▁m",
+ "etric"
+ ],
+ [
+ "▁met",
+ "ric"
+ ],
+ [
+ "▁",
+ "metric"
+ ],
+ [
+ "UR",
+ "I"
+ ],
+ [
+ "U",
+ "RI"
+ ],
+ [
+ "▁v",
+ "ice"
+ ],
+ [
+ "▁vi",
+ "ce"
+ ],
+ [
+ "▁vic",
+ "e"
+ ],
+ [
+ "RE",
+ "D"
+ ],
+ [
+ "R",
+ "ED"
+ ],
+ [
+ "▁n",
+ "ue"
+ ],
+ [
+ "▁nu",
+ "e"
+ ],
+ [
+ "re",
+ "vs"
+ ],
+ [
+ "rev",
+ "s"
+ ],
+ [
+ "▁col",
+ "lected"
+ ],
+ [
+ "▁collect",
+ "ed"
+ ],
+ [
+ "▁colle",
+ "cted"
+ ],
+ [
+ "oo",
+ "se"
+ ],
+ [
+ "o",
+ "ose"
+ ],
+ [
+ "▁m",
+ "ond"
+ ],
+ [
+ "▁mon",
+ "d"
+ ],
+ [
+ "▁mo",
+ "nd"
+ ],
+ [
+ "▁",
+ "mond"
+ ],
+ [
+ "▁n",
+ "as"
+ ],
+ [
+ "▁na",
+ "s"
+ ],
+ [
+ "▁",
+ "nas"
+ ],
+ [
+ "▁На",
+ "се"
+ ],
+ [
+ "▁",
+ "å"
+ ],
+ [
+ "Dr",
+ "op"
+ ],
+ [
+ "D",
+ "rop"
+ ],
+ [
+ "▁ab",
+ "use"
+ ],
+ [
+ "▁s",
+ "ees"
+ ],
+ [
+ "▁se",
+ "es"
+ ],
+ [
+ "▁see",
+ "s"
+ ],
+ [
+ "▁H",
+ "ence"
+ ],
+ [
+ "▁Hen",
+ "ce"
+ ],
+ [
+ "ex",
+ "ec"
+ ],
+ [
+ "}\\",
+ ","
+ ],
+ [
+ "}",
+ "\\,"
+ ],
+ [
+ "▁ar",
+ "bitr"
+ ],
+ [
+ "▁Ap",
+ "plication"
+ ],
+ [
+ "▁",
+ "Application"
+ ],
+ [
+ "f",
+ "amily"
+ ],
+ [
+ "ü",
+ "d"
+ ],
+ [
+ "▁mag",
+ "netic"
+ ],
+ [
+ "▁magn",
+ "etic"
+ ],
+ [
+ "▁magnet",
+ "ic"
+ ],
+ [
+ "▁new",
+ "ly"
+ ],
+ [
+ "▁re",
+ "produ"
+ ],
+ [
+ "▁rep",
+ "rodu"
+ ],
+ [
+ "▁writ",
+ "ers"
+ ],
+ [
+ "▁write",
+ "rs"
+ ],
+ [
+ "▁writer",
+ "s"
+ ],
+ [
+ "▁he",
+ "aders"
+ ],
+ [
+ "▁head",
+ "ers"
+ ],
+ [
+ "▁header",
+ "s"
+ ],
+ [
+ "▁",
+ "headers"
+ ],
+ [
+ "š",
+ "í"
+ ],
+ [
+ "р",
+ "т"
+ ],
+ [
+ "YP",
+ "E"
+ ],
+ [
+ "Y",
+ "PE"
+ ],
+ [
+ "▁s",
+ "chema"
+ ],
+ [
+ "▁sch",
+ "ema"
+ ],
+ [
+ "▁sche",
+ "ma"
+ ],
+ [
+ "▁",
+ "schema"
+ ],
+ [
+ "▁C",
+ "e"
+ ],
+ [
+ "▁Je",
+ "ws"
+ ],
+ [
+ "▁Jew",
+ "s"
+ ],
+ [
+ "▁Re",
+ "cord"
+ ],
+ [
+ "▁Rec",
+ "ord"
+ ],
+ [
+ "▁",
+ "Record"
+ ],
+ [
+ "pre",
+ "sent"
+ ],
+ [
+ "pres",
+ "ent"
+ ],
+ [
+ "p",
+ "resent"
+ ],
+ [
+ "▁так",
+ "же"
+ ],
+ [
+ "▁label",
+ "s"
+ ],
+ [
+ "▁lab",
+ "els"
+ ],
+ [
+ "▁",
+ "labels"
+ ],
+ [
+ "S",
+ "ocket"
+ ],
+ [
+ "▁equ",
+ "ations"
+ ],
+ [
+ "▁equation",
+ "s"
+ ],
+ [
+ "▁eq",
+ "uations"
+ ],
+ [
+ "▁medic",
+ "ine"
+ ],
+ [
+ "▁author",
+ "ities"
+ ],
+ [
+ "}",
+ "`"
+ ],
+ [
+ "ст",
+ "ви"
+ ],
+ [
+ "ств",
+ "и"
+ ],
+ [
+ "▁C",
+ "orn"
+ ],
+ [
+ "▁Co",
+ "rn"
+ ],
+ [
+ "▁Cor",
+ "n"
+ ],
+ [
+ "▁environment",
+ "al"
+ ],
+ [
+ "WAR",
+ "E"
+ ],
+ [
+ "WA",
+ "RE"
+ ],
+ [
+ "W",
+ "ARE"
+ ],
+ [
+ "Me",
+ "r"
+ ],
+ [
+ "M",
+ "er"
+ ],
+ [
+ "▁са",
+ "мо"
+ ],
+ [
+ "▁Techn",
+ "ology"
+ ],
+ [
+ "▁S",
+ "af"
+ ],
+ [
+ "▁Sa",
+ "f"
+ ],
+ [
+ "▁con",
+ "n"
+ ],
+ [
+ "▁co",
+ "nn"
+ ],
+ [
+ "▁",
+ "conn"
+ ],
+ [
+ "▁U",
+ "m"
+ ],
+ [
+ "▁Pac",
+ "ific"
+ ],
+ [
+ "те",
+ "л"
+ ],
+ [
+ "ja",
+ "n"
+ ],
+ [
+ "j",
+ "an"
+ ],
+ [
+ "▁unc",
+ "ertain"
+ ],
+ [
+ "▁bel",
+ "ief"
+ ],
+ [
+ "▁belie",
+ "f"
+ ],
+ [
+ "co",
+ "unter"
+ ],
+ [
+ "count",
+ "er"
+ ],
+ [
+ "c",
+ "ounter"
+ ],
+ [
+ "to",
+ "Be"
+ ],
+ [
+ "IN",
+ "S"
+ ],
+ [
+ "I",
+ "NS"
+ ],
+ [
+ "we",
+ "et"
+ ],
+ [
+ "Li",
+ "ght"
+ ],
+ [
+ "L",
+ "ight"
+ ],
+ [
+ "pr",
+ "imary"
+ ],
+ [
+ "prim",
+ "ary"
+ ],
+ [
+ "▁feature",
+ "d"
+ ],
+ [
+ "▁feat",
+ "ured"
+ ],
+ [
+ "▁touch",
+ "ed"
+ ],
+ [
+ "▁tou",
+ "ched"
+ ],
+ [
+ "HT",
+ "TP"
+ ],
+ [
+ "▁t",
+ "act"
+ ],
+ [
+ "▁ta",
+ "ct"
+ ],
+ [
+ "pos",
+ "itory"
+ ],
+ [
+ "p",
+ "ository"
+ ],
+ [
+ "▁e",
+ "ines"
+ ],
+ [
+ "▁ein",
+ "es"
+ ],
+ [
+ "▁eine",
+ "s"
+ ],
+ [
+ "la",
+ "ss"
+ ],
+ [
+ "las",
+ "s"
+ ],
+ [
+ "l",
+ "ass"
+ ],
+ [
+ "сь",
+ "ка"
+ ],
+ [
+ "▁prz",
+ "ez"
+ ],
+ [
+ "▁prze",
+ "z"
+ ],
+ [
+ "▁f",
+ "uer"
+ ],
+ [
+ "▁fue",
+ "r"
+ ],
+ [
+ "▁fu",
+ "er"
+ ],
+ [
+ "▁exc",
+ "iting"
+ ],
+ [
+ "▁excit",
+ "ing"
+ ],
+ [
+ "▁C",
+ "ub"
+ ],
+ [
+ "▁Cu",
+ "b"
+ ],
+ [
+ "ag",
+ "an"
+ ],
+ [
+ "aga",
+ "n"
+ ],
+ [
+ "a",
+ "gan"
+ ],
+ [
+ "V",
+ "O"
+ ],
+ [
+ "▁'",
+ "%"
+ ],
+ [
+ "▁\\",
+ "{"
+ ],
+ [
+ "▁",
+ "\\{"
+ ],
+ [
+ "ub",
+ "ble"
+ ],
+ [
+ "▁F",
+ "ol"
+ ],
+ [
+ "▁Fo",
+ "l"
+ ],
+ [
+ "▁K",
+ "ong"
+ ],
+ [
+ "▁Kon",
+ "g"
+ ],
+ [
+ "▁Ko",
+ "ng"
+ ],
+ [
+ "▁ver",
+ "sch"
+ ],
+ [
+ "▁vers",
+ "ch"
+ ],
+ [
+ "FA",
+ "IL"
+ ],
+ [
+ "F",
+ "AIL"
+ ],
+ [
+ "▁na",
+ "ar"
+ ],
+ [
+ "ö",
+ "s"
+ ],
+ [
+ "sp",
+ "eed"
+ ],
+ [
+ "spe",
+ "ed"
+ ],
+ [
+ "s",
+ "peed"
+ ],
+ [
+ "▁terr",
+ "itor"
+ ],
+ [
+ "▁territo",
+ "r"
+ ],
+ [
+ "▁w",
+ "rap"
+ ],
+ [
+ "▁wr",
+ "ap"
+ ],
+ [
+ "▁",
+ "wrap"
+ ],
+ [
+ "▁Jah",
+ "re"
+ ],
+ [
+ "▁Jahr",
+ "e"
+ ],
+ [
+ "▁Ja",
+ "hre"
+ ],
+ [
+ "le",
+ "e"
+ ],
+ [
+ "l",
+ "ee"
+ ],
+ [
+ "▁cross",
+ "ed"
+ ],
+ [
+ "res",
+ "olve"
+ ],
+ [
+ "▁s",
+ "tim"
+ ],
+ [
+ "▁st",
+ "im"
+ ],
+ [
+ "N",
+ "ative"
+ ],
+ [
+ "ur",
+ "sor"
+ ],
+ [
+ "urs",
+ "or"
+ ],
+ [
+ "Not",
+ "Null"
+ ],
+ [
+ "▁Al",
+ "bert"
+ ],
+ [
+ "▁Alber",
+ "t"
+ ],
+ [
+ "▁Alb",
+ "ert"
+ ],
+ [
+ "▁sign",
+ "ature"
+ ],
+ [
+ "▁",
+ "signature"
+ ],
+ [
+ "▁R",
+ "u"
+ ],
+ [
+ "id",
+ "as"
+ ],
+ [
+ "ida",
+ "s"
+ ],
+ [
+ "i",
+ "das"
+ ],
+ [
+ "▁de",
+ "cent"
+ ],
+ [
+ "▁dec",
+ "ent"
+ ],
+ [
+ "▁dece",
+ "nt"
+ ],
+ [
+ "▁f",
+ "aced"
+ ],
+ [
+ "▁face",
+ "d"
+ ],
+ [
+ "▁fac",
+ "ed"
+ ],
+ [
+ "▁fa",
+ "ced"
+ ],
+ [
+ "▁",
+ "faced"
+ ],
+ [
+ "▁",
+ "лю"
+ ],
+ [
+ "▁Sp",
+ "ain"
+ ],
+ [
+ "▁res",
+ "istance"
+ ],
+ [
+ "▁resist",
+ "ance"
+ ],
+ [
+ "▁B",
+ "rian"
+ ],
+ [
+ "▁Br",
+ "ian"
+ ],
+ [
+ "kw",
+ "args"
+ ],
+ [
+ "▁inter",
+ "val"
+ ],
+ [
+ "▁",
+ "interval"
+ ],
+ [
+ "▁Л",
+ "е"
+ ],
+ [
+ "▁ex",
+ "plo"
+ ],
+ [
+ "▁expl",
+ "o"
+ ],
+ [
+ "▁exp",
+ "lo"
+ ],
+ [
+ "▁s",
+ "emi"
+ ],
+ [
+ "▁se",
+ "mi"
+ ],
+ [
+ "▁sem",
+ "i"
+ ],
+ [
+ "▁wide",
+ "ly"
+ ],
+ [
+ "▁wid",
+ "ely"
+ ],
+ [
+ "d",
+ "x"
+ ],
+ [
+ "ko",
+ "v"
+ ],
+ [
+ "k",
+ "ov"
+ ],
+ [
+ "▁C",
+ "ome"
+ ],
+ [
+ "▁Com",
+ "e"
+ ],
+ [
+ "▁Co",
+ "me"
+ ],
+ [
+ "▁",
+ "Come"
+ ],
+ [
+ "▁kn",
+ "ife"
+ ],
+ [
+ "As",
+ "p"
+ ],
+ [
+ "A",
+ "sp"
+ ],
+ [
+ "un",
+ "o"
+ ],
+ [
+ "u",
+ "no"
+ ],
+ [
+ "line",
+ "to"
+ ],
+ [
+ "lin",
+ "eto"
+ ],
+ [
+ "▁B",
+ "und"
+ ],
+ [
+ "▁Bu",
+ "nd"
+ ],
+ [
+ "▁Bun",
+ "d"
+ ],
+ [
+ "C",
+ "ert"
+ ],
+ [
+ "▁t",
+ "odo"
+ ],
+ [
+ "▁to",
+ "do"
+ ],
+ [
+ "▁tod",
+ "o"
+ ],
+ [
+ "ta",
+ "gs"
+ ],
+ [
+ "tag",
+ "s"
+ ],
+ [
+ "t",
+ "ags"
+ ],
+ [
+ "▁guarante",
+ "e"
+ ],
+ [
+ "▁v",
+ "ital"
+ ],
+ [
+ "▁vi",
+ "tal"
+ ],
+ [
+ "▁vit",
+ "al"
+ ],
+ [
+ "▁vita",
+ "l"
+ ],
+ [
+ "▁f",
+ "ought"
+ ],
+ [
+ "▁fou",
+ "ght"
+ ],
+ [
+ "▁E",
+ "nv"
+ ],
+ [
+ "▁En",
+ "v"
+ ],
+ [
+ "▁",
+ "Env"
+ ],
+ [
+ "H",
+ "D"
+ ],
+ [
+ "Lo",
+ "wer"
+ ],
+ [
+ "Low",
+ "er"
+ ],
+ [
+ "L",
+ "ower"
+ ],
+ [
+ "T",
+ "x"
+ ],
+ [
+ "▁F",
+ "a"
+ ],
+ [
+ "▁ant",
+ "icip"
+ ],
+ [
+ "▁anti",
+ "cip"
+ ],
+ [
+ "Time",
+ "r"
+ ],
+ [
+ "Tim",
+ "er"
+ ],
+ [
+ "T",
+ "imer"
+ ],
+ [
+ "med",
+ "iate"
+ ],
+ [
+ "medi",
+ "ate"
+ ],
+ [
+ "media",
+ "te"
+ ],
+ [
+ "▁pro",
+ "ven"
+ ],
+ [
+ "▁pr",
+ "oven"
+ ],
+ [
+ "▁prov",
+ "en"
+ ],
+ [
+ "▁prove",
+ "n"
+ ],
+ [
+ "▁part",
+ "ir"
+ ],
+ [
+ "▁parti",
+ "r"
+ ],
+ [
+ "A",
+ "E"
+ ],
+ [
+ "cur",
+ "sor"
+ ],
+ [
+ "curs",
+ "or"
+ ],
+ [
+ "c",
+ "ursor"
+ ],
+ [
+ "▁wood",
+ "en"
+ ],
+ [
+ "▁wo",
+ "oden"
+ ],
+ [
+ "▁Cont",
+ "act"
+ ],
+ [
+ "▁",
+ "Contact"
+ ],
+ [
+ "re",
+ "gs"
+ ],
+ [
+ "reg",
+ "s"
+ ],
+ [
+ "▁prov",
+ "inc"
+ ],
+ [
+ "▁provin",
+ "c"
+ ],
+ [
+ "▁D",
+ "C"
+ ],
+ [
+ "▁",
+ "DC"
+ ],
+ [
+ "▁mem",
+ "ories"
+ ],
+ [
+ "▁memor",
+ "ies"
+ ],
+ [
+ "▁memo",
+ "ries"
+ ],
+ [
+ "▁f",
+ "t"
+ ],
+ [
+ "▁",
+ "ft"
+ ],
+ [
+ "▁b",
+ "attery"
+ ],
+ [
+ "▁batter",
+ "y"
+ ],
+ [
+ "▁batt",
+ "ery"
+ ],
+ [
+ "▁bat",
+ "tery"
+ ],
+ [
+ "ute",
+ "nant"
+ ],
+ [
+ "uten",
+ "ant"
+ ],
+ [
+ "u",
+ "tenant"
+ ],
+ [
+ "Log",
+ "in"
+ ],
+ [
+ "Lo",
+ "gin"
+ ],
+ [
+ "ount",
+ "ry"
+ ],
+ [
+ "oun",
+ "try"
+ ],
+ [
+ "▁comp",
+ "ens"
+ ],
+ [
+ "operator",
+ "name"
+ ],
+ [
+ "▁Jac",
+ "ob"
+ ],
+ [
+ "ze",
+ "d"
+ ],
+ [
+ "z",
+ "ed"
+ ],
+ [
+ "AD",
+ "DR"
+ ],
+ [
+ "ADD",
+ "R"
+ ],
+ [
+ "▁qu",
+ "ad"
+ ],
+ [
+ "▁",
+ "quad"
+ ],
+ [
+ "*)",
+ "."
+ ],
+ [
+ "*",
+ ")."
+ ],
+ [
+ "▁co",
+ "at"
+ ],
+ [
+ "▁f",
+ "ir"
+ ],
+ [
+ "▁fi",
+ "r"
+ ],
+ [
+ "▁Mich",
+ "el"
+ ],
+ [
+ "▁Mic",
+ "hel"
+ ],
+ [
+ "▁Mi",
+ "chel"
+ ],
+ [
+ "▁Miche",
+ "l"
+ ],
+ [
+ "▁Stand",
+ "ard"
+ ],
+ [
+ "▁",
+ "Standard"
+ ],
+ [
+ "r",
+ "f"
+ ],
+ [
+ "me",
+ "l"
+ ],
+ [
+ "m",
+ "el"
+ ],
+ [
+ "▁co",
+ "eff"
+ ],
+ [
+ "▁Ira",
+ "q"
+ ],
+ [
+ "▁G",
+ "iven"
+ ],
+ [
+ "▁Gi",
+ "ven"
+ ],
+ [
+ "▁Give",
+ "n"
+ ],
+ [
+ "ни",
+ "ма"
+ ],
+ [
+ "ним",
+ "а"
+ ],
+ [
+ "▁F",
+ "IT"
+ ],
+ [
+ "▁FI",
+ "T"
+ ],
+ [
+ "▁p",
+ "eu"
+ ],
+ [
+ "▁pe",
+ "u"
+ ],
+ [
+ "▁i",
+ "g"
+ ],
+ [
+ "▁",
+ "ig"
+ ],
+ [
+ "▁C",
+ "ase"
+ ],
+ [
+ "▁Cas",
+ "e"
+ ],
+ [
+ "▁Ca",
+ "se"
+ ],
+ [
+ "▁",
+ "Case"
+ ],
+ [
+ "m",
+ "é"
+ ],
+ [
+ "▁par",
+ "allel"
+ ],
+ [
+ "▁",
+ "parallel"
+ ],
+ [
+ "ci",
+ "o"
+ ],
+ [
+ "c",
+ "io"
+ ],
+ [
+ "ko",
+ "w"
+ ],
+ [
+ "k",
+ "ow"
+ ],
+ [
+ "▁institut",
+ "ions"
+ ],
+ [
+ "▁institution",
+ "s"
+ ],
+ [
+ "í",
+ "cul"
+ ],
+ [
+ "ab",
+ "an"
+ ],
+ [
+ "aba",
+ "n"
+ ],
+ [
+ "a",
+ "ban"
+ ],
+ [
+ "U",
+ "X"
+ ],
+ [
+ "▁Sa",
+ "rah"
+ ],
+ [
+ "▁Sar",
+ "ah"
+ ],
+ [
+ "▁Sara",
+ "h"
+ ],
+ [
+ "▁m",
+ "és"
+ ],
+ [
+ "▁mé",
+ "s"
+ ],
+ [
+ "▁at",
+ "mos"
+ ],
+ [
+ "▁atm",
+ "os"
+ ],
+ [
+ "▁slä",
+ "ktet"
+ ],
+ [
+ "▁br",
+ "others"
+ ],
+ [
+ "▁bro",
+ "thers"
+ ],
+ [
+ "▁brother",
+ "s"
+ ],
+ [
+ "▁want",
+ "ing"
+ ],
+ [
+ "aa",
+ "aa"
+ ],
+ [
+ "▁f",
+ "est"
+ ],
+ [
+ "▁fe",
+ "st"
+ ],
+ [
+ "=",
+ "-"
+ ],
+ [
+ "▁for",
+ "ty"
+ ],
+ [
+ "▁fort",
+ "y"
+ ],
+ [
+ "▁cre",
+ "ates"
+ ],
+ [
+ "▁create",
+ "s"
+ ],
+ [
+ "▁creat",
+ "es"
+ ],
+ [
+ "h",
+ "h"
+ ],
+ [
+ "▁And",
+ "roid"
+ ],
+ [
+ "▁Andr",
+ "oid"
+ ],
+ [
+ "▁",
+ "Android"
+ ],
+ [
+ "an",
+ "ches"
+ ],
+ [
+ "anc",
+ "hes"
+ ],
+ [
+ "anch",
+ "es"
+ ],
+ [
+ "anche",
+ "s"
+ ],
+ [
+ "B",
+ "T"
+ ],
+ [
+ "up",
+ "load"
+ ],
+ [
+ "u",
+ "pload"
+ ],
+ [
+ "xi",
+ "s"
+ ],
+ [
+ "x",
+ "is"
+ ],
+ [
+ "H",
+ "z"
+ ],
+ [
+ "бо",
+ "р"
+ ],
+ [
+ "б",
+ "ор"
+ ],
+ [
+ "RA",
+ "Y"
+ ],
+ [
+ "R",
+ "AY"
+ ],
+ [
+ "nt",
+ "il"
+ ],
+ [
+ "n",
+ "til"
+ ],
+ [
+ "▁le",
+ "aned"
+ ],
+ [
+ "▁lean",
+ "ed"
+ ],
+ [
+ "un",
+ "da"
+ ],
+ [
+ "und",
+ "a"
+ ],
+ [
+ "▁ult",
+ "imately"
+ ],
+ [
+ "▁ultimate",
+ "ly"
+ ],
+ [
+ "▁t",
+ "ok"
+ ],
+ [
+ "▁to",
+ "k"
+ ],
+ [
+ "▁",
+ "tok"
+ ],
+ [
+ "ne",
+ "h"
+ ],
+ [
+ "n",
+ "eh"
+ ],
+ [
+ "▁law",
+ "yer"
+ ],
+ [
+ "he",
+ "nd"
+ ],
+ [
+ "hen",
+ "d"
+ ],
+ [
+ "h",
+ "end"
+ ],
+ [
+ "▁V",
+ "in"
+ ],
+ [
+ "▁Vi",
+ "n"
+ ],
+ [
+ "▁fac",
+ "ility"
+ ],
+ [
+ "▁facil",
+ "ity"
+ ],
+ [
+ "▁l",
+ "ikes"
+ ],
+ [
+ "▁li",
+ "kes"
+ ],
+ [
+ "▁like",
+ "s"
+ ],
+ [
+ "▁lik",
+ "es"
+ ],
+ [
+ "en",
+ "to"
+ ],
+ [
+ "ent",
+ "o"
+ ],
+ [
+ "Node",
+ "s"
+ ],
+ [
+ "No",
+ "des"
+ ],
+ [
+ "N",
+ "odes"
+ ],
+ [
+ "▁entr",
+ "ance"
+ ],
+ [
+ "at",
+ "to"
+ ],
+ [
+ "att",
+ "o"
+ ],
+ [
+ "a",
+ "tto"
+ ],
+ [
+ "re",
+ "tt"
+ ],
+ [
+ "ret",
+ "t"
+ ],
+ [
+ "r",
+ "ett"
+ ],
+ [
+ "ac",
+ "cept"
+ ],
+ [
+ "th",
+ "eme"
+ ],
+ [
+ "the",
+ "me"
+ ],
+ [
+ "та",
+ "н"
+ ],
+ [
+ "т",
+ "ан"
+ ],
+ [
+ "os",
+ "i"
+ ],
+ [
+ "o",
+ "si"
+ ],
+ [
+ "▁{",
+ "},"
+ ],
+ [
+ "▁{}",
+ ","
+ ],
+ [
+ "▁",
+ "{},"
+ ],
+ [
+ "pgfpath",
+ "lineto"
+ ],
+ [
+ "go",
+ "od"
+ ],
+ [
+ "g",
+ "ood"
+ ],
+ [
+ "sl",
+ "ot"
+ ],
+ [
+ "s",
+ "lot"
+ ],
+ [
+ "▁in",
+ "noc"
+ ],
+ [
+ "▁inn",
+ "oc"
+ ],
+ [
+ "▁pro",
+ "port"
+ ],
+ [
+ "▁pr",
+ "oport"
+ ],
+ [
+ "▁prop",
+ "ort"
+ ],
+ [
+ "▁ar",
+ "rive"
+ ],
+ [
+ "▁arriv",
+ "e"
+ ],
+ [
+ "▁arr",
+ "ive"
+ ],
+ [
+ "é",
+ "ho"
+ ],
+ [
+ "▁p",
+ "airs"
+ ],
+ [
+ "▁pa",
+ "irs"
+ ],
+ [
+ "▁pair",
+ "s"
+ ],
+ [
+ "▁wr",
+ "apped"
+ ],
+ [
+ "▁wrap",
+ "ped"
+ ],
+ [
+ "▁un",
+ "w"
+ ],
+ [
+ "▁expl",
+ "os"
+ ],
+ [
+ "▁exp",
+ "los"
+ ],
+ [
+ "▁explo",
+ "s"
+ ],
+ [
+ "▁g",
+ "el"
+ ],
+ [
+ "▁ge",
+ "l"
+ ],
+ [
+ "▁",
+ "gel"
+ ],
+ [
+ "W",
+ "ill"
+ ],
+ [
+ "▁Ze",
+ "aland"
+ ],
+ [
+ "ía",
+ "s"
+ ],
+ [
+ "í",
+ "as"
+ ],
+ [
+ "▁J",
+ "r"
+ ],
+ [
+ "▁F",
+ "ra"
+ ],
+ [
+ "▁Fr",
+ "a"
+ ],
+ [
+ "▁le",
+ "git"
+ ],
+ [
+ "▁leg",
+ "it"
+ ],
+ [
+ "▁il",
+ "legal"
+ ],
+ [
+ "к",
+ "лю"
+ ],
+ [
+ "▁t",
+ "ort"
+ ],
+ [
+ "▁to",
+ "rt"
+ ],
+ [
+ "▁tor",
+ "t"
+ ],
+ [
+ "▁p",
+ "ron"
+ ],
+ [
+ "▁pro",
+ "n"
+ ],
+ [
+ "▁pr",
+ "on"
+ ],
+ [
+ "F",
+ "i"
+ ],
+ [
+ "▁f",
+ "org"
+ ],
+ [
+ "▁for",
+ "g"
+ ],
+ [
+ "▁fo",
+ "rg"
+ ],
+ [
+ "ex",
+ "port"
+ ],
+ [
+ "exp",
+ "ort"
+ ],
+ [
+ "▁Child",
+ "ren"
+ ],
+ [
+ "▁",
+ "Children"
+ ],
+ [
+ "▁A",
+ "bs"
+ ],
+ [
+ "▁Ab",
+ "s"
+ ],
+ [
+ "▁",
+ "Abs"
+ ],
+ [
+ "▁S",
+ "end"
+ ],
+ [
+ "▁Se",
+ "nd"
+ ],
+ [
+ "▁Sen",
+ "d"
+ ],
+ [
+ "▁",
+ "Send"
+ ],
+ [
+ "▁dis",
+ "count"
+ ],
+ [
+ "▁disc",
+ "ount"
+ ],
+ [
+ "▁disco",
+ "unt"
+ ],
+ [
+ "▁p",
+ "oster"
+ ],
+ [
+ "▁pos",
+ "ter"
+ ],
+ [
+ "▁po",
+ "ster"
+ ],
+ [
+ "▁post",
+ "er"
+ ],
+ [
+ "en",
+ "ted"
+ ],
+ [
+ "ent",
+ "ed"
+ ],
+ [
+ "ente",
+ "d"
+ ],
+ [
+ "an",
+ "im"
+ ],
+ [
+ "ani",
+ "m"
+ ],
+ [
+ "a",
+ "nim"
+ ],
+ [
+ "ve",
+ "rb"
+ ],
+ [
+ "ver",
+ "b"
+ ],
+ [
+ "st",
+ "o"
+ ],
+ [
+ "s",
+ "to"
+ ],
+ [
+ "▁B",
+ "ible"
+ ],
+ [
+ "▁Bi",
+ "ble"
+ ],
+ [
+ "pend",
+ "ing"
+ ],
+ [
+ "pen",
+ "ding"
+ ],
+ [
+ "p",
+ "ending"
+ ],
+ [
+ "▁P",
+ "hot"
+ ],
+ [
+ "▁Ph",
+ "ot"
+ ],
+ [
+ "st",
+ "rap"
+ ],
+ [
+ "str",
+ "ap"
+ ],
+ [
+ "stra",
+ "p"
+ ],
+ [
+ "ie",
+ "ron"
+ ],
+ [
+ "ier",
+ "on"
+ ],
+ [
+ "iero",
+ "n"
+ ],
+ [
+ "i",
+ "eron"
+ ],
+ [
+ "P",
+ "G"
+ ],
+ [
+ "cul",
+ "ar"
+ ],
+ [
+ "cu",
+ "lar"
+ ],
+ [
+ "c",
+ "ular"
+ ],
+ [
+ "cri",
+ "t"
+ ],
+ [
+ "cr",
+ "it"
+ ],
+ [
+ "c",
+ "rit"
+ ],
+ [
+ "ur",
+ "d"
+ ],
+ [
+ "u",
+ "rd"
+ ],
+ [
+ "EN",
+ "O"
+ ],
+ [
+ "E",
+ "NO"
+ ],
+ [
+ "▁nor",
+ "thern"
+ ],
+ [
+ "▁north",
+ "ern"
+ ],
+ [
+ "▁natural",
+ "ly"
+ ],
+ [
+ "▁natur",
+ "ally"
+ ],
+ [
+ "<",
+ "'"
+ ],
+ [
+ "we",
+ "g"
+ ],
+ [
+ "w",
+ "eg"
+ ],
+ [
+ "▁dr",
+ "unk"
+ ],
+ [
+ "▁D",
+ "al"
+ ],
+ [
+ "▁Da",
+ "l"
+ ],
+ [
+ "▁m",
+ "ouse"
+ ],
+ [
+ "▁mo",
+ "use"
+ ],
+ [
+ "▁mou",
+ "se"
+ ],
+ [
+ "▁",
+ "mouse"
+ ],
+ [
+ "▁contin",
+ "uous"
+ ],
+ [
+ "▁continu",
+ "ous"
+ ],
+ [
+ "▁init",
+ "ially"
+ ],
+ [
+ "▁initial",
+ "ly"
+ ],
+ [
+ "▁initi",
+ "ally"
+ ],
+ [
+ "ag",
+ "u"
+ ],
+ [
+ "a",
+ "gu"
+ ],
+ [
+ "м",
+ "пи"
+ ],
+ [
+ "AN",
+ "T"
+ ],
+ [
+ "A",
+ "NT"
+ ],
+ [
+ "Di",
+ "v"
+ ],
+ [
+ "D",
+ "iv"
+ ],
+ [
+ "▁rec",
+ "ording"
+ ],
+ [
+ "▁record",
+ "ing"
+ ],
+ [
+ "Bin",
+ "d"
+ ],
+ [
+ "Bi",
+ "nd"
+ ],
+ [
+ "B",
+ "ind"
+ ],
+ [
+ "▁correct",
+ "ly"
+ ],
+ [
+ "init",
+ "ial"
+ ],
+ [
+ "▁R",
+ "ights"
+ ],
+ [
+ "▁Right",
+ "s"
+ ],
+ [
+ "▁deb",
+ "ate"
+ ],
+ [
+ "WR",
+ "ITE"
+ ],
+ [
+ "bu",
+ "ilt"
+ ],
+ [
+ "▁per",
+ "mit"
+ ],
+ [
+ "▁perm",
+ "it"
+ ],
+ [
+ "▁professional",
+ "s"
+ ],
+ [
+ "▁profession",
+ "als"
+ ],
+ [
+ "c",
+ "v"
+ ],
+ [
+ "▁D",
+ "I"
+ ],
+ [
+ "▁",
+ "DI"
+ ],
+ [
+ "▁h",
+ "anded"
+ ],
+ [
+ "▁hand",
+ "ed"
+ ],
+ [
+ "▁han",
+ "ded"
+ ],
+ [
+ "▁",
+ "handed"
+ ],
+ [
+ "▁C",
+ "u"
+ ],
+ [
+ "▁H",
+ "ospital"
+ ],
+ [
+ "▁besk",
+ "revs"
+ ],
+ [
+ "не",
+ "й"
+ ],
+ [
+ "н",
+ "ей"
+ ],
+ [
+ "но",
+ "ст"
+ ],
+ [
+ "▁anx",
+ "iety"
+ ],
+ [
+ "▁heav",
+ "ily"
+ ],
+ [
+ "▁V",
+ "ar"
+ ],
+ [
+ "▁Va",
+ "r"
+ ],
+ [
+ "▁",
+ "Var"
+ ],
+ [
+ "▁dis",
+ "pos"
+ ],
+ [
+ "▁disp",
+ "os"
+ ],
+ [
+ "+",
+ "\""
+ ],
+ [
+ "▁E",
+ "ver"
+ ],
+ [
+ "▁Ev",
+ "er"
+ ],
+ [
+ "▁Eve",
+ "r"
+ ],
+ [
+ "iz",
+ "on"
+ ],
+ [
+ "izo",
+ "n"
+ ],
+ [
+ "i",
+ "zon"
+ ],
+ [
+ "▁oper",
+ "ators"
+ ],
+ [
+ "▁operator",
+ "s"
+ ],
+ [
+ "ne",
+ "go"
+ ],
+ [
+ "neg",
+ "o"
+ ],
+ [
+ "n",
+ "ego"
+ ],
+ [
+ "▁B",
+ "ry"
+ ],
+ [
+ "▁Br",
+ "y"
+ ],
+ [
+ "▁v",
+ "otes"
+ ],
+ [
+ "▁vo",
+ "tes"
+ ],
+ [
+ "▁vote",
+ "s"
+ ],
+ [
+ "▁vot",
+ "es"
+ ],
+ [
+ "iz",
+ "ione"
+ ],
+ [
+ "izi",
+ "one"
+ ],
+ [
+ "izio",
+ "ne"
+ ],
+ [
+ "i",
+ "zione"
+ ],
+ [
+ "▁ра",
+ "й"
+ ],
+ [
+ "▁fe",
+ "at"
+ ],
+ [
+ "▁",
+ "feat"
+ ],
+ [
+ "▁w",
+ "estern"
+ ],
+ [
+ "▁west",
+ "ern"
+ ],
+ [
+ "▁",
+ "western"
+ ],
+ [
+ "▁con",
+ "front"
+ ],
+ [
+ "▁strong",
+ "er"
+ ],
+ [
+ "▁ф",
+ "а"
+ ],
+ [
+ "▁",
+ "фа"
+ ],
+ [
+ "st",
+ "re"
+ ],
+ [
+ "str",
+ "e"
+ ],
+ [
+ "s",
+ "tre"
+ ],
+ [
+ "▁Val",
+ "id"
+ ],
+ [
+ "▁",
+ "Valid"
+ ],
+ [
+ "▁n",
+ "ad"
+ ],
+ [
+ "▁na",
+ "d"
+ ],
+ [
+ "▁check",
+ "ing"
+ ],
+ [
+ "▁bird",
+ "s"
+ ],
+ [
+ "▁North",
+ "ern"
+ ],
+ [
+ "▁Nor",
+ "thern"
+ ],
+ [
+ "▁int",
+ "ention"
+ ],
+ [
+ "▁intent",
+ "ion"
+ ],
+ [
+ "uc",
+ "e"
+ ],
+ [
+ "u",
+ "ce"
+ ],
+ [
+ "▁co",
+ "vers"
+ ],
+ [
+ "▁cover",
+ "s"
+ ],
+ [
+ "▁cov",
+ "ers"
+ ],
+ [
+ "▁wonder",
+ "ing"
+ ],
+ [
+ "▁Option",
+ "al"
+ ],
+ [
+ "▁Opt",
+ "ional"
+ ],
+ [
+ "▁",
+ "Optional"
+ ],
+ [
+ "pro",
+ "tocol"
+ ],
+ [
+ "proto",
+ "col"
+ ],
+ [
+ "prot",
+ "ocol"
+ ],
+ [
+ "▁ag",
+ "gress"
+ ],
+ [
+ "—",
+ "—"
+ ],
+ [
+ "V",
+ "ec"
+ ],
+ [
+ "▁d",
+ "ates"
+ ],
+ [
+ "▁da",
+ "tes"
+ ],
+ [
+ "▁dat",
+ "es"
+ ],
+ [
+ "▁date",
+ "s"
+ ],
+ [
+ "▁",
+ "dates"
+ ],
+ [
+ "qu",
+ "ot"
+ ],
+ [
+ "▁b",
+ "om"
+ ],
+ [
+ "▁bo",
+ "m"
+ ],
+ [
+ "▁s",
+ "can"
+ ],
+ [
+ "▁sc",
+ "an"
+ ],
+ [
+ "▁",
+ "scan"
+ ],
+ [
+ "▁I",
+ "tem"
+ ],
+ [
+ "▁It",
+ "em"
+ ],
+ [
+ "▁",
+ "Item"
+ ],
+ [
+ "▁N",
+ "avy"
+ ],
+ [
+ "▁Na",
+ "vy"
+ ],
+ [
+ "▁Nav",
+ "y"
+ ],
+ [
+ "▁G",
+ "ran"
+ ],
+ [
+ "▁Gr",
+ "an"
+ ],
+ [
+ "▁Gra",
+ "n"
+ ],
+ [
+ "▁every",
+ "body"
+ ],
+ [
+ "▁un",
+ "expected"
+ ],
+ [
+ "▁une",
+ "xpected"
+ ],
+ [
+ "▁di",
+ "vor"
+ ],
+ [
+ "▁div",
+ "or"
+ ],
+ [
+ "▁e",
+ "ase"
+ ],
+ [
+ "▁eas",
+ "e"
+ ],
+ [
+ "um",
+ "bled"
+ ],
+ [
+ "umb",
+ "led"
+ ],
+ [
+ "umble",
+ "d"
+ ],
+ [
+ "^",
+ "+"
+ ],
+ [
+ "cu",
+ "ss"
+ ],
+ [
+ "c",
+ "uss"
+ ],
+ [
+ "▁p",
+ "ale"
+ ],
+ [
+ "▁pal",
+ "e"
+ ],
+ [
+ "▁pa",
+ "le"
+ ],
+ [
+ "▁In",
+ "ga"
+ ],
+ [
+ "▁Ing",
+ "a"
+ ],
+ [
+ "▁B",
+ "road"
+ ],
+ [
+ "▁Br",
+ "oad"
+ ],
+ [
+ "▁Bro",
+ "ad"
+ ],
+ [
+ "▁",
+ "Broad"
+ ],
+ [
+ "▁Med",
+ "ic"
+ ],
+ [
+ "▁R",
+ "oy"
+ ],
+ [
+ "▁Ro",
+ "y"
+ ],
+ [
+ "▁I",
+ "nn"
+ ],
+ [
+ "▁In",
+ "n"
+ ],
+ [
+ "▁p",
+ "ens"
+ ],
+ [
+ "▁pe",
+ "ns"
+ ],
+ [
+ "▁pen",
+ "s"
+ ],
+ [
+ "P",
+ "N"
+ ],
+ [
+ ".",
+ ":"
+ ],
+ [
+ "▁princip",
+ "le"
+ ],
+ [
+ "▁let",
+ "ting"
+ ],
+ [
+ "▁lett",
+ "ing"
+ ],
+ [
+ "▁condu",
+ "cted"
+ ],
+ [
+ "▁conduct",
+ "ed"
+ ],
+ [
+ "F",
+ "ALSE"
+ ],
+ [
+ "▁O",
+ "S"
+ ],
+ [
+ "▁",
+ "OS"
+ ],
+ [
+ "F",
+ "ocus"
+ ],
+ [
+ "▁measure",
+ "d"
+ ],
+ [
+ "▁meas",
+ "ured"
+ ],
+ [
+ "▁Dem",
+ "ocratic"
+ ],
+ [
+ "▁Democr",
+ "atic"
+ ],
+ [
+ "▁Democrat",
+ "ic"
+ ],
+ [
+ "Hi",
+ "gh"
+ ],
+ [
+ "H",
+ "igh"
+ ],
+ [
+ "▁p",
+ "ré"
+ ],
+ [
+ "▁pr",
+ "é"
+ ],
+ [
+ "en",
+ "nes"
+ ],
+ [
+ "enn",
+ "es"
+ ],
+ [
+ "enne",
+ "s"
+ ],
+ [
+ "▁ind",
+ "icates"
+ ],
+ [
+ "▁indic",
+ "ates"
+ ],
+ [
+ "▁indicate",
+ "s"
+ ],
+ [
+ "▁en",
+ "ding"
+ ],
+ [
+ "▁end",
+ "ing"
+ ],
+ [
+ "▁",
+ "ending"
+ ],
+ [
+ "▁Sm",
+ "all"
+ ],
+ [
+ "▁",
+ "Small"
+ ],
+ [
+ "▁<",
+ "!--"
+ ],
+ [
+ "▁",
+ "