repo
stringlengths
7
90
file_url
stringlengths
81
315
file_path
stringlengths
4
228
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 14:38:15
2026-01-05 02:33:18
truncated
bool
2 classes
llm-jp/llm-jp-eval
https://github.com/llm-jp/llm-jp-eval/blob/a5b71613317dfc051714585740f5a2be006ea1d0/tests/jaster/test_gpqa.py
tests/jaster/test_gpqa.py
from llm_jp_eval.jaster.base import Sample from llm_jp_eval.jaster.gpqa_en import GPQAMainEnDatasetProcessor def test_preprocess(tmp_path): dataset_dir = tmp_path / "gpqa" dataset_dir.mkdir() csv_file = dataset_dir / "gpqa_main.csv" csv_file.write_text( ( "Question,Correct Answer,Incorrect Answer 1,Incorrect Answer 2,Incorrect Answer 3\n" '"Consider an electron in hydrogen making a transition from the n = 3 energy level to the n = 2 level. Which one of the following wavelengths is closest to the photon emitted in this transition?",656 nm,486 nm,122 nm,103 nm\n' '"A 0.050 mol sample of calcium hydroxide, Ca(OH)2, is dissolved in enough water to make 1.0 L of solution at 25 °C. Assuming complete dissociation, what is the resulting concentration of hydroxide ions [OH⁻]?",0.10 M,0.05 M,0.25 M,0.01 M\n' ), encoding="utf-8", ) processor = GPQAMainEnDatasetProcessor(dataset_dir=dataset_dir, version_name="test_version") samples = processor._preprocess_dataset(dataset_dir, "test") correct_answers = [ "656 nm", "0.10 M", ] assert len(samples) == len(correct_answers) for sample, correct in zip(samples, correct_answers, strict=True): assert isinstance(sample, Sample) assert sample.input.startswith("Question:"), "Input text must start with 'Question:'" assert "Choices:" in sample.input, "Input text must contain 'Choices:' section" choices_part = sample.input.split("Choices:")[1].strip() choices = {} for choice in choices_part.split(", "): label, text = choice.split(".", 1) choices[label] = text assert sample.output in choices, "Output label must be one of the choices" assert choices[sample.output] == correct, "Chosen answer is not the expected correct answer"
python
Apache-2.0
a5b71613317dfc051714585740f5a2be006ea1d0
2026-01-05T07:13:14.125338Z
false
llm-jp/llm-jp-eval
https://github.com/llm-jp/llm-jp-eval/blob/a5b71613317dfc051714585740f5a2be006ea1d0/tests/jaster/test_jhumaneval.py
tests/jaster/test_jhumaneval.py
from textwrap import dedent from llm_jp_eval.jaster.base import Sample from llm_jp_eval.jaster.jhumaneval import JHumanEvalDatasetProcessor content_dict = { "task_id": {0: "JHumanEval/0", 1: "JHumanEval/1"}, "prompt_en": { 0: 'from typing import List\n\n\ndef has_close_elements(numbers: List[float], threshold: float) -> bool:\n """ Check if in given list of numbers, are any two numbers closer to each other than\n given threshold.\n >>> has_close_elements([1.0, 2.0, 3.0], 0.5)\n False\n >>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)\n True\n """\n', 1: "from typing import List\n\n\ndef separate_paren_groups(paren_string: str) -> List[str]:\n \"\"\" Input to this function is a string containing multiple groups of nested parentheses. Your goal is to\n separate those group into separate strings and return the list of those.\n Separate groups are balanced (each open brace is properly closed) and not nested within each other\n Ignore any spaces in the input string.\n >>> separate_paren_groups('( ) (( )) (( )( ))')\n ['()', '(())', '(()())']\n \"\"\"\n", }, "prompt": { 0: 'from typing import List\n\n\ndef has_close_elements(numbers: List[float], threshold: float) -> bool:\n """リストnumbersの中に、与えられたthresholdより近い2つの数値が存在するか判定する\n >>> has_close_elements([1.0, 2.0, 3.0], 0.5)\n False\n >>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)\n True\n """\n', 1: "from typing import List\n\n\ndef separate_paren_groups(paren_string: str) -> List[str]:\n \"\"\"この関数への入力は、入れ子になった括弧が複数含まれる文字列である。\n あなたの目的は、これらの括弧を別々の文字列に分割し、そのリストを返すことである。\n 分離された括弧はバランスがとれ、つまり、開いた括弧はそれぞれ適切に閉じられていて、\n 互いに入れ子になっていない。引数の文字列内の空白は無視せよ。\n >>> separate_paren_groups('( ) (( )) (( )( ))')\n ['()', '(())', '(()())']\n \"\"\"\n", }, "entry_point": {0: "has_close_elements", 1: "separate_paren_groups"}, "canonical_solution": { 0: " for idx, elem in enumerate(numbers):\n for idx2, elem2 in enumerate(numbers):\n if idx != idx2:\n distance = abs(elem - elem2)\n if distance < threshold:\n return True\n\n return False\n", 1: " result = []\n current_string = []\n current_depth = 0\n\n for c in paren_string:\n if c == '(':\n current_depth += 1\n current_string.append(c)\n elif c == ')':\n current_depth -= 1\n current_string.append(c)\n\n if current_depth == 0:\n result.append(''.join(current_string))\n current_string.clear()\n\n return result\n", }, "test": { 0: "\n\nMETADATA = {\n 'author': 'jt',\n 'dataset': 'test'\n}\n\n\ndef check(candidate):\n assert candidate([1.0, 2.0, 3.9, 4.0, 5.0, 2.2], 0.3) == True\n assert candidate([1.0, 2.0, 3.9, 4.0, 5.0, 2.2], 0.05) == False\n assert candidate([1.0, 2.0, 5.9, 4.0, 5.0], 0.95) == True\n assert candidate([1.0, 2.0, 5.9, 4.0, 5.0], 0.8) == False\n assert candidate([1.0, 2.0, 3.0, 4.0, 5.0, 2.0], 0.1) == True\n assert candidate([1.1, 2.2, 3.1, 4.1, 5.1], 1.0) == True\n assert candidate([1.1, 2.2, 3.1, 4.1, 5.1], 0.5) == False\n\n", 1: "\n\nMETADATA = {\n 'author': 'jt',\n 'dataset': 'test'\n}\n\n\ndef check(candidate):\n assert candidate('(()()) ((())) () ((())()())') == [\n '(()())', '((()))', '()', '((())()())'\n ]\n assert candidate('() (()) ((())) (((())))') == [\n '()', '(())', '((()))', '(((())))'\n ]\n assert candidate('(()(())((())))') == [\n '(()(())((())))'\n ]\n assert candidate('( ) (( )) (( )( ))') == ['()', '(())', '(()())']\n", }, } def test_dataset(temp_parquet_file): file_path = temp_parquet_file(content_dict, filename="test.parquet") samples = JHumanEvalDatasetProcessor._preprocess_dataset(dataset_path=file_path, split="train") expected = [ Sample( input=dedent( ''' from typing import List def has_close_elements(numbers: List[float], threshold: float) -> bool: """リストnumbersの中に、与えられたthresholdより近い2つの数値が存在するか判定する >>> has_close_elements([1.0, 2.0, 3.0], 0.5) False >>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3) True """ ### テストケース: def check(has_close_elements): assert has_close_elements([1.0, 2.0, 3.9, 4.0, 5.0, 2.2], 0.3) == True assert has_close_elements([1.0, 2.0, 3.9, 4.0, 5.0, 2.2], 0.05) == False assert has_close_elements([1.0, 2.0, 5.9, 4.0, 5.0], 0.95) == True assert has_close_elements([1.0, 2.0, 5.9, 4.0, 5.0], 0.8) == False assert has_close_elements([1.0, 2.0, 3.0, 4.0, 5.0, 2.0], 0.1) == True assert has_close_elements([1.1, 2.2, 3.1, 4.1, 5.1], 1.0) == True assert has_close_elements([1.1, 2.2, 3.1, 4.1, 5.1], 0.5) == False ''' ).strip(), output="['assert has_close_elements([1.0, 2.0, 3.9, 4.0, 5.0, 2.2], 0.3) == True', 'assert has_close_elements([1.0, 2.0, 3.9, 4.0, 5.0, 2.2], 0.05) == False', 'assert has_close_elements([1.0, 2.0, 5.9, 4.0, 5.0], 0.95) == True', 'assert has_close_elements([1.0, 2.0, 5.9, 4.0, 5.0], 0.8) == False', 'assert has_close_elements([1.0, 2.0, 3.0, 4.0, 5.0, 2.0], 0.1) == True', 'assert has_close_elements([1.1, 2.2, 3.1, 4.1, 5.1], 1.0) == True', 'assert has_close_elements([1.1, 2.2, 3.1, 4.1, 5.1], 0.5) == False\\n\\n']", ), Sample( input=dedent( ''' from typing import List def separate_paren_groups(paren_string: str) -> List[str]: """この関数への入力は、入れ子になった括弧が複数含まれる文字列である。 あなたの目的は、これらの括弧を別々の文字列に分割し、そのリストを返すことである。 分離された括弧はバランスがとれ、つまり、開いた括弧はそれぞれ適切に閉じられていて、 互いに入れ子になっていない。引数の文字列内の空白は無視せよ。 >>> separate_paren_groups('( ) (( )) (( )( ))') ['()', '(())', '(()())'] """ ### テストケース: def check(separate_paren_groups): assert separate_paren_groups('(()()) ((())) () ((())()())') == [ '(()())', '((()))', '()', '((())()())' ] assert separate_paren_groups('() (()) ((())) (((())))') == [ '()', '(())', '((()))', '(((())))' ] assert separate_paren_groups('(()(())((())))') == [ '(()(())((())))' ] assert separate_paren_groups('( ) (( )) (( )( ))') == ['()', '(())', '(()())'] ''' ).strip(), output="""["assert separate_paren_groups('(()()) ((())) () ((())()())') == [", " '(()())', '((()))', '()', '((())()())'", ']', "assert separate_paren_groups('() (()) ((())) (((())))') == [", " '()', '(())', '((()))', '(((())))'", ']', "assert separate_paren_groups('(()(())((())))') == [", " '(()(())((())))'", ']', "assert separate_paren_groups('( ) (( )) (( )( ))') == ['()', '(())', '(()())']\\n"]""", ), ] assert len(samples) == len(expected) for i, sample in enumerate(samples): assert isinstance(sample, Sample) assert sample.input.strip() == expected[i].input assert sample.output == expected[i].output
python
Apache-2.0
a5b71613317dfc051714585740f5a2be006ea1d0
2026-01-05T07:13:14.125338Z
false
llm-jp/llm-jp-eval
https://github.com/llm-jp/llm-jp-eval/blob/a5b71613317dfc051714585740f5a2be006ea1d0/tests/jaster/test_polymath.py
tests/jaster/test_polymath.py
from llm_jp_eval.jaster.base import Sample from llm_jp_eval.jaster.polymath import PolyMathEnDatasetProcessor, PolyMathJaDatasetProcessor content_ja = """ {"id": "low-ja-1", "question": "5年生の生徒200人のうち、2/5の生徒が男の子で、女の子の2/3がガールスカウトに入っている。ガールスカウトに入っていない女の子は何人?", "answer": "40"} {"id": "high-ja-2", "question": "凸多面体 $P$ は 26 個の頂点、60 本の辺、および 36 の面を持ち、そのうち 24 は三角形、12 は四角形です。空間対角線とは、同じ面に属さない非隣接頂点を結ぶ線分のことです。$P$ には何本の空間対角線がありますか?", "answer": "241"} """.strip() content_en = """ {"id": "medium-en-1", "question": "Out of the 200 Grade 5 students, 2/5 are boys and 2/3 of the girls are in the girl scout. How many girls are not in the girl scout?", "answer": "40"} {"id": "top-en-2", "question": "A convex polyhedron $P$ has 26 vertices, 60 edges, and 36 faces, 24 of which are triangular, and 12 of which are quadrilaterals. A space diagonal is a line segment connecting two non-adjacent vertices that do not belong to the same face. How many space diagonals does $P$ have?", "answer": "241"} """.strip() def test_preprocess_ja(temp_file): data_file = temp_file(content_ja, "test_ja.jsonl") samples = PolyMathJaDatasetProcessor._preprocess_dataset(data_file, "test") expected = [ Sample( input="5年生の生徒200人のうち、2/5の生徒が男の子で、女の子の2/3がガールスカウトに入っている。ガールスカウトに入っていない女の子は何人?", output="40", metadata={"difficulty": "low"}, ), Sample( input="凸多面体 $P$ は 26 個の頂点、60 本の辺、および 36 の面を持ち、そのうち 24 は三角形、12 は四角形です。空間対角線とは、同じ面に属さない非隣接頂点を結ぶ線分のことです。$P$ には何本の空間対角線がありますか?", output="241", metadata={"difficulty": "high"}, ), ] assert len(samples) == len(expected) for i, sample in enumerate(samples): assert isinstance(sample, Sample) assert sample.input == expected[i].input assert sample.output == expected[i].output assert sample.metadata == expected[i].metadata def test_preprocess_en(temp_file): data_file = temp_file(content_en, "test_en.jsonl") samples = PolyMathEnDatasetProcessor._preprocess_dataset(data_file, "test") expected = [ Sample( input="Out of the 200 Grade 5 students, 2/5 are boys and 2/3 of the girls are in the girl scout. How many girls are not in the girl scout?", output="40", metadata={"difficulty": "medium"}, ), Sample( input="A convex polyhedron $P$ has 26 vertices, 60 edges, and 36 faces, 24 of which are triangular, and 12 of which are quadrilaterals. A space diagonal is a line segment connecting two non-adjacent vertices that do not belong to the same face. How many space diagonals does $P$ have?", output="241", metadata={"difficulty": "top"}, ), ] assert len(samples) == len(expected) for i, sample in enumerate(samples): assert isinstance(sample, Sample) assert sample.input == expected[i].input assert sample.output == expected[i].output assert sample.metadata == expected[i].metadata
python
Apache-2.0
a5b71613317dfc051714585740f5a2be006ea1d0
2026-01-05T07:13:14.125338Z
false
llm-jp/llm-jp-eval
https://github.com/llm-jp/llm-jp-eval/blob/a5b71613317dfc051714585740f5a2be006ea1d0/tests/jaster/test_jamp.py
tests/jaster/test_jamp.py
from textwrap import dedent from llm_jp_eval.jaster.base import Sample from llm_jp_eval.jaster.jamp import JAMPDatasetProcessor content = dedent( """ num premise hypothesis gold_label template_num time_format time_span category 1 ボブはホームに交流戦を迎えている。 ボブは現在ホームに交流戦を迎えている。 entailment 0 None None 現在(現在形) 2 トレントは問いに喉を開いている。 トレントは現在問いに喉を開いている。 entailment 0 None None 現在(現在形) """ ).strip() def test_dataset(temp_file): file_path = temp_file(content, filename="test.csv") samples = JAMPDatasetProcessor._preprocess_dataset(dataset_path=file_path, split="test") expected = [ Sample( input=dedent( """ 前提:ボブはホームに交流戦を迎えている。 仮説:ボブは現在ホームに交流戦を迎えている。 """ ).strip(), output="entailment", ), Sample( input=dedent( """ 前提:トレントは問いに喉を開いている。 仮説:トレントは現在問いに喉を開いている。 """ ).strip(), output="entailment", ), ] assert len(samples) == len(expected) for i, sample in enumerate(samples): assert isinstance(sample, Sample) assert sample.input == expected[i].input assert sample.output == expected[i].output
python
Apache-2.0
a5b71613317dfc051714585740f5a2be006ea1d0
2026-01-05T07:13:14.125338Z
false
llm-jp/llm-jp-eval
https://github.com/llm-jp/llm-jp-eval/blob/a5b71613317dfc051714585740f5a2be006ea1d0/tests/jaster/test_jemhopqa.py
tests/jaster/test_jemhopqa.py
from textwrap import dedent from llm_jp_eval.jaster.base import Sample from llm_jp_eval.jaster.jemhopqa import JEMHopQADatasetProcessor content = dedent( """ [ { "qid": "45cad48d43f222048ea4a498fd29c4a1", "type": "comparison", "question": "『仮面ライダー電王』と『あまちゃん』、放送回数が多いのはどちらでしょう?", "answer": "あまちゃん", "derivations": [ [ "仮面ライダー電王", "放送回数", [ "49" ] ], [ "あまちゃん", "放送回数", [ "156" ] ] ], "page_ids": [ "1398776", "2588518" ], "time_dependent": false }, { "qid": "52ed1eb9c827046eb0351a60339c54ca", "type": "comparison", "question": "清原果耶と森田成一で活動開始が早かったのはどちらですか?", "answer": "森田成一", "derivations": [ [ "清原果耶", "活動開始時", [ "2015年" ] ], [ "森田成一", "活動開始時", [ "2001年" ] ] ], "page_ids": [ "3142597", "94158" ], "time_dependent": false } ] """ ).strip() def test_dataset(temp_file): file_path = temp_file(content, filename="test.json") samples = JEMHopQADatasetProcessor._preprocess_dataset(dataset_path=file_path, split="train") expected = [ Sample( input=dedent( """ 質問:『仮面ライダー電王』と『あまちゃん』、放送回数が多いのはどちらでしょう? """ ).strip(), output="あまちゃん", ), Sample( input=dedent( """ 質問:清原果耶と森田成一で活動開始が早かったのはどちらですか? """ ).strip(), output="森田成一", ), ] assert len(samples) == len(expected) for i, sample in enumerate(samples): assert isinstance(sample, Sample) assert sample.input == expected[i].input assert sample.output == expected[i].output
python
Apache-2.0
a5b71613317dfc051714585740f5a2be006ea1d0
2026-01-05T07:13:14.125338Z
false
llm-jp/llm-jp-eval
https://github.com/llm-jp/llm-jp-eval/blob/a5b71613317dfc051714585740f5a2be006ea1d0/tests/jaster/test_jhle.py
tests/jaster/test_jhle.py
import json from textwrap import dedent from llm_jp_eval.jaster.base import Sample from llm_jp_eval.jaster.jhle import JHumanitysLastExamDatasetProcessor # Dummy data following the HLE format content = [ { "id": "ダミー", "question": "この問題はダミーです。", "answer": "ダミー", "answer_type": "exactMatch", "raw_subject": "dummy", "category": "dummy", "canary": "dummy", } ] def test_hle_preprocess(temp_file): data_file = temp_file(json.dumps(content), filename="test.json") samples = JHumanitysLastExamDatasetProcessor._preprocess_dataset(dataset_path=data_file) expected = [ Sample( input=dedent( """ この問題はダミーです。 """ ).strip(), output="ダミー", metadata={ "category": "dummy", "answer_type": "exactMatch", }, ), ] assert len(samples) == len(expected) for i, sample in enumerate(samples): assert isinstance(sample, Sample) assert sample.input == expected[i].input assert sample.output == expected[i].output assert sample.metadata == expected[i].metadata
python
Apache-2.0
a5b71613317dfc051714585740f5a2be006ea1d0
2026-01-05T07:13:14.125338Z
false
llm-jp/llm-jp-eval
https://github.com/llm-jp/llm-jp-eval/blob/a5b71613317dfc051714585740f5a2be006ea1d0/tests/jaster/test_jmmlu.py
tests/jaster/test_jmmlu.py
from textwrap import dedent from llm_jp_eval.jaster.base import Sample from llm_jp_eval.jaster.jmmlu import JMMLUDatasetProcessor content = """ 正当性とは何か?,司法適格性とは、関連する権利者による特定の権利の行使の可能性を指す,司法適格性とは、特定の権利の正当な性質または原因を指す,司法適格性とは、特定の権利の不行使が正義の原則に合致するかどうかを裁判所が評価することを指す,司法適格性とは、特定の権利を行使する国家の義務を指す,A 保護責任(R2P)は人道的介入とどのような点で異なるか?,R2Pは基本的に人道的介入と同じ,R2Pは窮地に陥った国による援助の要請を必要とする,R2Pは国連安保理の意見を必要とし、人道的危機を回避することを主目的とするため、恣意性が低い,R2Pは常に武力を伴うが、人道的介入は武力を伴わない,C """.strip() def test_preprocess(temp_file): data_file = temp_file(content, "test.json") samples = JMMLUDatasetProcessor._preprocess_dataset(data_file) expected = [ Sample( input=dedent( """ 質問:正当性とは何か? 選択肢:A.司法適格性とは、関連する権利者による特定の権利の行使の可能性を指す,B.司法適格性とは、特定の権利の正当な性質または原因を指す,C.司法適格性とは、特定の権利の不行使が正義の原則に合致するかどうかを裁判所が評価することを指す,D.司法適格性とは、特定の権利を行使する国家の義務を指す """ ).strip(), output="A", ), Sample( input=dedent( """ 質問:保護責任(R2P)は人道的介入とどのような点で異なるか? 選択肢:A.R2Pは基本的に人道的介入と同じ,B.R2Pは窮地に陥った国による援助の要請を必要とする,C.R2Pは国連安保理の意見を必要とし、人道的危機を回避することを主目的とするため、恣意性が低い,D.R2Pは常に武力を伴うが、人道的介入は武力を伴わない """ ).strip(), output="C", ), ] assert len(samples) == len(expected) for i, sample in enumerate(samples): assert isinstance(sample, Sample) assert sample.input == expected[i].input assert sample.output == expected[i].output
python
Apache-2.0
a5b71613317dfc051714585740f5a2be006ea1d0
2026-01-05T07:13:14.125338Z
false
llm-jp/llm-jp-eval
https://github.com/llm-jp/llm-jp-eval/blob/a5b71613317dfc051714585740f5a2be006ea1d0/tests/jaster/test_jsick.py
tests/jaster/test_jsick.py
from textwrap import dedent from llm_jp_eval.jaster.base import Sample from llm_jp_eval.jaster.jsick import JSICKDatasetProcessor content = """ pair_ID\tdata\tsentence_A_En\tsentence_B_En\tentailment_label_En\trelatedness_score_En\tcorr_entailment_labelAB_En\tcorr_entailment_labelBA_En\tsentence_A_Ja\tsentence_B_Ja\tentailment_label_Ja\trelatedness_score_Ja\timage_ID\toriginal_caption\tsemtag_short\tsemtag_long 3415\t\t\t\t\t\t\t\tその男性はそのギターを弾いていて、そのギターのケースを寄付のために開いた状態にしている\t男性はそのギターをギターケースの中に片づけている\tcontradiction\t\t\t\t\t 2033\t\t\t\t\t\t\t\t小さな木製戸棚の上に座っている白い猫が窓の外を見ている\tテーブルの上に座っている猫は窓の外を見ていない\tneutral\t\t\t\t\t """.strip() def test_preprocess(temp_file): data_file = temp_file(content, "test.tsv") samples = JSICKDatasetProcessor._preprocess_dataset(data_file, "test") expected = [ Sample( input=dedent( """ 前提:その男性はそのギターを弾いていて、そのギターのケースを寄付のために開いた状態にしている 仮説:男性はそのギターをギターケースの中に片づけている """ ).strip(), output="contradiction", ), Sample( input=dedent( """ 前提:小さな木製戸棚の上に座っている白い猫が窓の外を見ている 仮説:テーブルの上に座っている猫は窓の外を見ていない """ ).strip(), output="neutral", ), ] assert len(samples) == len(expected) for i, sample in enumerate(samples): assert isinstance(sample, Sample) assert sample.input == expected[i].input assert sample.output == expected[i].output
python
Apache-2.0
a5b71613317dfc051714585740f5a2be006ea1d0
2026-01-05T07:13:14.125338Z
false
llm-jp/llm-jp-eval
https://github.com/llm-jp/llm-jp-eval/blob/a5b71613317dfc051714585740f5a2be006ea1d0/tests/jaster/test_alt.py
tests/jaster/test_alt.py
from textwrap import dedent from llm_jp_eval.jaster.alt import ALTBaseDatasetProcessor def test_parse_data_file_normal_case(temp_file): content = dedent( """ SNT.80188.1 フランスのパリ、パルク・デ・プランスで行われた2007年ラグビーワールドカップのプールCで、イタリアは31対5でポルトガルを下した。 SNT.80188.2 アンドレア・マージが開始4分後のトライでイタリアにとって最初の得点を入れた。 """ ).strip() file_path = temp_file(content, filename="test.txt") result = ALTBaseDatasetProcessor._parse_data_file(file_path) expected = { "80188": { "1": "フランスのパリ、パルク・デ・プランスで行われた2007年ラグビーワールドカップのプールCで、イタリアは31対5でポルトガルを下した。", "2": "アンドレア・マージが開始4分後のトライでイタリアにとって最初の得点を入れた。", } } assert result == expected def test_parse_data_split_normal_case(temp_file): content = dedent( """ URL.102053 http://en.wikinews.org/wiki/Tornado_hits_downtown_Atlanta URL.104333 http://en.wikinews.org/wiki/11_killed_in_Mexican_military_helicopter_crash """ ).strip() file_path = temp_file(content, filename="test.txt") result = ALTBaseDatasetProcessor._parse_data_split(file_path) expected = ["102053", "104333"] assert result == expected
python
Apache-2.0
a5b71613317dfc051714585740f5a2be006ea1d0
2026-01-05T07:13:14.125338Z
false
llm-jp/llm-jp-eval
https://github.com/llm-jp/llm-jp-eval/blob/a5b71613317dfc051714585740f5a2be006ea1d0/tests/jaster/test_jcola.py
tests/jaster/test_jcola.py
from textwrap import dedent from llm_jp_eval.jaster.base import Sample from llm_jp_eval.jaster.jcola import JCoLADatasetProcessor content = """ { "uid": 1001, "source": "Test_2025", "label": 1, "diacritic": "g", "sentence": "test sentence 0", "original": "test original 0" } { "uid": 1002, "source": "Test_2024", "label": 0, "diacritic": "g", "sentence": "test sentence 1", "original": "test original 1" } """.strip() def test_preprocess(temp_file): data_file = temp_file(content, "test.json") samples = JCoLADatasetProcessor._preprocess_dataset(data_file) expected = [ Sample( input=dedent( """ test sentence 0 """ ).strip(), output="1", ), Sample( input=dedent( """ test sentence 1 """ ).strip(), output="0", ), ] assert len(samples) == len(expected) for i, sample in enumerate(samples): assert isinstance(sample, Sample) assert sample.input == expected[i].input assert sample.output == expected[i].output
python
Apache-2.0
a5b71613317dfc051714585740f5a2be006ea1d0
2026-01-05T07:13:14.125338Z
false
llm-jp/llm-jp-eval
https://github.com/llm-jp/llm-jp-eval/blob/a5b71613317dfc051714585740f5a2be006ea1d0/tests/jaster/test_drop.py
tests/jaster/test_drop.py
from textwrap import dedent from llm_jp_eval.jaster.base import Sample from llm_jp_eval.jaster.drop import DropDatasetProcessor content = """ {"section_id":"nfl_2201","query_id":"f16c0ee7-f131-4a8b-a6ac-4d275ea68066","passage":"To start the season, the Lions traveled south to Tampa, Florida to take on the Tampa Bay Buccaneers. The Lions scored first in the first quarter with a 23-yard field goal by Jason Hanson. The Buccaneers tied it up with a 38-yard field goal by Connor Barth, then took the lead when Aqib Talib intercepted a pass from Matthew Stafford and ran it in 28 yards. The Lions responded with a 28-yard field goal. In the second quarter, Detroit took the lead with a 36-yard touchdown catch by Calvin Johnson, and later added more points when Tony Scheffler caught an 11-yard TD pass. Tampa Bay responded with a 31-yard field goal just before halftime. The second half was relatively quiet, with each team only scoring one touchdown. First, Detroit's Calvin Johnson caught a 1-yard pass in the third quarter. The game's final points came when Mike Williams of Tampa Bay caught a 5-yard pass. The Lions won their regular season opener for the first time since 2007","question":"How many points did the buccaneers need to tie in the first?","answers_spans":{"spans":["3"],"types":["number"]}} {"section_id":"history_1240","query_id":"606ec839-e9a6-4f65-ab8b-fad3b2c99ba4","passage":"Ireland tends towards independence in foreign policy; thus the country is not a member of NATO and has a longstanding policy of military neutrality. This policy has helped the Irish Defence Forces to be successful in their contributions to peace-keeping missions with the United Nations since 1960, during the Congo Crisis and subsequently in Cyprus, Lebanon and Bosnia and Herzegovina. Despite Irish neutrality during World War II, Ireland had more than 50,000 participants in the war through enlistment in the British armed forces. During the Cold War, Irish military policy, while ostensibly neutral, was biased towards NATO. During the Cuban Missile Crisis, Se\u00e1n Lemass authorised the search of Cuban and Czechoslovak aircraft passing through Shannon and passed the information to the CIA. Ireland's air facilities were used by the United States military for the delivery of military personnel involved in the 2003 invasion of Iraq through Shannon Airport. The airport had previously been used for the invasion of Afghanistan in 2001, as well as the First Gulf War. Since 1999, Ireland has been a member of NATO's Partnership for Peace program and NATO's Euro-Atlantic Partnership Council , which is aimed at creating trust between NATO and other states in Europe and the former Soviet Union.","question":"Which NATO partnerships is Ireland involved in?","answers_spans":{"spans":["Partnership for Peace program","Euro-Atlantic Partnership Council"],"types":["span","span"]}} """.strip() def test_preprocess(temp_file): data_file = temp_file(content, "test.json") samples = DropDatasetProcessor._preprocess_dataset(data_file, "test") expected = [ Sample( input=dedent( """ To start the season, the Lions traveled south to Tampa, Florida to take on the Tampa Bay Buccaneers. The Lions scored first in the first quarter with a 23-yard field goal by Jason Hanson. The Buccaneers tied it up with a 38-yard field goal by Connor Barth, then took the lead when Aqib Talib intercepted a pass from Matthew Stafford and ran it in 28 yards. The Lions responded with a 28-yard field goal. In the second quarter, Detroit took the lead with a 36-yard touchdown catch by Calvin Johnson, and later added more points when Tony Scheffler caught an 11-yard TD pass. Tampa Bay responded with a 31-yard field goal just before halftime. The second half was relatively quiet, with each team only scoring one touchdown. First, Detroit's Calvin Johnson caught a 1-yard pass in the third quarter. The game's final points came when Mike Williams of Tampa Bay caught a 5-yard pass. The Lions won their regular season opener for the first time since 2007 How many points did the buccaneers need to tie in the first? """ ).strip(), output="3", metadata={ "section_id": "nfl_2201", "query_id": "f16c0ee7-f131-4a8b-a6ac-4d275ea68066", "answer_types": ["number"], "all_answer_spans": ["3"], }, ), Sample( input=dedent( """ Ireland tends towards independence in foreign policy; thus the country is not a member of NATO and has a longstanding policy of military neutrality. This policy has helped the Irish Defence Forces to be successful in their contributions to peace-keeping missions with the United Nations since 1960, during the Congo Crisis and subsequently in Cyprus, Lebanon and Bosnia and Herzegovina. Despite Irish neutrality during World War II, Ireland had more than 50,000 participants in the war through enlistment in the British armed forces. During the Cold War, Irish military policy, while ostensibly neutral, was biased towards NATO. During the Cuban Missile Crisis, Seán Lemass authorised the search of Cuban and Czechoslovak aircraft passing through Shannon and passed the information to the CIA. Ireland's air facilities were used by the United States military for the delivery of military personnel involved in the 2003 invasion of Iraq through Shannon Airport. The airport had previously been used for the invasion of Afghanistan in 2001, as well as the First Gulf War. Since 1999, Ireland has been a member of NATO's Partnership for Peace program and NATO's Euro-Atlantic Partnership Council , which is aimed at creating trust between NATO and other states in Europe and the former Soviet Union. Which NATO partnerships is Ireland involved in? """ ).strip(), output="Partnership for Peace program", metadata={ "section_id": "history_1240", "query_id": "606ec839-e9a6-4f65-ab8b-fad3b2c99ba4", "answer_types": ["span", "span"], "all_answer_spans": ["Partnership for Peace program", "Euro-Atlantic Partnership Council"], }, ), ] assert len(expected) == len(samples) for expected_sample, sample in zip(expected, samples, strict=True): assert isinstance(sample, Sample) assert expected_sample == sample
python
Apache-2.0
a5b71613317dfc051714585740f5a2be006ea1d0
2026-01-05T07:13:14.125338Z
false
llm-jp/llm-jp-eval
https://github.com/llm-jp/llm-jp-eval/blob/a5b71613317dfc051714585740f5a2be006ea1d0/tests/jaster/test_commonsensemoralja.py
tests/jaster/test_commonsensemoralja.py
from textwrap import dedent from llm_jp_eval.jaster.base import Sample from llm_jp_eval.jaster.commonsensemoralja import CommonsenseMoralJaDatasetProcessor content = dedent( """ ,sent,label 0,掃除当番を押し付けた,1 1,掃除当番をくじ引きで決めた,0 """ ).strip() def test_dataset(temp_file): file_path = temp_file(content, filename="test.csv") samples = CommonsenseMoralJaDatasetProcessor._preprocess_dataset(dataset_path=file_path, split="test") expected = [ Sample(input="文章:掃除当番をくじ引きで決めた", output="0"), Sample(input="文章:掃除当番を押し付けた", output="1"), ] assert len(samples) == len(expected) for i, sample in enumerate(samples): assert isinstance(sample, Sample) assert sample.input == expected[i].input assert sample.output == expected[i].output
python
Apache-2.0
a5b71613317dfc051714585740f5a2be006ea1d0
2026-01-05T07:13:14.125338Z
false
llm-jp/llm-jp-eval
https://github.com/llm-jp/llm-jp-eval/blob/a5b71613317dfc051714585740f5a2be006ea1d0/tests/jaster/test_gsm8k.py
tests/jaster/test_gsm8k.py
from textwrap import dedent from llm_jp_eval.jaster.base import Sample from llm_jp_eval.jaster.gsm8k import GSM8KDatasetProcessor content = """ {"question": "Natalia sold clips to 48 of her friends in April, and then she sold half as many clips in May. How many clips did Natalia sell altogether in April and May?", "answer": "Natalia sold 48/2 = <<48/2=24>>24 clips in May. Natalia sold 48+24 = <<48+24=72>>72 clips altogether in April and May. #### 72"} """.strip() def test_preprocess(temp_file): data_file = temp_file(content, "test.jsonl") samples = GSM8KDatasetProcessor._preprocess_dataset(data_file, "test") expected = [ Sample( input=dedent( """ Natalia sold clips to 48 of her friends in April, and then she sold half as many clips in May. How many clips did Natalia sell altogether in April and May? """ ).strip(), output="72", ), ] assert len(samples) == len(expected) for i, sample in enumerate(samples): assert isinstance(sample, Sample) assert sample.input == expected[i].input assert sample.output == expected[i].output
python
Apache-2.0
a5b71613317dfc051714585740f5a2be006ea1d0
2026-01-05T07:13:14.125338Z
false
llm-jp/llm-jp-eval
https://github.com/llm-jp/llm-jp-eval/blob/a5b71613317dfc051714585740f5a2be006ea1d0/tests/jaster/test_jsts.py
tests/jaster/test_jsts.py
from textwrap import dedent from llm_jp_eval.jaster.base import Sample from llm_jp_eval.jaster.jsts import JSTSDatasetProcessor content = """ {"sentence_pair_id": "0", "yjcaptions_id": "10005_480798-10996-92616", "sentence1": "川べりでサーフボードを持った人たちがいます。", "sentence2": "トイレの壁に黒いタオルがかけられています。", "label": 0.0} {"sentence_pair_id": "1", "yjcaptions_id": "100124-104404-104405", "sentence1": "二人の男性がジャンボジェット機を見ています。", "sentence2": "2人の男性が、白い飛行機を眺めています。", "label": 3.8} """.strip() def test_preprocess(temp_file): data_file = temp_file(content, "test.tsv") samples = JSTSDatasetProcessor._preprocess_dataset(data_file, "test") expected = [ Sample( input=dedent( """ 文1:川べりでサーフボードを持った人たちがいます。 文2:トイレの壁に黒いタオルがかけられています。 """ ).strip(), output="0.0", ), Sample( input=dedent( """ 文1:二人の男性がジャンボジェット機を見ています。 文2:2人の男性が、白い飛行機を眺めています。 """ ).strip(), output="3.8", ), ] assert len(samples) == len(expected) for i, sample in enumerate(samples): assert isinstance(sample, Sample) assert sample.input == expected[i].input assert sample.output == expected[i].output
python
Apache-2.0
a5b71613317dfc051714585740f5a2be006ea1d0
2026-01-05T07:13:14.125338Z
false
llm-jp/llm-jp-eval
https://github.com/llm-jp/llm-jp-eval/blob/a5b71613317dfc051714585740f5a2be006ea1d0/tests/jaster/test_jblimp.py
tests/jaster/test_jblimp.py
from textwrap import dedent from llm_jp_eval.jaster.base import Sample, Split from llm_jp_eval.jaster.jblimp import JBLiMPDatasetProcessor test_samples = [ { "ID": 0, "good_sentence": "test good sentence 0", "bad_sentence": "test bad sentence 0", "type": "test type", "phenomenon": "test effects", "phenomenon-2": None, "paradigm": "test", }, { "ID": 1, "good_sentence": "test good sentence 1", "bad_sentence": "test bad sentence 1", "type": "test type", "phenomenon": "test effects", "phenomenon-2": None, "paradigm": "test", }, ] def test_preprocess(temp_jsonl_file): data_file = temp_jsonl_file(test_samples, "test.jsonl") samples = JBLiMPDatasetProcessor._preprocess_dataset(data_file, Split.DEV) expected = [ Sample( input=dedent( """ a: test good sentence 0 b: test bad sentence 0 """ ).strip(), output="a", ), Sample( input=dedent( """ a: test good sentence 1 b: test bad sentence 1 """ ).strip(), output="a", ), ] assert len(samples) == len(expected) for i, sample in enumerate(samples): assert isinstance(sample, Sample) assert sample.input == expected[i].input assert sample.output == expected[i].output
python
Apache-2.0
a5b71613317dfc051714585740f5a2be006ea1d0
2026-01-05T07:13:14.125338Z
false
llm-jp/llm-jp-eval
https://github.com/llm-jp/llm-jp-eval/blob/a5b71613317dfc051714585740f5a2be006ea1d0/tests/exporter/__init__.py
tests/exporter/__init__.py
python
Apache-2.0
a5b71613317dfc051714585740f5a2be006ea1d0
2026-01-05T07:13:14.125338Z
false
llm-jp/llm-jp-eval
https://github.com/llm-jp/llm-jp-eval/blob/a5b71613317dfc051714585740f5a2be006ea1d0/tests/exporter/test_exporter.py
tests/exporter/test_exporter.py
from unittest.mock import MagicMock import pytest from llm_jp_eval.exporter.base import EvaluationExporter, ExportConfigBase, OutputRecord class DummyConfig(ExportConfigBase): exporter_type = "dummy" @property def exporter_class(self): return DummyExporter class DummyExporter(EvaluationExporter[DummyConfig]): def initialize(self) -> None: pass def output_results(self, result) -> None: pass def finalize(self) -> None: pass @pytest.fixture def sample_records(): return [ OutputRecord( run_name="test_run", basemodel_name="test_model", target_dataset="dataset1", num_few_shots=0, target_split="test", id=str(i), input="test input", output="test output", pred="test pred", true="test true", prompt="test prompt", exact=1, char_f1=0.9, ) for i in range(5) ] + [ OutputRecord( run_name="test_run", basemodel_name="test_model", target_dataset="dataset2", num_few_shots=0, target_split="test", id=str(i), input="test input", output="test output", pred="test pred", true="test true", prompt="test prompt", exact=1, char_f1=0.9, ) for i in range(5) ] @pytest.mark.parametrize( "export_output_table,top_n,expected_count", [ (False, 2, 0), # 出力無効の場合は0件 (True, 2, 4), # dataset1とdataset2からそれぞれ2件ずつ (True, 10, 10), # 各データセット5件ずつで合計10件 ], ) def test_filter_records(sample_records, export_output_table, top_n, expected_count): config = DummyConfig(export_output_table=export_output_table, output_top_n=top_n) eval_config = MagicMock() exporter = DummyExporter(eval_config, config) filtered = exporter._filter_records(sample_records) assert len(filtered) == expected_count if export_output_table: dataset_counts = {} for record in filtered: dataset_counts[record.target_dataset] = dataset_counts.get(record.target_dataset, 0) + 1 for count in dataset_counts.values(): assert count <= top_n
python
Apache-2.0
a5b71613317dfc051714585740f5a2be006ea1d0
2026-01-05T07:13:14.125338Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/__init__.py
rl_gp_mpc/__init__.py
from .control_objects.controllers.gp_mpc_controller import GpMpcController from .visu_objects.visu_object import ControlVisualizations
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/run_env_function.py
rl_gp_mpc/run_env_function.py
import time import numpy as np import matplotlib.pyplot as plt from gym.core import Env from rl_gp_mpc import GpMpcController from rl_gp_mpc import ControlVisualizations from rl_gp_mpc.config_classes.visu_config import VisuConfig from rl_gp_mpc.config_classes.total_config import Config NUM_DECIMALS_REPR = 3 np.set_printoptions(precision=NUM_DECIMALS_REPR, suppress=True) def run_env(env: Env, control_config:Config, visu_config: VisuConfig, random_actions_init=10, num_steps=150, verbose=True): visu_obj = ControlVisualizations(env=env, num_steps=num_steps, control_config=control_config, visu_config=visu_config) ctrl_obj = GpMpcController(observation_low=env.observation_space.low, observation_high=env.observation_space.high, action_low=env.action_space.low, action_high=env.action_space.high, config=control_config) obs = env.reset() for idx_ctrl in range(num_steps): action_is_random = (idx_ctrl < random_actions_init) action = ctrl_obj.get_action(obs_mu=obs, random=action_is_random) iter_info = ctrl_obj.get_iter_info() cost, cost_var = ctrl_obj.compute_cost_unnormalized(obs, action) visu_obj.update(obs=obs, reward=-cost, action=action, env=env, iter_info=iter_info) obs_new, reward, done, info = env.step(action) ctrl_obj.add_memory(obs=obs, action=action, obs_new=obs_new, reward=-cost, predicted_state=iter_info.predicted_states[1], predicted_state_std=iter_info.predicted_states_std[1]) obs = obs_new if verbose: print(str(iter_info)) visu_obj.save(ctrl_obj) ctrl_obj.check_and_close_processes() env.__exit__() visu_obj.close() return visu_obj.get_costs() def run_env_multiple(env, env_name, control_config:Config, visu_config: VisuConfig, num_runs, random_actions_init=10, num_steps=150, verbose=True): costs_runs = [] for run_idx in range(num_runs): costs_iter = run_env(env, control_config, visu_config, random_actions_init, num_steps, verbose=verbose) costs_runs.append(costs_iter) time.sleep(1) costs_runs = np.array(costs_runs) costs_runs_mean = np.mean(costs_runs, axis=0) costs_runs_std = np.std(costs_runs, axis=0) x_axis = np.arange(len(costs_runs_mean)) fig, ax = plt.subplots(figsize=(10, 5)) ax.plot(x_axis, costs_runs_mean) ax.fill_between(x_axis, costs_runs_mean - costs_runs_std, costs_runs_mean + costs_runs_std, alpha=0.4) plt.title(f"Costs of multiples {env_name} runs") plt.ylabel("Cost") plt.xlabel("Env iteration") plt.savefig(f'multiple_runs_costs_{env_name}.png') plt.show()
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/visu_objects/visu_object.py
rl_gp_mpc/visu_objects/visu_object.py
import os import multiprocessing import copy import time import numpy as np from gym.wrappers.monitoring.video_recorder import VideoRecorder from gym.core import Env from rl_gp_mpc.config_classes.visu_config import VisuConfig from rl_gp_mpc.config_classes.total_config import Config from rl_gp_mpc.control_objects.controllers.iteration_info_class import IterationInformation from rl_gp_mpc.control_objects.controllers.gp_mpc_controller import GpMpcController from .utils import get_env_name, create_folder_save from .static_3d_graph import save_plot_model_3d from .static_2d_graph import save_plot_2d from .dynamic_2d_graph import LivePlotParallel class ControlVisualizations: def __init__(self, env:Env, num_steps:int, control_config: Config, visu_config: VisuConfig): self.control_config = control_config self.visu_config = visu_config self.states = [] self.actions = [] self.rewards = [] self.model_iter_infos = [] self.rec = None self.use_thread = False self.env_str = get_env_name(env) self.folder_save = create_folder_save(self.env_str) self.obs_min = env.observation_space.low self.obs_max = env.observation_space.high self.action_min = env.action_space.low self.action_max = env.action_space.high self.init = True if self.visu_config.render_live_plot_2d: self.live_plot_obj = LivePlotParallel(num_steps, use_constraints=bool(self.control_config.reward.use_constraints), dim_states=len(self.obs_min), dim_actions=len(self.action_min), state_min=self.control_config.reward.state_min, state_max=self.control_config.reward.state_max, use_thread=self.use_thread, path_save=os.path.join(self.folder_save, 'control_animation.mp4'), save=self.visu_config.save_live_plot_2d) if self.visu_config.save_render_env: self.rec = VideoRecorder(env, path=os.path.join(self.folder_save, 'gym_animation.mp4')) self.processes_running = True def update(self, obs:np.ndarray, action:np.ndarray, reward:float, env:Env, iter_info:IterationInformation=None): state = (obs - self.obs_min) / (self.obs_max - self.obs_min) action_norm = (action - self.action_min) / (self.action_max - self.action_min) self.states.append(state) self.actions.append(action_norm) self.rewards.append(reward) iter_info_arrays = copy.deepcopy(iter_info) iter_info_arrays.to_arrays() self.model_iter_infos.append(iter_info_arrays) if self.visu_config.render_live_plot_2d: self.live_plot_obj.update(state=state, action=action_norm, cost=-reward, iter_info=iter_info_arrays) self.env_render_step(env) if self.init: time.sleep(2) self.init = False def env_render_step(self, env): if self.visu_config.render_env: env.render() if self.visu_config.save_render_env: self.rec.capture_frame() def save_plot_model_3d(self, controller: GpMpcController): save_plot_model_3d(controller.transition_model.save_state(), self.folder_save, controller.memory) def save_plot_2d(self): save_plot_2d(np.array(self.states), np.array(self.actions), -np.array(self.rewards), self.model_iter_infos, self.folder_save, self.control_config.reward.use_constraints, self.control_config.reward.state_min, self.control_config.reward.state_max) def save(self, controller: GpMpcController): self.save_plot_2d() self.save_plot_model_3d(controller) def close(self): if self.visu_config.save_render_env: self.rec.close() self.close_running_processes() def close_running_processes(self): if 'live_plot_obj' in self.__dict__ and self.live_plot_obj.graph_p.is_alive(): self.live_plot_obj.close() self.processes_running = False def __exit__(self): if self.processes_running: self.close() def get_costs(self): return -np.array(self.rewards)
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/visu_objects/static_2d_graph.py
rl_gp_mpc/visu_objects/static_2d_graph.py
import datetime import os import matplotlib.pyplot as plt import numpy as np from rl_gp_mpc.control_objects.controllers.iteration_info_class import IterationInformation ALPHA_CONFIDENCE_BOUNDS = 0.4 UPDATE_X_LIMIT = True MUL_STD_BOUNDS = 3 def save_plot_2d(states, actions, costs, model_iter_infos: "list[IterationInformation]", folder_save, use_constraints=False, state_min_constraint=None, state_max_constraint=None, iter_ahead_show=3): fig, axes = plt.subplots(nrows=3, figsize=(6, 5), sharex=True) axes[0].set_title('Normed states and predictions') axes[1].set_title('Normed actions') axes[2].set_title('Cost and horizon cost') plt.xlabel("time") max_state_plot = np.max([states.max(), 1.03]) min_state_plot = np.min([states.min(), -0.03]) axes[0].set_ylim(min_state_plot, max_state_plot) axes[1].set_ylim(0, 1.02) plt.tight_layout() fig, axes = draw_plots_2d(states=states, actions=actions, costs=costs, model_iter_infos=model_iter_infos, use_constraints=use_constraints, state_min_constraint=state_min_constraint, state_max_constraint=state_max_constraint, iter_ahead_show=iter_ahead_show, fig=fig, axes=axes) axes[0].legend() axes[0].grid() axes[1].legend() axes[1].grid() axes[2].legend() axes[2].grid() fig.savefig(os.path.join(folder_save, f'history_{datetime.datetime.now().strftime("%Y_%m_%d_%H_%M_%S")}.png')) plt.close() def draw_plots_2d(states, actions, costs, model_iter_infos: "list[IterationInformation]"=None, use_constraints=False, state_min_constraint=None, state_max_constraint=None, iter_ahead_show=15, fig=None, axes=None, ): obs_idxs = np.arange(0, len(states)) colors = plt.get_cmap('tab20').colors if model_iter_infos is not None: pred_idxs = np.array([info.iteration for info in model_iter_infos]) pred_mean_costs = np.array([info.mean_predicted_cost for info in model_iter_infos]) pred_mean_cost_std = np.array([info.mean_predicted_cost_std for info in model_iter_infos]) states_pred_ahead_idxs = np.array([info.predicted_idxs[iter_ahead_show] for info in model_iter_infos]) states_pred_ahead = np.array([info.predicted_states[iter_ahead_show] for info in model_iter_infos]) states_pred_std_ahead = np.array([info.predicted_states_std[iter_ahead_show] for info in model_iter_infos]) for state_idx in range(len(states[0])): axes[0].plot(obs_idxs, np.array(states)[:, state_idx], label='state' + str(state_idx), color=colors[2 * state_idx]) if model_iter_infos is not None: axes[0].plot(states_pred_ahead_idxs, states_pred_ahead[:, state_idx], label=str(iter_ahead_show) + 'step_' + 'prediction' + str(state_idx), color=colors[2 * state_idx], linestyle='--') axes[0].fill_between(states_pred_ahead_idxs, states_pred_ahead[:, state_idx] - MUL_STD_BOUNDS * states_pred_std_ahead[:, state_idx], states_pred_ahead[:, state_idx] + MUL_STD_BOUNDS * states_pred_std_ahead[:, state_idx], color=colors[2 * state_idx], alpha=ALPHA_CONFIDENCE_BOUNDS) if use_constraints and (state_min_constraint is not None) and (state_max_constraint is not None): axes[0].axhline(y=state_min_constraint[state_idx], color=colors[2 * state_idx], linestyle='-.') axes[0].axhline(y=state_max_constraint[state_idx], color=colors[2 * state_idx], linestyle='-.') axes[2].plot(obs_idxs, costs, label='cost', color='k') axes[2].plot(pred_idxs, pred_mean_costs, label='mean predicted cost', color='orange') axes[2].fill_between(pred_idxs, pred_mean_costs - MUL_STD_BOUNDS * pred_mean_cost_std, pred_mean_costs + MUL_STD_BOUNDS * pred_mean_cost_std, color='orange', alpha=ALPHA_CONFIDENCE_BOUNDS) for idx_action in range(len(actions[0])): axes[1].step(obs_idxs, np.array(actions)[:, idx_action], label='a_' + str(idx_action), color=colors[1 + 2 * idx_action]) if UPDATE_X_LIMIT is True: axes[2].set_xlim(0, np.max(obs_idxs)) axes[2].set_ylim(0, np.max([np.max(pred_mean_costs), np.max(costs)]) * 1.2) plt.xlabel('Env steps') return fig, axes
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/visu_objects/dynamic_2d_graph.py
rl_gp_mpc/visu_objects/dynamic_2d_graph.py
import time import threading import queue import multiprocessing import matplotlib import matplotlib.pyplot as plt import numpy as np import imageio from rl_gp_mpc.control_objects.controllers.iteration_info_class import IterationInformation # matplotlib.use('TkAgg') matplotlib.rc('font', size='6') ALPHA_CONFIDENCE_BOUNDS = 0.3 FONTSIZE = 6 MUL_STD_BOUNDS = 3 FONTSIZE = 6 class LivePlotParallel: def __init__(self, num_steps_total, dim_states, dim_actions, use_constraints=False, state_min=None, state_max=None, time_between_updates=0.75, use_thread=False, path_save="control_animation.gif", save=False): self.use_thread = use_thread if use_thread: self.pqueue = queue.Queue() self.graph_p = threading.Thread(target=anim_plots_2d_p, args=(self.pqueue, num_steps_total, dim_states, dim_actions, use_constraints, state_min, state_max, time_between_updates, path_save, save)) else: self.pqueue = multiprocessing.Queue() self.graph_p = multiprocessing.Process(target=anim_plots_2d_p, args=(self.pqueue, num_steps_total, dim_states, dim_actions, use_constraints, state_min, state_max, time_between_updates, path_save, save)) self.graph_p.start() def update(self, state, action, cost, iter_info:IterationInformation=None): if iter_info is None: self.pqueue.put([state, action, cost]) else: self.pqueue.put([state, action, cost, iter_info.mean_predicted_cost, iter_info.mean_predicted_cost_std, iter_info.predicted_states, iter_info.predicted_states_std, iter_info.predicted_actions, iter_info.predicted_costs, iter_info.predicted_costs_std, iter_info.predicted_idxs]) def close(self): self.pqueue.put(None) if self.use_thread: self.graph_p.join() else: self.pqueue.close() self.graph_p.join() self.graph_p.close() def __exit__(self): if 'graph_p' in self.__dict__ and self.graph_p.is_alive(): self.close() def anim_plots_2d_p(queue, num_steps_total, dim_states, dim_actions, use_constraints=False, state_min=None, state_max=None, time_between_updates=0.5, path_save="control_animation.gif", save=False): fig, axes = plt.subplots(nrows=3, figsize=(10, 6), sharex=True) axes[0].set_title('Normed states and predictions') axes[1].set_title('Normed actions') axes[2].set_title('Cost and horizon cost') plt.xlabel("Env steps") min_state = -0.03 max_state = 1.03 axes[0].set_ylim(min_state, max_state) axes[0].grid() axes[1].set_ylim(-0.03, 1.03) axes[1].grid() axes[2].set_xlim(0, num_steps_total) axes[2].grid() plt.tight_layout() states = np.empty((num_steps_total, dim_states)) actions = np.empty((num_steps_total, dim_actions)) costs = np.empty(num_steps_total) mean_cost_pred = np.empty_like(costs) mean_cost_std_pred = np.empty_like(mean_cost_pred) if use_constraints: if state_min is not None: for idx_state in range(dim_states): axes[0].axhline(y=state_min[idx_state], color=plt.get_cmap('tab20').colors[2 * idx_state], linestyle='-.') if state_max is not None: for idx_state in range(dim_states): axes[0].axhline(y=state_max[idx_state], color=plt.get_cmap('tab20').colors[2 * idx_state], linestyle='-.') lines_states = [axes[0].plot([], [], label='state' + str(state_idx), color=plt.get_cmap('tab20').colors[2 * state_idx]) for state_idx in range(dim_states)] lines_actions = [axes[1].step([], [], label='action' + str(action_idx), color=plt.get_cmap('tab20').colors[1 + 2 * action_idx], where='post') for action_idx in range(dim_actions)] line_cost = axes[2].plot([], [], label='cost', color='k') line_costs_mean_pred = axes[2].plot([], [], label='mean predicted cost', color='orange') lines_predicted_states = [axes[0].plot([], [], label='predicted_states' + str(state_idx), color=plt.get_cmap('tab20').colors[2 * state_idx], linestyle='dashed') for state_idx in range(dim_states)] lines_actions_pred = [axes[1].step([], [], label='predicted_action' + str(action_idx), color=plt.get_cmap('tab20').colors[1 + 2 * action_idx], linestyle='dashed', where='post') for action_idx in range(dim_actions)] line_predicted_costs = axes[2].plot([], [], label='predicted cost', color='k', linestyle='dashed') axes[0].legend(fontsize=FONTSIZE) axes[1].legend(fontsize=FONTSIZE) axes[2].legend(fontsize=FONTSIZE) fig.canvas.draw() plt.show(block=False) num_pts_show = 0 last_update_time = time.time() exit_loop = False if save: writer = imageio.get_writer(path_save, mode='I') while True: if exit_loop: break if (time.time() - last_update_time) > time_between_updates: got_data = False last_update_time = time.time() while not (queue.empty()): msg = queue.get() # Read from the queue if msg is None: # Allows to gracefully end the process by send None, then join and close exit_loop = True continue obs_norm = np.array(msg[0]) action_norm = np.array(msg[1]) states[num_pts_show] = obs_norm actions[num_pts_show] = action_norm costs[num_pts_show] = msg[2] update_limits = False min_state_actual = np.min(states[num_pts_show]) if min_state_actual < min_state: min_state = min_state_actual update_limits = True max_state_actual = np.max(states[num_pts_show]) if max_state_actual > max_state: max_state = max_state_actual update_limits = True if update_limits: axes[0].set_ylim(min_state, max_state) if len(msg) > 3: mean_cost_pred[num_pts_show] = np.nan_to_num(msg[3], nan=-1, posinf=-1, neginf=-1) mean_cost_std_pred[num_pts_show] = np.nan_to_num(msg[4], copy=True, nan=99, posinf=99, neginf=99) else: if num_pts_show == 0: # Init values for mean cost pred and mean cost std pred at the beginning when the control # has not started yet and only random actions has been applied. # Thus, the mean future cost and the std of the mean future cost has not been computed yet # and have not been provided to the graph object for the update mean_cost_pred[num_pts_show] = 0 mean_cost_std_pred[num_pts_show] = 0 else: # For when the information about prediction have not been provided to the graph object # for the update, for example at the init period when random actions are applied mean_cost_pred[num_pts_show] = mean_cost_pred[num_pts_show - 1] mean_cost_std_pred[num_pts_show] = mean_cost_std_pred[num_pts_show - 1] num_pts_show += 1 got_data = True last_update_time = time.time() if got_data: for idx_axes in range(len(axes)): axes[idx_axes].collections.clear() idxs = np.arange(0, num_pts_show) for idx_axes in range(len(axes)): axes[idx_axes].collections.clear() for idx_state in range(len(obs_norm)): lines_states[idx_state][0].set_data(idxs, states[:num_pts_show, idx_state]) for idx_action in range(len(action_norm)): lines_actions[idx_action][0].set_data(idxs, actions[:num_pts_show, idx_action]) line_cost[0].set_data(idxs, costs[:num_pts_show]) line_costs_mean_pred[0].set_data(idxs, mean_cost_pred[:num_pts_show]) axes[2].fill_between(idxs, mean_cost_pred[:num_pts_show] - mean_cost_std_pred[:num_pts_show] * MUL_STD_BOUNDS, mean_cost_pred[:num_pts_show] + mean_cost_std_pred[:num_pts_show] * MUL_STD_BOUNDS, facecolor='orange', alpha=ALPHA_CONFIDENCE_BOUNDS, label='mean predicted ' + str(MUL_STD_BOUNDS) + ' std cost bounds') axes[2].set_ylim(0, np.max([np.max(mean_cost_pred[:num_pts_show]), np.max(costs[:num_pts_show])]) * 1.1) axes[2].set_ylim(0, np.max(line_cost[0].get_ydata()) * 1.1) # If predictions are not given in the last update, do not show them on the graph if len(msg) > 3: predicted_states = np.nan_to_num(msg[5], nan=-1, posinf=-1, neginf=-1) predicted_states_std = np.nan_to_num(msg[6], copy=False, nan=99, posinf=99, neginf=99) actions_pred = msg[7] predicted_costs = np.nan_to_num(msg[8], nan=-1, posinf=-1, neginf=-1) predicted_costs_std = np.nan_to_num(msg[9], nan=99, posinf=99, neginf=0) predicted_idxs = np.array(msg[10]) step_idx = (predicted_idxs[-1] - predicted_idxs[-2]) predicted_idxs_states = np.concatenate((predicted_idxs, [predicted_idxs[-1] + step_idx])) for idx_state in range(len(obs_norm)): future_states_show = predicted_states[:, idx_state] lines_predicted_states[idx_state][0].set_data(predicted_idxs_states, future_states_show) future_states_std_show = predicted_states_std[:, idx_state] axes[0].fill_between(predicted_idxs_states, future_states_show - future_states_std_show * MUL_STD_BOUNDS, future_states_show + future_states_std_show * MUL_STD_BOUNDS, facecolor=plt.get_cmap('tab20').colors[2 * idx_state], alpha=ALPHA_CONFIDENCE_BOUNDS, label='predicted ' + str(MUL_STD_BOUNDS) + ' std bounds state ' + str(idx_state)) for idx_action in range(len(action_norm)): future_actions_show = actions_pred[:, idx_action] lines_actions_pred[idx_action][0].set_data(predicted_idxs, future_actions_show) # future_costs_show = predicted_costs line_predicted_costs[0].set_data(predicted_idxs_states, future_costs_show) future_cost_std_show = predicted_costs_std axes[2].fill_between(predicted_idxs_states, future_costs_show - future_cost_std_show * MUL_STD_BOUNDS, future_costs_show + future_cost_std_show * MUL_STD_BOUNDS, facecolor='black', alpha=ALPHA_CONFIDENCE_BOUNDS, label='predicted ' + str(MUL_STD_BOUNDS) + ' std cost bounds') for ax in axes: ax.relim() ax.autoscale_view(True, True, True) fig.canvas.draw() if save: image = np.frombuffer(fig.canvas.tostring_rgb(), dtype='uint8') image = image.reshape(fig.canvas.get_width_height()[::-1] + (3,)) # Append the image to the animation writer.append_data(image) plt.pause(0.01) time.sleep(0.05) if save: writer.close() plt.close('all') class LivePlotSequential: ''' DEPRECATED CLASS, WILL NOT WORK ''' def __init__(self, num_steps_total, obs_space, action_space, step_pred, use_constraints=False, state_min=None, state_max=None): self.fig, self.axes = plt.subplots(nrows=3, figsize=(6, 5), sharex=True) self.axes[0].set_title('Normed states and predictions') self.axes[1].set_title('Normed actions') self.axes[2].set_title('Cost and horizon cost') plt.xlabel("Env steps") self.min_state = -0.03 self.max_state = 1.03 self.axes[0].set_ylim(self.min_state, self.max_state) self.axes[1].set_ylim(-0.03, 1.03) self.axes[2].set_xlim(0, num_steps_total) plt.tight_layout() self.step_pred = step_pred self.states = np.empty((num_steps_total, obs_space.shape[0])) self.actions = np.empty((num_steps_total, action_space.shape[0])) self.costs = np.empty(num_steps_total) self.mean_predicted_cost = np.empty_like(self.costs) self.mean_predicted_cost_std = np.empty_like(self.costs) self.min_obs = obs_space.low self.max_obs = obs_space.high self.min_action = action_space.low self.max_action = action_space.high self.num_points_show = 0 if use_constraints: if state_min is not None: for state_idx in range(obs_space.shape[0]): self.axes[0].axhline(y=state_min[state_idx], color=plt.get_cmap('tab20').colors[2 * state_idx], linestyle='-.') if state_max is not None: for state_idx in range(obs_space.shape[0]): self.axes[0].axhline(y=state_max[state_idx], color=plt.get_cmap('tab20').colors[2 * state_idx], linestyle='-.') self.lines_states = [self.axes[0].plot([], [], label='state' + str(state_idx), color=plt.get_cmap('tab20').colors[2 * state_idx]) for state_idx in range(obs_space.shape[0])] self.line_cost = self.axes[2].plot([], [], label='cost', color='k') self.lines_actions = [self.axes[1].step([], [], label='action' + str(action_idx), color=plt.get_cmap('tab20').colors[1 + 2 * action_idx]) for action_idx in range(action_space.shape[0])] self.line_mean_predicted_cost = self.axes[2].plot([], [], label='mean predicted cost', color='orange') self.lines_predicted_states = [self.axes[0].plot([], [], label='predicted_states' + str(state_idx), color=plt.get_cmap('tab20').colors[2 * state_idx], linestyle='dashed') for state_idx in range(obs_space.shape[0])] self.lines_actions_pred = [self.axes[1].step([], [], label='predicted_action' + str(action_idx), color=plt.get_cmap('tab20').colors[1 + 2 * action_idx], linestyle='dashed') for action_idx in range(action_space.shape[0])] self.line_predicted_costs = self.axes[2].plot([], [], label='predicted cost', color='k', linestyle='dashed') self.axes[0].legend(fontsize=FONTSIZE) self.axes[0].grid() self.axes[1].legend(fontsize=FONTSIZE) self.axes[1].grid() self.axes[2].legend(fontsize=FONTSIZE) self.axes[2].grid() plt.show(block=False) def update(self, obs, action, cost, iter_info:IterationInformation=None): obs_norm = (obs - self.min_obs) / (self.max_obs - self.min_obs) action_norm = (action - self.min_action) / (self.max_action - self.min_action) self.states[self.num_points_show] = obs_norm self.costs[self.num_points_show] = cost update_limits = False min_state_actual = np.min(obs_norm) if min_state_actual < self.min_state: self.min_state = min_state_actual update_limits = True max_state_actual = np.max(obs_norm) if max_state_actual > self.max_state: self.max_state = max_state_actual update_limits = True if update_limits: self.axes[0].set_ylim(self.min_state, self.max_state) idxs = np.arange(0, (self.num_points_show + 1)) for idx_axes in range(len(self.axes)): self.axes[idx_axes].collections.clear() for idx_state in range(len(obs_norm)): self.lines_states[idx_state][0].set_data(idxs, self.states[:(self.num_points_show + 1), idx_state]) self.actions[self.num_points_show] = action_norm for idx_action in range(len(action_norm)): self.lines_actions[idx_action][0].set_data(idxs, self.actions[:(self.num_points_show + 1), idx_action]) self.line_cost[0].set_data(idxs, self.costs[:(self.num_points_show + 1)]) if iter_info is not None: np.nan_to_num(iter_info.mean_predicted_cost, copy=False, nan=-1, posinf=-1, neginf=-1) np.nan_to_num(iter_info.mean_predicted_cost_std, copy=False, nan=99, posinf=99, neginf=99) np.nan_to_num(iter_info.predicted_states, copy=False, nan=-1, posinf=-1, neginf=-1) np.nan_to_num(iter_info.predicted_states_std, copy=False, nan=99, posinf=99, neginf=0) np.nan_to_num(iter_info.predicted_costs, copy=False, nan=-1, posinf=-1, neginf=-1) np.nan_to_num(iter_info.predicted_costs_std, copy=False, nan=99, posinf=99, neginf=0) # if num_repeat_action is not 1, the control do not happen at each iteration, # we must select the last index where the control happened as the start of the prediction horizon idx_prev_control = (idxs[-1] // self.step_pred) * self.step_pred idxs_future = iter_info.predicted_idxs #np.arange(idx_prev_control, # idx_prev_control + self.step_pred + len(iter_info.predicted_states) * self.step_pred, # self.step_pred) self.mean_predicted_cost[self.num_points_show] = iter_info.mean_predicted_cost self.mean_predicted_cost_std[self.num_points_show] = iter_info.mean_predicted_cost_std for idx_state in range(len(obs_norm)): future_states_show = np.concatenate(([self.states[idx_prev_control, idx_state]], iter_info.predicted_states[:, idx_state])) self.lines_predicted_states[idx_state][0].set_data(idxs_future, future_states_show) future_states_std_show = np.concatenate(([0], iter_info.predicted_states_std[:, idx_state])) self.axes[0].fill_between(idxs_future, future_states_show - future_states_std_show * MUL_STD_BOUNDS, future_states_show + future_states_std_show * MUL_STD_BOUNDS, facecolor=plt.get_cmap('tab20').colors[2 * idx_state], alpha=ALPHA_CONFIDENCE_BOUNDS, label='predicted ' + str(MUL_STD_BOUNDS) + ' std bounds state ' + str(idx_state)) for idx_action in range(len(action_norm)): self.lines_actions_pred[idx_action][0].set_data(idxs_future, iter_info.actions_pred[:, idx_action]) #) np.concatenate(([self.actions[idx_prev_control, idx_action]], ) future_costs_show = np.concatenate(([self.costs[idx_prev_control]], iter_info.predicted_costs)) self.line_predicted_costs[0].set_data(idxs_future, future_costs_show) future_cost_std_show = np.concatenate(([0], iter_info.predicted_costs_std)) self.axes[2].fill_between(idxs_future, future_costs_show - future_cost_std_show * MUL_STD_BOUNDS, future_costs_show + future_cost_std_show * MUL_STD_BOUNDS, facecolor='black', alpha=ALPHA_CONFIDENCE_BOUNDS, label='predicted ' + str(MUL_STD_BOUNDS) + ' std cost bounds') else: if self.num_points_show == 0: self.mean_predicted_cost[self.num_points_show] = MEAN_PRED_COST_INIT self.mean_predicted_cost_std[self.num_points_show] = STD_MEAN_PRED_COST_INIT else: self.mean_predicted_cost[self.num_points_show] = self.mean_predicted_cost[self.num_points_show - 1] self.mean_predicted_cost_std[self.num_points_show] = self.mean_predicted_cost_std[self.num_points_show - 1] self.line_mean_predicted_cost[0].set_data(idxs, self.mean_predicted_cost[:(self.num_points_show + 1)]) self.axes[2].set_ylim(0, np.max([np.max(self.mean_predicted_cost[:(self.num_points_show + 1)]), np.max(self.costs[:(self.num_points_show + 1)])]) * 1.1) self.axes[2].fill_between(idxs, self.mean_predicted_cost[:(self.num_points_show + 1)] - self.mean_predicted_cost_std[:(self.num_points_show + 1)] * MUL_STD_BOUNDS, self.mean_predicted_cost[:(self.num_points_show + 1)] + self.mean_predicted_cost_std[:(self.num_points_show + 1)] * MUL_STD_BOUNDS, facecolor='orange', alpha=ALPHA_CONFIDENCE_BOUNDS, label='mean predicted ' + str(MUL_STD_BOUNDS) + ' std cost bounds') self.fig.canvas.draw() plt.pause(0.01) self.num_points_show += 1
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/visu_objects/static_3d_graph.py
rl_gp_mpc/visu_objects/static_3d_graph.py
import os import datetime import torch import gpytorch import matplotlib.pyplot as plt from sklearn.pipeline import Pipeline from sklearn.preprocessing import StandardScaler from sklearn.neighbors import KNeighborsRegressor import numpy as np from rl_gp_mpc.control_objects.models.gp_model import create_models, SavedState from rl_gp_mpc.control_objects.memories.gp_memory import Memory PROP_EXTEND_DOMAIN = 1 N_TICKS = 75 TOTAL_COL_MAX= 3 FONTSIZE = 6 def save_plot_model_3d(saved_state: SavedState, folder_save, memory:Memory, plot_points_memory=True): models = create_models(saved_state.parameters, saved_state.constraints_hyperparams, saved_state.inputs, saved_state.states_change) inputs_total, targets_total = memory.get_memory_total() active_memory_mask = memory.get_mask_model_inputs() inputs_model = inputs_total[active_memory_mask] targets_model = targets_total[active_memory_mask] with torch.no_grad(), gpytorch.settings.fast_pred_var(): torch.set_num_threads(1) num_input_model = len(inputs_model[0]) num_models = len(targets_model[0]) for idx_model in range(len(models)): models[idx_model].eval() targets_model = targets_model.numpy() num_figures = int(num_input_model / TOTAL_COL_MAX + 0.5) figs = [] axes_s = [] for index_figure in range(num_figures): fig = plt.figure(figsize=(15, 6)) axes = [] columns_active = np.min([num_models - (TOTAL_COL_MAX * index_figure), TOTAL_COL_MAX]) for idx_ax in range(columns_active * 2): axes.append(fig.add_subplot(2, columns_active, idx_ax + 1, projection='3d')) axes_s.append(axes) figs.append(fig) for idx_subplot in range(columns_active): idx_obs_repr = index_figure * columns_active + idx_subplot if idx_obs_repr >= num_models: break feat_importance = (1 / models[idx_obs_repr].covar_module.base_kernel.lengthscale).numpy() feat_importance = feat_importance / np.sum(feat_importance) best_features = np.argsort(-feat_importance)[0, :2] estimator_other_columns = Pipeline( steps=[('standardscaler', StandardScaler()), ('features', KNeighborsRegressor(n_neighbors=3, weights='distance'))]) estimator_other_columns.fit(inputs_total[:, best_features].numpy(), np.delete(inputs_total.numpy(), best_features, axis=1)) domain_extension_x = (PROP_EXTEND_DOMAIN - 1) * ( inputs_total[:, best_features[0]].max() - inputs_total[:, best_features[0]].min()) domain_extension_y = (PROP_EXTEND_DOMAIN - 1) * ( inputs_total[:, best_features[1]].max() - inputs_total[:, best_features[1]].min()) x_grid = np.linspace(inputs_total[:, best_features[0]].min() - domain_extension_x / 2, inputs_total[:, best_features[0]].max() + domain_extension_x / 2, N_TICKS) y_grid = np.linspace(inputs_total[:, best_features[1]].min() - domain_extension_y / 2, inputs_total[:, best_features[1]].max() + domain_extension_y / 2, N_TICKS) x_grid, y_grid = np.meshgrid(x_grid, y_grid) x_grid_1d = np.expand_dims(x_grid.flatten(), axis=-1) y_grid_1d = np.expand_dims(y_grid.flatten(), axis=-1) xy_grid = np.concatenate((x_grid_1d, y_grid_1d), axis=1) pred_other_columns = estimator_other_columns.predict(xy_grid) all_col = np.zeros((xy_grid.shape[0], num_input_model)) all_col[:, best_features] = xy_grid all_col[:, np.delete(np.arange(num_input_model), best_features)] = pred_other_columns gauss_pred = models[idx_obs_repr].likelihood(models[idx_obs_repr](torch.from_numpy( all_col.astype(np.float64)))) z_grid_1d_mean = gauss_pred.mean.numpy() z_grid_1d_std = np.sqrt(gauss_pred.stddev.numpy()) z_grid_mean = z_grid_1d_mean.reshape(x_grid.shape) z_grid_std = z_grid_1d_std.reshape(x_grid.shape) surf1 = axes[idx_subplot].contour3D(x_grid, y_grid, z_grid_mean, N_TICKS, cmap='cool') axes[idx_subplot].set_xlabel('Input ' + str(best_features[0]), fontsize=FONTSIZE, rotation=150) axes[idx_subplot].set_ylabel('Input ' + str(best_features[1]), fontsize=FONTSIZE, rotation=150) axes[idx_subplot].set_zlabel('Variation state ' + str(idx_obs_repr), fontsize=FONTSIZE, rotation=60) surf2 = axes[idx_subplot + columns_active].contour3D(x_grid, y_grid, z_grid_std, N_TICKS, cmap='cool') axes[idx_subplot + columns_active].set_xlabel('Input ' + str(best_features[0]), fontsize=FONTSIZE, rotation=150) axes[idx_subplot + columns_active].set_ylabel('Input ' + str(best_features[1]), fontsize=FONTSIZE, rotation=150) axes[idx_subplot + columns_active].set_zlabel('Uncertainty: std state ' + str(idx_obs_repr), fontsize=FONTSIZE, rotation=60) if plot_points_memory: axes[idx_subplot].scatter( inputs_total[~active_memory_mask, best_features[0]], inputs_total[~active_memory_mask, best_features[1]], targets_total[~active_memory_mask, idx_obs_repr], marker='x', c='k') axes[idx_subplot].scatter( inputs_model[:, best_features[0]], inputs_model[:, best_features[1]], targets_model[:, idx_obs_repr], marker='x', c='g') axes[idx_subplot].quiver(inputs_total[:-1, best_features[0]], inputs_total[:-1, best_features[1]], targets_total[:-1, idx_obs_repr], inputs_total[1:, best_features[0]] - inputs_total[:-1, best_features[0]], inputs_total[1:, best_features[1]] - inputs_total[:-1, best_features[1]], targets_total[1:, idx_obs_repr] - targets_total[:-1, idx_obs_repr], color='k', linestyle="solid", alpha=0.3, arrow_length_ratio=0.001, length=0.9) preds_total = models[idx_obs_repr].likelihood(models[idx_obs_repr](inputs_total)) errors_total = np.abs(preds_total.mean.numpy() - targets_total[:, idx_obs_repr].numpy()) axes[idx_subplot + columns_active].scatter( inputs_total[~active_memory_mask, best_features[0]], inputs_total[~active_memory_mask, best_features[1]], errors_total[~active_memory_mask], marker='x', c='k') axes[idx_subplot + columns_active].scatter( inputs_total[active_memory_mask, best_features[0]], inputs_total[active_memory_mask, best_features[1]], errors_total[active_memory_mask], marker='x', c='g') axes[idx_subplot + columns_active].quiver( inputs_total[:-1, best_features[0]], inputs_total[:-1, best_features[1]], errors_total[:-1], inputs_total[1:, best_features[0]] - inputs_total[:-1, best_features[0]], inputs_total[1:, best_features[1]] - inputs_total[:-1, best_features[1]], errors_total[1:] - errors_total[:-1], color='k', linestyle="solid", alpha=0.3, arrow_length_ratio=0.001, length=0.9) plt.tight_layout() fig.savefig(os.path.join(folder_save, f'model_3d_{datetime.datetime.now().strftime("%Y_%m_%d_%H_%M_%S")}.png')) plt.close()
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/visu_objects/utils.py
rl_gp_mpc/visu_objects/utils.py
import datetime import os def get_env_name(env): try: env_name = env.env.spec.entry_point.replace('-', '_').replace(':', '_').replace('.', '_') except: env_name = env.name return env_name def create_folder_save(env_name): datetime_now = datetime.datetime.now() folder_save = os.path.join('folder_save', env_name, datetime_now.strftime("%Y_%m_%d_%H_%M_%S")) if not os.path.exists(folder_save): os.makedirs(folder_save) return folder_save
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/envs/process_control.py
rl_gp_mpc/envs/process_control.py
import numpy as np import gym from gym import spaces class ProcessControl(gym.Env): """ Example of custom Environment that follows gym interface. The additional methods: - define_params: used to be able to vary the internal parameters of the env to check the robustness of the control to change in the env. - get_obs: methods that returns the observation given the state of the env. It is used to add noise to measurement. For example, the state is the volume, but the level is measured If change_params is set to True, the internal params will be changed every period_change iterations with the env Params are ranges because they are chosen at random between the two values provided. """ metadata = {'render.modes': []} def __init__(self, dt=1, s_range=(9, 11), fi_range=(0., 0.2), ci_range=(0, 0.2), cr_range=(0.5, 1), noise_l_prop_range=(1e-5, 1e-3), noise_co_prop_range=(1e-5, 1e-3), sp_l_range=(0.2, 0.8), sp_co_range=(0.2, 0.4), change_params=True, period_change=50): """ The environment represents the simulation of a cuve with two input flow with different concentrations of product. The flow fi with concentration ci is not controllable The flow with concentration cr is controllable with action_1 action_0 controls the output flow state0 = level state1 = concentration of product s: surface of the cuve ci: concentration of product with flow fi cr: concentration of product for input input_1 There is a gaussian noise on the measurement of the level noise_l_prop: noise on level noise_co: noise on concentration The parameters are taken randomly between the ranges provided. If change_params is set to True, the parameters are changed every period_change timesteps. Can be used to see how the method is robust to environment changes dt: time between time steps """ super().__init__() self.name = 'processcontrol' # Definitions for gym env: action and observation space # They must be gym.spaces objects self.observation_space = spaces.Box( low=np.array([0, 0]), high=np.array([10, 1]), shape=(2,), dtype=np.float32) self.reward_range = (0, 1) self.action_space = spaces.Box( low=np.array([0, 0]), high=np.array([1, 1]), shape=(2,), dtype=np.float32) # definitions specific to my env self.dt = dt self.s_range = s_range self.fi_range = fi_range self.ci_range = ci_range self.cr_range = cr_range self.noise_l_prop_range = noise_l_prop_range self.noise_co_prop_range = noise_co_prop_range self.sp_l_range = sp_l_range self.sp_co_range = sp_co_range self.change_params = change_params self.period_change = period_change # initialization of parameters and states self.define_params() def define_params(self): """ Define the env parameters. It is not specific to gym """ self.s = np.random.uniform(self.s_range[0], self.s_range[1]) self.fi = np.random.uniform(self.fi_range[0], self.fi_range[1]) self.ci = np.random.uniform(self.ci_range[0], self.ci_range[1]) self.cr = np.random.uniform(self.cr_range[0], self.cr_range[1]) self.noise_l_prop = np.exp(np.random.uniform( np.log(self.noise_l_prop_range[0]), np.log(self.noise_l_prop_range[1]))) self.noise_co_prop = np.exp(np.random.uniform( np.log(self.noise_co_prop_range[0]), np.log(self.noise_co_prop_range[1]))) self.sp_l = np.random.uniform(self.sp_l_range[0], self.sp_l_range[1]) self.sp_co = np.random.uniform(self.sp_co_range[0], self.sp_co_range[1]) if hasattr(self, 'v'): self.clip_parameters() print("New params value: s: {:.2f}, fi: {:.2f}, ci: {:.2f}, cr: {:.2f}, " "noise_l: {:.4f}, noise_co: {:.4f}, sp_l: {:.2f}, sp_co: {:.2f}".format( self.s, self.fi, self.ci, self.cr, self.noise_l_prop, self.noise_co_prop, self.sp_l, self.sp_co)) def step(self, action): dv = (self.fi + action[1] - action[0]) dr = (self.fi * self.ci + action[1] * self.cr - action[0] * self.r / (self.v + 1e-3)) self.v += dv * self.dt self.r += dr * self.dt self.iter += 1 self.v = np.clip(self.v, self.observation_space.low[0] * self.s, self.observation_space.high[0] * self.s) self.r = np.clip(self.r, self.observation_space.low[1] * self.v, self.observation_space.high[1] * self.v) reward = - (pow(self.v / self.s - self.sp_l, 2) + pow(self.r / (self.v + 1e-6) - self.sp_co, 2)) done = 0 info = {} if self.change_params and self.iter % self.period_change == 0: self.define_params() return self.get_obs(), reward, done, info def reset(self, min_prop=0.3, max_prop=0.7): # Reset the state of the environment to an initial state self.iter = 0 ranges = (self.observation_space.high - self.observation_space.low) obs_sample = np.clip(self.observation_space.sample(), min_prop * ranges + self.observation_space.low, max_prop * ranges + self.observation_space.low) self.v = obs_sample[0] * self.s self.r = obs_sample[1] * self.v return self.get_obs() def get_obs(self): """ Get observation given the internal states. It is not specific to gym """ l_mes = self.v / self.s co_mes = self.r / (self.v + 1e-6) if self.noise_l_prop != 0: l_mes += np.random.normal(0, self.noise_l_prop * self.observation_space.high[0]) if self.noise_co_prop != 0: co_mes += np.random.normal(0, self.noise_co_prop * self.observation_space.high[1]) l_mes = np.clip(l_mes, self.observation_space.low[0], self.observation_space.high[0]) co_mes = np.clip(co_mes, self.observation_space.low[1], self.observation_space.high[1]) return np.array([l_mes, co_mes]) def render(self, mode='human', close=False): # Render the environment pass def clip_parameters(self, prop_level_max_after_reset=0.9): v_p = self.v self.v = np.clip(self.v, a_min=0, a_max=prop_level_max_after_reset * (self.s * self.observation_space.high[0]) ) self.r = self.r * self.v / v_p
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/config_classes/memory_config.py
rl_gp_mpc/config_classes/memory_config.py
from .utils.functions_process_config import convert_config_lists_to_tensor class MemoryConfig: def __init__(self, check_errors_for_storage:bool=True, min_error_prediction_state_for_memory:"list[float]"=[3e-4, 3e-4, 3e-4], min_prediction_state_std_for_memory:"list[float]"=[3e-3, 3e-3, 3e-3], points_batch_memory:int=1500 ): """ check_errors_for_storage: If true, when adding a new point in memory, it will be checked if it is worth adding it to the model memory depending on the prediction by checking the error or uncertainty min_error_prediction_state_for_memory: if check_errors_for_storage is true, a point will only be used by the model if the error is above this threshold (any) min_prediction_state_std_for_memory: if check_errors_for_storage is true, a point will only be used by the model if the predicted standard deviation is above this threshold (any) """ self.check_errors_for_storage = check_errors_for_storage self.min_error_prediction_state_for_memory = min_error_prediction_state_for_memory self.min_prediction_state_std_for_memory = min_prediction_state_std_for_memory self.points_batch_memory = points_batch_memory self = convert_config_lists_to_tensor(self)
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/config_classes/actions_config.py
rl_gp_mpc/config_classes/actions_config.py
from .utils.functions_process_config import convert_config_lists_to_tensor class ActionsConfig: def __init__( self, limit_action_change:bool=False, max_change_action_norm:"list[float]"=[0.05], ): """ limit_action_change: if set to true, the actions will stay close to the action of the previous time step max_change_action_norm: determines the maximum change of the normalized action if limit_action_change is set to true """ self.limit_action_change = limit_action_change self.max_change_action_norm = max_change_action_norm self = convert_config_lists_to_tensor(self)
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/config_classes/visu_config.py
rl_gp_mpc/config_classes/visu_config.py
class VisuConfig: def __init__( self, save_render_env:bool=True, render_live_plot_2d:bool=True, render_env:bool=True, save_live_plot_2d:bool=False ): """ save_render_env: if set to true and render_env is true, will save the env animation render_live_plot_2d: if true, will show the dynamic 2d graph update in real time render_env: if set to true, will show the env save_live_plot_2d: if set to true and render_live_plot_2d is true, will save the 2d graph animation """ self.save_render_env = save_render_env self.render_live_plot_2d = render_live_plot_2d self.render_env = render_env self.save_live_plot_2d = save_live_plot_2d
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/config_classes/observation_config.py
rl_gp_mpc/config_classes/observation_config.py
import torch class ObservationConfig: def __init__( self, obs_var_norm:"list[float]"=[1e-6, 1e-6, 1e-6] ): """ obs_var_norm: uncertainty of the observation used by the model """ self.obs_var_norm = torch.diag(torch.Tensor(obs_var_norm))
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/config_classes/training_config.py
rl_gp_mpc/config_classes/training_config.py
class TrainingConfig: def __init__( self, lr_train:float=7e-3, iter_train:int=15, training_frequency:int=25, clip_grad_value:float=1e-3, print_train:bool=False, step_print_train:int=5 ): """ lr_train: learning rate when training the hyperparameters of the model iter_train: number of iteration when training the model training_frequency: the model will be trained periodically. The interval in number of control iteration is this parameter clip_grad_value: if the gradient is above this number, it will be clipped to that number. Allows for better stability print_train: if set to true, will print the training loop information step_print_train: if print_train is true, will print the info every step_print_train of the training iterations """ self.lr_train = lr_train self.iter_train = iter_train self.training_frequency = training_frequency self.clip_grad_value = clip_grad_value self.print_train = print_train self.step_print_train = step_print_train
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/config_classes/reward_config.py
rl_gp_mpc/config_classes/reward_config.py
import torch from .utils.functions_process_config import convert_config_lists_to_tensor class RewardConfig: def __init__( self, target_state_norm:"list[float]"=[1, 0.5, 0.5], weight_state:"list[float]"=[1, 0.1, 0.1], weight_state_terminal:"list[float]"=[10, 5, 5], target_action_norm:"list[float]"=[0.5], weight_action:"list[float]"=[0.05], exploration_factor:float=3, use_constraints:bool=False, state_min:"list[float]"=[-0.1, 0.05, 0.05], state_max:"list[float]"=[1.1, 0.95, 0.925], area_multiplier:float=1, clip_lower_bound_cost_to_0:bool=False ): """ target_state_norm: target states. The cost relative to observations will be 0 if the normalized observations match this. Dim=No weight_state: target states. The weight attributed to each observation. Dim=No weight_state_terminal. The weight attributed to each observation at the end of the prediction horizon. Dim=No target_action_norm: Target actions. The cost relative to actions will be 0 if the nromalized actions match this. Dim=Na weight_action: weight attributed for each action in the cost. Dim=Na exploration_factor: bonus attributed to the uncertainty of the predicted cost function. use_constraints: if set to True, constraints for states will be used for the cost function state_min: minimum bound of the constraints for the normalized observation if use-constraints is True. Dim=No state_max: maximum bound of the constraints for the normalized observation if use-constraints is True. Dim=No area_multiplier: When including constraints in the cost function. The area of the predicted states outside the allowed region is used as penalty. This parameter multiply this area to adjust the effect of constraints on the cost clip_lower_bound_cost_to_0: if set to True, the cost will be clipped to 0. Should always be False Na: dimension of the env actions No: dimension of the env observations """ self.target_state_norm = target_state_norm self.weight_state = weight_state self.weight_state_terminal = weight_state_terminal self.target_action_norm = target_action_norm self.weight_action = weight_action self.exploration_factor = exploration_factor self.use_constraints=use_constraints self.state_min=state_min self.state_max=state_max self.area_multiplier=area_multiplier self.clip_lower_bound_cost_to_0 = clip_lower_bound_cost_to_0 # Computed after init self.target_state_action_norm = None self.weight_matrix_cost = None self.weight_matrix_cost_terminal = None self = convert_config_lists_to_tensor(self) self = combine_weight_matrix(self) self.target_state_action_norm = torch.cat((self.target_state_norm, self.target_action_norm)) def combine_weight_matrix(reward_config:RewardConfig): reward_config.weight_matrix_cost = \ torch.block_diag( torch.diag(reward_config.weight_state), torch.diag(reward_config.weight_action)) reward_config.weight_matrix_cost_terminal = torch.diag(reward_config.weight_state_terminal) return reward_config
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/config_classes/model_config.py
rl_gp_mpc/config_classes/model_config.py
from .utils.functions_process_config import convert_dict_lists_to_dict_tensor, extend_dim, extend_dim_lengthscale_time class ModelConfig: def __init__( self, gp_init:dict = { "noise_covar.noise": [1e-4, 1e-4, 1e-4], # variance = (std)² "base_kernel.lengthscale": [[0.75, 0.75, 0.75, 0.75], [0.75, 0.75, 0.75, 0.75], [0.75, 0.75, 0.75, 0.75]], "outputscale": [5e-2, 5e-2, 5e-2] }, init_lengthscale_time:float=100, min_std_noise:float=1e-3, max_std_noise:float=3e-1, min_outputscale:float=1e-5, max_outputscale:float=0.95, min_lengthscale:float=4e-3, max_lengthscale:float=25.0, min_lengthscale_time:float=10, max_lengthscale_time:float=10000, include_time_model:bool=False, ): """ include_time_model: if set to true, the time of the observation (index of control) will be an additional contextual input to the model, which will allow to weight recent points more in the case that the environment changes over time gp_init: dict containing the initialization of the gp model parameters. The dimension are the number of observation, since there is 1 model per state/observation. Lengthcales dimensions are (No, No+Na). No: numver of ovservations, Na: number of actions All parameters are for normalized observations expect for the time lengthscales The noise parameter represents the noise variance. Take the sqrt to get the corresponding std. """ self.include_time_model = include_time_model self.min_std_noise = min_std_noise self.max_std_noise = max_std_noise self.min_outputscale = min_outputscale self.max_outputscale = max_outputscale self.min_lengthscale = min_lengthscale self.max_lengthscale = max_lengthscale self.min_lengthscale_time = min_lengthscale_time self.max_lengthscale_time = max_lengthscale_time self.init_lengthscale_time = init_lengthscale_time self.include_time_model = include_time_model self.gp_init = convert_dict_lists_to_dict_tensor(gp_init) def extend_dimensions_params(self, dim_state, dim_input): self.min_std_noise = extend_dim(self.min_std_noise, dim=(dim_state,)) self.max_std_noise = extend_dim(self.max_std_noise, dim=(dim_state,)) self.min_outputscale = extend_dim(self.min_outputscale, dim=(dim_state,)) self.max_outputscale = extend_dim(self.max_outputscale, dim=(dim_state,)) self.gp_init["noise_covar.noise"] = extend_dim(self.gp_init["noise_covar.noise"], dim=(dim_state,)) self.gp_init["outputscale"] = extend_dim(self.gp_init["outputscale"], dim=(dim_state,)) if self.include_time_model: self.min_lengthscale = extend_dim_lengthscale_time(self.min_lengthscale, self.min_lengthscale_time, dim_state, dim_input) self.max_lengthscale = extend_dim_lengthscale_time(self.max_lengthscale, self.max_lengthscale_time, dim_state, dim_input) self.gp_init["base_kernel.lengthscale"] = extend_dim_lengthscale_time( lengthscale=self.gp_init["base_kernel.lengthscale"], lengthscale_time=self.init_lengthscale_time, num_models=dim_state, num_inputs=dim_input, ) else: self.min_lengthscale = extend_dim(self.min_lengthscale, dim=(dim_state, dim_input)) self.max_lengthscale = extend_dim(self.max_lengthscale, dim=(dim_state, dim_input)) self.gp_init["base_kernel.lengthscale"] = extend_dim(self.gp_init["base_kernel.lengthscale"], dim=(dim_state, dim_input))
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/config_classes/controller_config.py
rl_gp_mpc/config_classes/controller_config.py
class ControllerConfig: def __init__(self, len_horizon:int=15, actions_optimizer_params:dict={ "disp": None, "maxcor": 30, "ftol": 1e-99, "gtol": 1e-99, "eps": 1e-2, "maxfun": 30, "maxiter": 30, "iprint": -1, "maxls": 30, "finite_diff_rel_step": None }, init_from_previous_actions:bool=True, restarts_optim:int=1, optimize:bool=True, num_repeat_actions:int=1 ): """ len_horizon: number of timesteps used by the mpc to find the optimal action. The effective number of predicted timtesteps will be len_horizon * num_repeat_actions actions_optimizer_params: parameters of the optimizer used. Refer to the scipy documentation for more information about the parameters init_from_previous_actions: if set to true, the actions of the optimizer will be initialized with actions of the previous timestep, shifted by 1 restarts_optim: number of restarts for the optimization. Can help escape local mimimum of the actions if init_from_previous_actions is set to true optimize: if set to false, the optimal actions will be taken at random. Used for debugging only num_repeat_actions: The actions will be repeated this number of time. It allows to predict more timesteps in advance while keeping the same len_horizon number. It can also be used to be more resilient to noise with the model. The model only predicts the change after all actions have been repeated. """ self.len_horizon = len_horizon self.actions_optimizer_params = actions_optimizer_params self.init_from_previous_actions = init_from_previous_actions self.restarts_optim=restarts_optim self.optimize = optimize self.num_repeat_actions=num_repeat_actions
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/config_classes/total_config.py
rl_gp_mpc/config_classes/total_config.py
import torch from .controller_config import ControllerConfig from .actions_config import ActionsConfig from .model_config import ModelConfig from .training_config import TrainingConfig from .reward_config import RewardConfig from .observation_config import ObservationConfig from .memory_config import MemoryConfig torch.set_default_tensor_type(torch.DoubleTensor) class Config: def __init__( self, observation_config: ObservationConfig=ObservationConfig(), reward_config:RewardConfig=RewardConfig(), actions_config:ActionsConfig=ActionsConfig(), model_config:ModelConfig=ModelConfig(), memory_config:MemoryConfig=MemoryConfig(), training_config:TrainingConfig=TrainingConfig(), controller_config:ControllerConfig=ControllerConfig() ): self.observation = observation_config self.reward = reward_config self.actions =actions_config self.model = model_config self.memory = memory_config self.training = training_config self.controller = controller_config
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/config_classes/utils/functions_process_config.py
rl_gp_mpc/config_classes/utils/functions_process_config.py
import torch def convert_config_lists_to_tensor(self): for attr, value in self.__dict__.items(): if isinstance(value, list): setattr(self, attr, torch.Tensor(value)) return self def convert_dict_lists_to_dict_tensor(dict_list: "dict[list]"): for attr, value in dict_list.items(): if isinstance(value, list): dict_list[attr] = torch.Tensor(value) return dict_list def extend_dim_lengthscale_time(lengthscale, lengthscale_time:float, num_models:int, num_inputs:int): lengthscales = torch.empty((num_models, num_inputs)) if isinstance(lengthscale, float) or isinstance(lengthscale, int) or lengthscale.dim() != 1: lengthscales[:, :-1] = lengthscale lengthscales[:, -1] = lengthscale_time else: lengthscales[:, :-1] = lengthscale[None].t().repeat((1, num_inputs - 1)) lengthscales[:, -1] = lengthscale_time return lengthscales def extend_dim(orig_tensor, dim:int): orig_tensor = orig_tensor if isinstance(orig_tensor, torch.Tensor) else torch.tensor(orig_tensor) ones_multiply = torch.ones(dim) if orig_tensor.ndim < ones_multiply.ndim: orig_tensor = orig_tensor.unsqueeze(-1) return orig_tensor * ones_multiply
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/control_objects/__init__.py
rl_gp_mpc/control_objects/__init__.py
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/control_objects/actions_mappers/action_init_functions.py
rl_gp_mpc/control_objects/actions_mappers/action_init_functions.py
import numpy as np def generate_mpc_action_init_random(len_horizon, dim_action): return np.random.uniform(low=0, high=1, size=(len_horizon, dim_action)).reshape(-1) def generate_mpc_action_init_frompreviousiter(actions_mpc, dim_action): actions_mpc[:-dim_action] = actions_mpc[dim_action:] return actions_mpc def get_init_action_change(len_horizon, max_change_action_norm): return np.dot(np.expand_dims(np.random.uniform(low=-1, high=1, size=(len_horizon)), 1), np.expand_dims(max_change_action_norm, 0)) def get_init_action(len_horizon, num_actions): return np.random.uniform(low=0, high=1, size=(len_horizon, num_actions))
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/control_objects/actions_mappers/abstract_action_mapper.py
rl_gp_mpc/control_objects/actions_mappers/abstract_action_mapper.py
from typing import Union import numpy as np import torch from rl_gp_mpc.config_classes.actions_config import ActionsConfig class AbstractActionMapper: ''' Class use to map actions to space between 0 and 1. The optimiser constraints the embed_action to be between 0 and 1. By using this class, the action can be constrainted in interesting ways. It is used by DerivativeActionsMapper to limit the rate of change of actions. Another application would be to use a autoencoder to allow the mpc controller to work with high dimension actions by optimizing in the embed space. ''' def __init__(self, action_low:Union[np.ndarray, torch.Tensor], action_high:Union[np.ndarray, torch.Tensor], len_horizon: int, config:ActionsConfig): self.config = config self.action_low = torch.Tensor(action_low) self.action_high = torch.Tensor(action_high) self.dim_action = len(action_low) self.len_horizon = len_horizon self.n_iter_ctrl = 0 def transform_action_raw_to_action_model(self, action_raw: Union[np.ndarray, torch.Tensor]) -> torch.Tensor: raise NotImplementedError def transform_action_model_to_action_raw(self, action_model: Union[np.ndarray, torch.Tensor], update_internals:bool=False) -> torch.Tensor: raise NotImplementedError def transform_action_mpc_to_action_model(self, action_mpc: np.ndarray) -> torch.Tensor: raise NotImplementedError def transform_action_mpc_to_action_raw(self, action_mpc: Union[np.ndarray, torch.Tensor], update_internals:bool=False) -> torch.Tensor: action_model = self.transform_action_mpc_to_action_model(action_mpc) action_raw = self.transform_action_model_to_action_raw(action_model, update_internals=update_internals) return action_raw def norm_action(self, action: Union[np.ndarray, torch.Tensor]) -> torch.Tensor: return (torch.Tensor(action) - self.action_low) / (self.action_high - self.action_low) def denorm_action(self, normed_action: Union[np.ndarray, torch.Tensor], update_internals=False) -> torch.Tensor: # Update internals is True when the action will be applied to the environment if update_internals: self.n_iter_ctrl += 1 return torch.Tensor(normed_action) * (self.action_high - self.action_low) + self.action_low
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/control_objects/actions_mappers/derivative_action_mapper.py
rl_gp_mpc/control_objects/actions_mappers/derivative_action_mapper.py
from typing import Union import numpy as np import torch from rl_gp_mpc.config_classes.actions_config import ActionsConfig from rl_gp_mpc.control_objects.utils.pytorch_utils import Clamp from .abstract_action_mapper import AbstractActionMapper class DerivativeActionMapper(AbstractActionMapper): def __init__(self, action_low:Union[np.ndarray, torch.Tensor], action_high:Union[np.ndarray, torch.Tensor], len_horizon: int, config:ActionsConfig): super().__init__(action_low, action_high, len_horizon, config) self.action_model_previous_iter = torch.rand(self.dim_action) self.bounds = [(0, 1)] * self.dim_action * len_horizon self.clamp_class = Clamp() def transform_action_raw_to_action_model(self, action_raw: Union[np.ndarray, torch.Tensor]) -> torch.Tensor: action_model = self.norm_action(action_raw) return action_model def transform_action_model_to_action_raw(self, action_model: Union[np.ndarray, torch.Tensor], update_internals:bool=False) -> torch.Tensor: if update_internals: self.action_model_previous_iter = action_model[0] action_raw = self.denorm_action(action_model, update_internals=update_internals) return action_raw def transform_action_mpc_to_action_model(self, action_mpc: torch.Tensor) -> torch.Tensor: actions_mpc_2d = torch.atleast_2d(action_mpc.reshape(self.len_horizon, -1)) actions_mpc_2d = actions_mpc_2d * 2 * self.config.max_change_action_norm - self.config.max_change_action_norm actions_mpc_2d[0] = actions_mpc_2d[0] + self.action_model_previous_iter actions_model = torch.cumsum(actions_mpc_2d, dim=0) clamp = self.clamp_class.apply actions_model = clamp(actions_model, 0, 1) return actions_model
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/control_objects/actions_mappers/normalization_action_mapper.py
rl_gp_mpc/control_objects/actions_mappers/normalization_action_mapper.py
from typing import Union import numpy as np import torch from rl_gp_mpc.config_classes.actions_config import ActionsConfig from .abstract_action_mapper import AbstractActionMapper class NormalizationActionMapper(AbstractActionMapper): def __init__(self, action_low:Union[np.ndarray, torch.Tensor], action_high:Union[np.ndarray, torch.Tensor], len_horizon: int, config:ActionsConfig): super().__init__(action_low, action_high, len_horizon, config) self.bounds = [(0, 1)] * self.dim_action * len_horizon def transform_action_raw_to_action_model(self, action_raw: Union[np.ndarray, torch.Tensor]) -> torch.Tensor: return self.norm_action(action_raw) def transform_action_model_to_action_raw(self, action_model: Union[np.ndarray, torch.Tensor], update_internals:bool=False) -> torch.Tensor: return self.denorm_action(action_model, update_internals=update_internals) def transform_action_mpc_to_action_model(self, action_mpc: np.ndarray) -> torch.Tensor: actions = torch.atleast_2d(action_mpc.reshape(self.len_horizon, -1)) return actions
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/control_objects/memories/gp_memory.py
rl_gp_mpc/control_objects/memories/gp_memory.py
import torch import numpy as np from rl_gp_mpc.config_classes.memory_config import MemoryConfig from rl_gp_mpc.control_objects.utils.data_utils import form_model_input class Memory: def __init__(self, config:MemoryConfig, dim_input, dim_state, include_time_model=False, step_model=1): self.config = config self.include_time_model = include_time_model self.dim_input = dim_input self.dim_state = dim_state self.step_model = step_model self.inputs = torch.empty(config.points_batch_memory, dim_input) self.states_next = torch.empty(config.points_batch_memory, dim_state) self.rewards = torch.empty(config.points_batch_memory) self.iter_ctrls = torch.empty(config.points_batch_memory) self.errors = torch.empty(config.points_batch_memory, dim_state) self.stds = torch.empty(config.points_batch_memory, dim_state) self.model_inputs = torch.empty(config.points_batch_memory, dim_input) self.model_targets = torch.empty(config.points_batch_memory, dim_state) self.active_data_mask = np.empty(config.points_batch_memory, dtype=bool) self.len_mem = 0 self.len_mem_last_processed = 0 self.len_mem_model = 0 def add(self, state, action_model, state_next, reward, iter_ctrl=0, **kwargs): # Add all points to memory but must use the "prepare" function to form the model inputs # it is needed if the model predicts multiple steps ahead if len(self.inputs) < (self.len_mem + 1): self.inputs = torch.cat(self.inputs, torch.empty(self.config.points_batch_memory, self.dim_input)) self.states_next = torch.cat(self.states_next, torch.empty(self.config.points_batch_memory, self.dim_state)) self.rewards = torch.cat(self.rewards, torch.empty(self.config.points_batch_memory)) self.iter_ctrls = torch.cat(self.rewards, torch.empty(self.config.points_batch_memory)) self.errors = torch.cat(self.errors, torch.empty(self.config.points_batch_memory, self.dim_state)) self.stds = torch.cat(self.stds, torch.empty(self.config.points_batch_memory, self.dim_state)) x = form_model_input(state=state, action_model=action_model, time_idx=iter_ctrl, include_time_model=self.include_time_model, dim_input=self.dim_input) self.inputs[self.len_mem] = x self.states_next[self.len_mem] = state_next self.rewards[self.len_mem] = reward self.iter_ctrls[self.len_mem] = iter_ctrl store_gp_mem = True if self.config.check_errors_for_storage: if 'predicted_state' in kwargs.keys() and kwargs['predicted_state'] is not None: error_prediction = torch.abs(kwargs['predicted_state'] - state_next) store_gp_mem = torch.any(error_prediction > self.config.min_error_prediction_state_for_memory) self.errors[self.len_mem] = error_prediction else: self.errors[self.len_mem] = np.nan if 'predicted_state_std' in kwargs.keys() and kwargs['predicted_state_std'] is not None: store_gp_mem = store_gp_mem and (torch.any(kwargs['predicted_state_std'] > self.config.min_prediction_state_std_for_memory)) self.stds[self.len_mem] = kwargs['predicted_state_std'] else: self.stds[self.len_mem] = np.nan self.active_data_mask[self.len_mem] = store_gp_mem self.len_mem += 1 def prepare_for_model(self): # form the model inputs from the memory. # It is defined in this way to allow to have a model that predicts multiple steps ahead if len(self.model_inputs) < (self.len_mem_model + 1): self.model_inputs = torch.cat(self.model_inputs, torch.empty(self.config.points_batch_memory, self.model_inputs.shape[1])) self.model_targets = torch.cat(self.model_targets, torch.empty(self.config.points_batch_memory, self.model_targets.shape[1])) idxs_add_memory = self.get_indexes_to_process() mask_add_memory = self.active_data_mask[idxs_add_memory] idxs_add_memory = idxs_add_memory[mask_add_memory] len_to_add = len(idxs_add_memory) new_model_inputs, new_model_targets = self.get_memory_by_index(idxs_add_memory) self.model_inputs[self.len_mem_model:self.len_mem_model+len_to_add] = new_model_inputs self.model_targets[self.len_mem_model:self.len_mem_model+len_to_add] = new_model_targets self.len_mem_model += len_to_add self.len_mem_last_processed = self.len_mem def get_memory_total(self): indexes= self.get_indexes_processed() inputs, targets = self.get_memory_by_index(indexes) return inputs, targets def get_memory_by_index(self, indexes): inputs = self.inputs[indexes] targets = self.states_next[indexes + self.step_model - 1] - self.inputs[indexes, :self.dim_state] return inputs, targets def get_indexes_to_process(self): return np.arange(self.len_mem_last_processed, self.len_mem, self.step_model) def get_indexes_processed(self): return np.arange(0, self.len_mem_last_processed, self.step_model) def get_mask_model_inputs(self): indexes = self.get_indexes_processed() return self.active_data_mask[indexes] def get(self): if self.len_mem_model > 0: x_mem = self.model_inputs[:self.len_mem_model] y_mem = self.model_targets[:self.len_mem_model] else: # If no datapoints, initialize with dummy 0 variable to avoid errors x_mem = torch.zeros((1, self.dim_input)) y_mem = torch.zeros((1, self.dim_state)) return x_mem, y_mem
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/control_objects/controllers/iteration_info_class.py
rl_gp_mpc/control_objects/controllers/iteration_info_class.py
import numpy as np import torch NUM_DECIMALS_REPR = 3 class IterationInformation: def __init__(self, iteration:int, state:float, cost:float, cost_std:float, mean_predicted_cost:float, mean_predicted_cost_std:float, lower_bound_mean_predicted_cost:float, predicted_idxs:np.array, predicted_states:np.array, predicted_states_std:np.array, predicted_actions:np.array, predicted_costs:np.array, predicted_costs_std:np.array, ): self.iteration=iteration self.state=state self.cost=cost self.cost_std=cost_std self.mean_predicted_cost=mean_predicted_cost self.mean_predicted_cost_std=mean_predicted_cost_std self.lower_bound_mean_predicted_cost=lower_bound_mean_predicted_cost self.predicted_idxs=predicted_idxs self.predicted_states=predicted_states self.predicted_states_std=predicted_states_std self.predicted_actions=predicted_actions self.predicted_costs=predicted_costs self.predicted_costs_std=predicted_costs_std def to_arrays(self): for key in self.__dict__.keys(): if isinstance(self.__dict__[key], torch.Tensor): self.__setattr__(key, np.array(self.__dict__[key])) def to_tensors(self): for key in self.__dict__.keys(): if isinstance(self.__dict__[key], np.array): self.__setattr__(key, torch.Tensor(self.__dict__[key])) def __str__(self): np.set_printoptions(precision=NUM_DECIMALS_REPR, suppress=True) iter_info_dict = self.__dict__ str_rep = "\n" for key, item in iter_info_dict.items(): #if key in ['predicted_states', 'predicted_states_std']: # continue if isinstance(item, np.ndarray) or isinstance(item, torch.Tensor): item = np.array2string(np.array(item), threshold=np.inf, max_line_width=np.inf, separator=',').replace('\n', '') # np.round( ), NUM_DECIMALS_REPR)# .tolist() else: item = np.round(item, NUM_DECIMALS_REPR) str_rep += f"{key}: {item}\n" return str_rep
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/control_objects/controllers/abstract_controller.py
rl_gp_mpc/control_objects/controllers/abstract_controller.py
from rl_gp_mpc.config_classes.total_config import Config class BaseControllerObject: def __init__(self, config: Config): raise NotImplementedError def add_memory(self, obs, action, obs_new, reward, **kwargs): raise NotImplementedError() def get_action(self, obs_mu, obs_var=None): raise NotImplementedError() def get_action_random(self, obs_mu, obs_var=None): raise NotImplementedError() def train(self): raise NotImplementedError()
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/control_objects/controllers/gp_mpc_controller.py
rl_gp_mpc/control_objects/controllers/gp_mpc_controller.py
import multiprocessing import numpy as np import torch from scipy.optimize import minimize from rl_gp_mpc.control_objects.actions_mappers.action_init_functions import generate_mpc_action_init_random, generate_mpc_action_init_frompreviousiter from rl_gp_mpc.config_classes.total_config import Config from rl_gp_mpc.control_objects.observations_states_mappers.normalization_observation_state_mapper import NormalizationObservationStateMapper from rl_gp_mpc.control_objects.actions_mappers.derivative_action_mapper import DerivativeActionMapper from rl_gp_mpc.control_objects.actions_mappers.normalization_action_mapper import NormalizationActionMapper from rl_gp_mpc.control_objects.models.gp_model import GpStateTransitionModel from rl_gp_mpc.control_objects.states_reward_mappers.setpoint_distance_reward_mapper import SetpointStateRewardMapper from rl_gp_mpc.control_objects.memories.gp_memory import Memory from rl_gp_mpc.control_objects.utils.pytorch_utils import Clamp from .abstract_controller import BaseControllerObject from .iteration_info_class import IterationInformation class GpMpcController(BaseControllerObject): def __init__(self, observation_low:np.array, observation_high:np.array, action_low:np.array, action_high:np.array, config:Config): self.config = config self.observation_state_mapper = NormalizationObservationStateMapper(config=config.observation, observation_low=observation_low, observation_high=observation_high) if self.config.actions.limit_action_change: self.actions_mapper = DerivativeActionMapper(config=self.config.actions, action_low=action_low, action_high=action_high, len_horizon=self.config.controller.len_horizon) else: self.actions_mapper = NormalizationActionMapper(config=self.config.actions, action_low=action_low, action_high=action_high, len_horizon=self.config.controller.len_horizon) self.transition_model = GpStateTransitionModel(config=self.config.model, dim_state=self.observation_state_mapper.dim_observation, dim_action=self.actions_mapper.dim_action) self.state_reward_mapper = SetpointStateRewardMapper(config=self.config.reward) self.memory = Memory(self.config.memory, dim_input=self.transition_model.dim_input, dim_state=self.transition_model.dim_state, include_time_model=self.transition_model.config.include_time_model, step_model=self.config.controller.num_repeat_actions) self.actions_mpc_previous_iter = None self.clamp_lcb_class = Clamp() self.iter_ctrl = 0 self.num_cores_main = multiprocessing.cpu_count() self.ctx = multiprocessing.get_context('spawn') self.queue_train = self.ctx.Queue() self.info_iters = {} def get_action(self, obs_mu:np.array, obs_var:np.array=None, random:bool=False): """ Get the optimal action given the observation by optimizing the actions of the simulated trajectory with the gaussian process models such that the lower confidence bound of the mean cost of the trajectory is minimized. Only the first action of the prediction window is returned. Args: obs_mu (numpy.array): unnormalized observation from the gym environment. dim=(Ns) obs_var (numpy.array): unnormalized variance of the observation from the gym environment. dim=(Ns, Ns). default=None. If it is set to None, the observation noise from the json parameters will be used for every iteration. Ns is the dimension of states in the gym environment. Returns: next_action_raw (numpy.array): action to use in the gym environment. It is denormalized, so it can be used directly. dim=(Na), where Ns is the dimension of the action_space """ # Check for parallel process that are open but not alive at each iteration to retrieve the results and close them self.check_and_close_processes() if (self.iter_ctrl % self.config.controller.num_repeat_actions) == 0: self.memory.prepare_for_model() with torch.no_grad(): state_mu, state_var = self.observation_state_mapper.get_state(obs=obs_mu, obs_var=obs_var, update_internals=True) if random: actions_model = self._get_random_actions(state_mu, state_var) else: actions_model = self._get_optimal_actions(state_mu, state_var) actions_raw = self.actions_mapper.transform_action_model_to_action_raw(actions_model, update_internals=True) next_action_raw = actions_raw[0] # all the information fo the trajectory are stored in the object when using the function compute_mean_lcb_trajectory with torch.no_grad(): reward, reward_var = self.state_reward_mapper.get_reward(state_mu, state_var, actions_model[0]) states_std_pred = torch.diagonal(self.states_var_pred, dim1=-2, dim2=-1).sqrt() indexes_predicted = np.arange(self.iter_ctrl, self.iter_ctrl + self.config.controller.len_horizon * self.config.controller.num_repeat_actions, self.config.controller.num_repeat_actions) self.iter_info = IterationInformation( iteration=self.iter_ctrl, state=self.states_mu_pred[0], cost=-reward.item(), cost_std=reward_var.sqrt().item(), mean_predicted_cost=np.min([-self.rewards_trajectory.mean().item(), 3]), mean_predicted_cost_std=self.rewards_traj_var.sqrt().mean().item(), lower_bound_mean_predicted_cost=self.cost_traj_mean_lcb.item(), predicted_idxs=indexes_predicted, predicted_states=self.states_mu_pred, predicted_states_std=states_std_pred, predicted_actions=actions_model, predicted_costs=-self.rewards_trajectory, predicted_costs_std=self.rewards_traj_var.sqrt(), ) self.store_iter_info(self.iter_info) self.past_action = next_action_raw else: next_action_raw = self.past_action self.iter_ctrl += 1 return np.array(next_action_raw) def _get_optimal_actions(self, state_mu:torch.Tensor, state_var:torch.Tensor) -> torch.Tensor: x_mem, y_mem = self.memory.get() with torch.no_grad(): self.transition_model.prepare_inference(x_mem, y_mem) # The optimize function from the scipy library. # It is used to get the optimal actions_norm in the prediction window # that minimizes the lower bound of the predicted cost. The jacobian is used, # otherwise the computation times would be 5 to 10x slower (for the tests I used) opt_fun = np.inf actions_mpc_optim = None for idx_restart in range(self.config.controller.restarts_optim): # only init from previous actions for the first restart, if it's available in the object and specified to be used in the config if self.config.controller.init_from_previous_actions and (self.actions_mpc_previous_iter is not None) and (idx_restart == 0): actions_mpc_init = generate_mpc_action_init_frompreviousiter(self.actions_mpc_previous_iter, dim_action=self.actions_mapper.dim_action) else: actions_mpc_init = generate_mpc_action_init_random(len_horizon=self.config.controller.len_horizon, dim_action=self.actions_mapper.dim_action) if self.config.controller.optimize: iter_res = minimize(fun=self.compute_mean_lcb_trajectory, x0=actions_mpc_init, jac=True, args=(state_mu, state_var), method='L-BFGS-B', bounds=self.actions_mapper.bounds, options=self.config.controller.actions_optimizer_params) actions_mpc_restart = iter_res.x func_val = iter_res.fun else: actions_mpc_restart = generate_mpc_action_init_random(len_horizon=self.config.controller.len_horizon, dim_action=self.actions_mapper.dim_action) func_val, grad_val = self.compute_mean_lcb_trajectory(actions_mpc_restart, state_mu, state_var) if func_val < opt_fun or (actions_mpc_optim is None and np.isnan(func_val)): opt_fun = func_val actions_mpc_optim = actions_mpc_restart self.actions_mpc_previous_iter = actions_mpc_optim.copy() actions_mpc_optim = torch.Tensor(actions_mpc_optim) actions_model_optim = self.actions_mapper.transform_action_mpc_to_action_model(actions_mpc_optim) return actions_model_optim def _get_random_actions(self, state_mu:torch.Tensor, state_var:torch.Tensor) -> torch.Tensor: actions_mpc = generate_mpc_action_init_random(len_horizon=self.config.controller.len_horizon, dim_action=self.actions_mapper.dim_action) actions_model = self.actions_mapper.transform_action_mpc_to_action_model(torch.Tensor(actions_mpc)) x_mem, y_mem = self.memory.get() with torch.no_grad(): self.transition_model.prepare_inference(x_mem, y_mem) func_val, grad_val = self.compute_mean_lcb_trajectory(actions_mpc, state_mu, state_var) # store the trajectory info in the object return actions_model def add_memory(self, obs: np.array, action: np.array, obs_new:np.array, reward:float, predicted_state:np.array=None, predicted_state_std:np.array=None): """ Add an observation, action and observation after applying the action to the memory that is used by the gaussian process models. At regular number of points interval (self.training_frequency), the training process of the gaussian process models will be launched to optimize the hyper-parameters. Args: obs: non-normalized observation. Dim=(Ns,) action: non-normalized action. Dim=(Ns,) obs_new: non-normalized observation obtained after applying the action on the observation. Dim=(Ns,) reward: reward obtained from the gym env. Unused at the moment. The cost given state and action is computed instead. predicted_state: if check_storage is True and predicted_state is not None, the prediction error for that point will be computed. and the point will only be stored in memory if the prediction error is larger than self.error_pred_memory. Dim=(Ns,) predicted_state_std: If check_storage is true, and predicted_state_std is not None, the point will only be stored in memory if it is larger than self.error_pred_memory. Dim=(Ns,) where Ns: dimension of states, Na: dimension of actions """ state_mu, state_var = self.observation_state_mapper.get_state(obs=obs, update_internals=False) state_mu_new, state_var_new = self.observation_state_mapper.get_state(obs=obs_new, update_internals=False) action_model = self.actions_mapper.transform_action_raw_to_action_model(action) self.memory.add(state_mu, action_model, state_mu_new, reward, iter_ctrl=self.iter_ctrl-1, predicted_state=predicted_state, predicted_state_std=predicted_state_std) if self.iter_ctrl % self.config.training.training_frequency == 0 and \ not ('p_train' in self.__dict__ and not self.p_train._closed): self.start_training_process() def start_training_process(self): saved_state = self.transition_model.save_state() saved_state.to_arrays() self.p_train = self.ctx.Process(target=self.transition_model.train, args=( self.queue_train, saved_state, self.config.training.lr_train, self.config.training.iter_train, self.config.training.clip_grad_value, self.config.training.print_train, self.config.training.step_print_train) ) self.p_train.start() def check_and_close_processes(self): """ Check active parallel processes, wait for their resolution, get the parameters and close them """ if 'p_train' in self.__dict__ and not self.p_train._closed and not (self.p_train.is_alive()): params_dict_list = self.queue_train.get() self.p_train.join() for model_idx in range(len(self.transition_model.models)): self.transition_model.models[model_idx].initialize(**params_dict_list[model_idx]) self.p_train.close() x_mem, y_mem = self.memory.get() self.transition_model.prepare_inference(x_mem, y_mem) def compute_mean_lcb_trajectory(self, actions_mpc:np.array, obs_mu:torch.Tensor, obs_var:torch.Tensor): """ Compute the mean lower bound cost of a trajectory given the actions of the trajectory and initial state distribution. The gaussian process models are used to predict the evolution of states (mean and variance). Then the cost is computed for each predicted state and the mean is returned. The partial derivatives of the mean lower bound cost with respect to the actions are also returned. They are computed automatically with autograd from pytorch. This function is called multiple times by an optimizer to find the optimal actions. Args: actions_mpc: actions to apply for the simulated trajectory. It is a flat 1d array so that this function can be used by the minimize function of the scipy library. It is reshaped and transformed into a tensor inside. If self.config.actions.limit_action_change is true, each element of the array contains the relative change with respect to the previous iteration, so that the change can be bounded by the optimizer. dim=(Nh x Na,) where Nh is the len of the horizon and Na the dimension of actions obs_mu: mean value of the inital state distribution. dim=(Ns,) where Ns is the dimension of state obs_var: covariance matrix of the inital state distribution. dim=(Ns, Ns) where Ns is the dimension of state Returns: mean_cost_traj_lcb.item() (float): lower bound of the mean cost distribution of the predicted trajectory. gradients_dcost_dactions.detach().numpy() (numpy.array): Derivative of the lower bound of the mean cost distribution with respect to each of the mpc actions in the prediction horizon. Dim=(Nh * Na,) where Nh is the len of the horizon and Ma the dimension of actions """ # reshape actions from flat 1d numpy array into 2d tensor actions_mpc = torch.Tensor(actions_mpc) actions_mpc.requires_grad = True actions_model = self.actions_mapper.transform_action_mpc_to_action_model(action_mpc=actions_mpc) states_mu_pred, states_var_pred = self.transition_model.predict_trajectory(actions_model, obs_mu, obs_var, len_horizon=self.config.controller.len_horizon, current_time_idx=self.iter_ctrl) rewards_traj, rewards_traj_var = self.state_reward_mapper.get_rewards_trajectory(states_mu_pred, states_var_pred, actions_model) rewards_traj_ucb = rewards_traj + self.config.reward.exploration_factor * torch.sqrt(rewards_traj_var) if self.config.reward.clip_lower_bound_cost_to_0: clamp = self.clamp_lcb_class.apply rewards_traj_ucb = clamp(rewards_traj_ucb, float('-inf'), 0) mean_reward_traj_ucb = rewards_traj_ucb.mean() mean_cost_traj_ucb = -mean_reward_traj_ucb gradients_dcost_dactions = torch.autograd.grad(mean_cost_traj_ucb, actions_mpc, retain_graph=False)[0] self.cost_traj_mean_lcb = mean_reward_traj_ucb.detach() self.states_mu_pred = states_mu_pred.detach() self.rewards_trajectory = rewards_traj.detach() self.rewards_traj_var = rewards_traj_var.detach() self.states_var_pred = states_var_pred.detach() return mean_cost_traj_ucb.item(), gradients_dcost_dactions.detach().numpy() def compute_cost_unnormalized(self, obs:np.array, action:np.array, obs_var:np.array=None)-> "tuple[float]": """ Compute the cost on un-normalized state and actions. Takes in numpy array and returns numpy array. Meant to be used to compute the cost outside the object. Args: obs (numpy.array): state (or observation). shape=(Ns,) action (numpy.array): action. Shape=(Na,) obs_var (numpy.array): state (or observation) variance. Default=None. shape=(Ns, Ns) If set to None, the observation constant stored inside the object will be used Returns: cost_mu (float): Mean of the cost cost_var (float): variance of the cost """ state_mu, state_var = self.observation_state_mapper.get_state(obs=obs, obs_var=obs_var, update_internals=False) action_model = self.actions_mapper.transform_action_raw_to_action_model(action) reward_mu, reward_var = self.state_reward_mapper.get_reward(state_mu, state_var, action_model) return -reward_mu.item(), reward_var.item() def get_iter_info(self): return self.iter_info def store_iter_info(self, iter_info: IterationInformation): iter_info_dict = iter_info.__dict__ for key in iter_info_dict.keys(): if not key in self.info_iters: self.info_iters[key] = [iter_info_dict[key]] else: self.info_iters[key].append(iter_info_dict[key])
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/control_objects/observations_states_mappers/abstract_observation_state_mapper.py
rl_gp_mpc/control_objects/observations_states_mappers/abstract_observation_state_mapper.py
from typing import Union import numpy as np import torch from rl_gp_mpc.config_classes.observation_config import ObservationConfig class AbstractObservationStateMapper: def __init__(self, observation_low: Union[np.ndarray, torch.Tensor], observation_high: Union[np.ndarray, torch.Tensor], config:ObservationConfig): self.config=config self.obs_low = torch.Tensor(observation_low) self.obs_high = torch.Tensor(observation_high) self.var_norm_factor = torch.pow((self.obs_high - self.obs_low), 2) self.dim_observation = len(observation_low) self.dim_state = self.dim_observation def get_state(self, obs:Union[np.ndarray, torch.Tensor], obs_var:Union[np.ndarray, torch.Tensor], update_internals:bool) -> torch.Tensor: raise(NotImplementedError()) def get_obs(self): raise(NotImplementedError())
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/control_objects/observations_states_mappers/normalization_observation_state_mapper.py
rl_gp_mpc/control_objects/observations_states_mappers/normalization_observation_state_mapper.py
from typing import Union import numpy as np import torch from rl_gp_mpc.config_classes.observation_config import ObservationConfig from .abstract_observation_state_mapper import AbstractObservationStateMapper class NormalizationObservationStateMapper(AbstractObservationStateMapper): def __init__(self, observation_low: Union[np.ndarray, torch.Tensor], observation_high: Union[np.ndarray, torch.Tensor], config:ObservationConfig): super().__init__(observation_low, observation_high, config) def get_state(self, obs: Union[np.ndarray, torch.Tensor], obs_var:Union[np.ndarray, torch.Tensor]=None, update_internals=False) -> torch.Tensor: state = (torch.Tensor(obs) - self.obs_low) / (self.obs_high - self.obs_low) if obs_var is not None: state_var = torch.Tensor(obs_var) / self.var_norm_factor else: state_var = self.config.obs_var_norm return state, state_var
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/control_objects/models/gp_model.py
rl_gp_mpc/control_objects/models/gp_model.py
import multiprocessing import time import torch import gpytorch import numpy as np from rl_gp_mpc.config_classes.model_config import ModelConfig from .abstract_model import AbstractStateTransitionModel class SavedState: def __init__(self, inputs:torch.Tensor, states_change: torch.Tensor, parameters:"list[dict]", constraints_hyperparams:dict, models: "list[gpytorch.models.ExactGP]"=None): self.inputs = inputs self.states_change = states_change self.parameters = parameters self.constraints_hyperparams = constraints_hyperparams if models is not None: self.state_dicts_models = [model.state_dict() for model in models] else: self.state_dicts_models = None def to_arrays(self): # convert tensors to arrays so that they can be sent to the multiprocessing process for training without errors self.inputs = self.inputs.numpy() self.states_change = self.states_change.numpy() self.parameters = [{k: v.numpy() for k, v in parameter.items()} for parameter in self.parameters] self.constraints_hyperparams = {k: v.numpy() if isinstance(v, torch.Tensor) else v for k, v in self.constraints_hyperparams.items()} def to_tensors(self): # used to reconstruct tensors inside the multiprocess process for training self.inputs = torch.Tensor(self.inputs) self.states_change = torch.Tensor(self.states_change) self.parameters = [{k: torch.Tensor(v) for k, v in parameter.items()} for parameter in self.parameters] self.constraints_hyperparams = {k: torch.Tensor(v) if isinstance(v, np.ndarray) else v for k, v in self.constraints_hyperparams.items()} class GpStateTransitionModel(AbstractStateTransitionModel): def __init__(self, config: ModelConfig, dim_state, dim_action): super().__init__(config, dim_state, dim_action) if self.config.include_time_model: self.dim_input += 1 self.config.extend_dimensions_params(dim_state=self.dim_state, dim_input=self.dim_input) self.models = create_models( gp_init_dict=self.config.gp_init, constraints_gp=self.config.__dict__, inputs=None, targets=None, num_models=self.dim_state, num_inputs=self.dim_input ) for idx_model in range(len(self.models)): self.models[idx_model].eval() def predict_trajectory(self, actions: torch.Tensor, obs_mu: torch.Tensor, obs_var: torch.Tensor, len_horizon: int, current_time_idx: int) -> "tuple[torch.Tensor]": """ Compute the future predicted states distribution for the simulated trajectory given the current initial state (or observation) distribution (obs_mu and obs_var) and planned actions It also returns the costs, the variance of the costs, and the lower confidence bound of the cost along the trajectory Args: actions: actions to apply for the simulated trajectory. dim=(Nh, Na) where Nh is the len of the horizon and Na the dimension of actions obs_mu: mean value of the inital state distribution. dim=(Ns,) where Ns is the dimension of state obs_var: variance matrix of the inital state distribution. dim=(Ns, Ns) where Ns is the dimension of state Returns: states_mu_pred: predicted states of the trajectory. The first element contains the initial state. Dim=(Nh + 1, Ns) states_var_pred: covariance matrix of the predicted states of the trajectory. The first element contains the initial state. Dim=(Nh + 1, Ns, Ns) """ dim_state = self.dim_state dim_input_model = self.dim_input dim_action = self.dim_action states_mu_pred = torch.empty((len_horizon + 1, len(obs_mu))) states_var_pred = torch.empty((len_horizon + 1, dim_state, dim_state)) states_mu_pred[0] = obs_mu states_var_pred[0] = obs_var state_dim = obs_mu.shape[0] # Input of predict_next_state_change is not a state, but the concatenation of state and action for idx_time in range(1, len_horizon + 1): input_var = torch.zeros((dim_input_model, dim_input_model)) input_var[:state_dim, :state_dim] = states_var_pred[idx_time - 1] input_mean = torch.empty((dim_input_model,)) input_mean[:dim_state] = states_mu_pred[idx_time - 1] input_mean[dim_state:(dim_state + dim_action)] = actions[idx_time - 1] if self.config.include_time_model: input_mean[-1] = current_time_idx + idx_time - 1 state_change, state_change_var, v = self.predict_next_state_change(input_mean, input_var) # use torch.clamp(states_mu_pred[idx_time], 0, 1) ? states_mu_pred[idx_time] = states_mu_pred[idx_time - 1] + state_change states_var_pred[idx_time] = state_change_var + states_var_pred[idx_time - 1] + \ input_var[:states_var_pred.shape[1]] @ v + \ v.t() @ input_var[:states_var_pred.shape[1]].t() return states_mu_pred, states_var_pred def predict_next_state_change(self, input_mu: torch.Tensor, input_var: torch.Tensor): """ Approximate GP regression at noisy inputs via moment matching IN: mean (m) (row vector) and (s) variance of the state OUT: mean (M) (row vector), variance (S) of the action and inv(s)*input-ouputcovariance Function inspired from https://github.com/nrontsis/PILCO/blob/6a962c8e4172f9e7f29ed6e373c4be2dd4b69cb7/pilco/models/mgpr.py#L81, reinterpreted from tensorflow to pytorch. Must be called after self.prepare_inference Args: input_mu: mean value of the input distribution. Dim=(Ns + Na,) input_var: covariance matrix of the input distribution. Dim=(Ns + Na, Ns + Na) Returns: M.t() (torch.Tensor): mean value of the predicted change distribution. Dim=(Ns,) S (torch.Tensor): covariance matrix of the predicted change distribution. Dim=(Ns, Ns) V.t() (torch.Tensor): Dim=(Ns, Ns + Na) where Ns: dimension of state, Na: dimension of action """ dim_state = self.dim_state dim_input_model = self.dim_input input_var = input_var[None, None, :, :].repeat([dim_state, dim_state, 1, 1]) inp = (self.x_mem - input_mu)[None, :, :].repeat([dim_state, 1, 1]) iN = inp @ self.iL B = self.iL @ input_var[0, ...] @ self.iL + torch.eye(dim_input_model) # Redefine iN as in^T and t --> t^T # B is symmetric, so it is equivalent iN_t = torch.transpose(iN, -1, -2) t = torch.transpose(torch.linalg.solve(B, iN_t), -1, -2) lb = torch.exp(-torch.sum(iN * t, -1) / 2) * self.beta tiL = t @ self.iL c = self.variances / torch.sqrt(torch.det(B)) M = (torch.sum(lb, -1) * c)[:, None] V = torch.matmul(torch.transpose(tiL.conj(), -1, -2), lb[:, :, None])[..., 0] * c[:, None] # Calculate S: Predictive Covariance R = torch.matmul(input_var, torch.diag_embed( 1 / torch.square(self.lengthscales[None, :, :]) + 1 / torch.square(self.lengthscales[:, None, :]) )) + torch.eye(dim_input_model) X = inp[None, :, :, :] / torch.square(self.lengthscales[:, None, None, :]) X2 = -inp[:, None, :, :] / torch.square(self.lengthscales[None, :, None, :]) Q = torch.linalg.solve(R, input_var) / 2 Xs = torch.sum(X @ Q * X, -1) X2s = torch.sum(X2 @ Q * X2, -1) maha = -2 * torch.matmul(torch.matmul(X, Q), torch.transpose(X2.conj(), -1, -2)) + Xs[:, :, :, None] + X2s[:, :, None, :] k = torch.log(self.variances)[:, None] - torch.sum(torch.square(iN), -1) / 2 L = torch.exp(k[:, None, :, None] + k[None, :, None, :] + maha) temp = self.beta[:, None, None, :].repeat([1, dim_state, 1, 1]) @ L S = (temp @ self.beta[None, :, :, None].repeat([dim_state, 1, 1, 1]))[:, :, 0, 0] diagL = torch.Tensor.permute(torch.diagonal(torch.Tensor.permute(L, dims=(3, 2, 1, 0)), dim1=-2, dim2=-1), dims=(2, 1, 0)) S = S - torch.diag_embed(torch.sum(torch.mul(self.iK, diagL), [1, 2])) S = S / torch.sqrt(torch.det(R)) S = S + torch.diag_embed(self.variances) S = S - M @ torch.transpose(M, -1, -2) return M.t(), S, V.t() def prepare_inference(self, inputs: torch.Tensor, state_changes: torch.Tensor): # compute all the parameters needed for inference that only depend on the memory, and not on the input, such that they can be computed in advance. # So they are not computed inside the mpc for each iteration self.x_mem = inputs self.y_mem = state_changes self.iK, self.beta = calculate_factorizations(inputs, state_changes, self.models) # self.x_mem = inputs[:self.beta.shape[1]] self.lengthscales = torch.stack([model.covar_module.base_kernel.lengthscale[0] for model in self.models]) self.variances = torch.stack([model.covar_module.outputscale for model in self.models]) self.iL = torch.diag_embed(1 / self.lengthscales) @staticmethod def train(queue:multiprocessing.Queue, saved_state:SavedState, lr_train:float, num_iter_train:int, clip_grad_value:float, print_train:bool=False, step_print_train:int=25): """ # , saved_state:SavedState , Train the gaussian process models hyper-parameters such that the marginal-log likelihood for the predictions of the points in memory is minimized. This function is launched in parallel of the main process, which is why a queue is used to tranfer information back to the main process and why the gaussian process models are reconstructed using the points in memory and hyper-parameters (the objects cant be sent directly as argument). If an error occurs, returns the parameters sent as init values (hyper-parameters obtained by the previous training process) Args: queue: queue object used to transfer information to the main process saved_state: SavedState, contains all the information to reconstruct the models lr_train: learning rate of the training num_iter_train: number of iteration for the training optimizer clip_grad_value: value at which the gradient are clipped, so that the training is more stable print_train: weither to print the information during training. default=False step_print_train: If print_train is True, only print the information every step_print_train iteration """ #torch.set_num_threads(1) start_time = time.time() saved_state.to_tensors() # create models, which is necessary since this function is used in a parallel process # that do not share memory with the principal process models = create_models(saved_state.parameters, saved_state.constraints_hyperparams, saved_state.inputs, saved_state.states_change) best_outputscales = [model.covar_module.outputscale.detach() for model in models] best_noises = [model.likelihood.noise.detach() for model in models] best_lengthscales = [model.covar_module.base_kernel.lengthscale.detach() for model in models] previous_losses = torch.empty(len(models)) for model_idx in range(len(models)): output = models[model_idx](models[model_idx].train_inputs[0]) mll = gpytorch.mlls.ExactMarginalLogLikelihood(models[model_idx].likelihood, models[model_idx]) previous_losses[model_idx] = -mll(output, models[model_idx].train_targets) best_losses = previous_losses.detach().clone() # Random initialization of the parameters showed better performance than # just taking the value from the previous iteration as init values. # If parameters found at the end do not better performance than previous iter, # return previous parameters for model_idx in range(len(models)): models[model_idx].covar_module.outputscale = \ models[model_idx].covar_module.raw_outputscale_constraint.lower_bound + \ torch.rand(models[model_idx].covar_module.outputscale.shape) * \ (models[model_idx].covar_module.raw_outputscale_constraint.upper_bound - \ models[model_idx].covar_module.raw_outputscale_constraint.lower_bound) models[model_idx].covar_module.base_kernel.lengthscale = \ models[model_idx].covar_module.base_kernel.raw_lengthscale_constraint.lower_bound + \ torch.rand(models[model_idx].covar_module.base_kernel.lengthscale.shape) * \ (models[model_idx].covar_module.base_kernel.raw_lengthscale_constraint.upper_bound - \ models[model_idx].covar_module.base_kernel.raw_lengthscale_constraint.lower_bound) models[model_idx].likelihood.noise = \ models[model_idx].likelihood.noise_covar.raw_noise_constraint.lower_bound + \ torch.rand(models[model_idx].likelihood.noise.shape) * \ (models[model_idx].likelihood.noise_covar.raw_noise_constraint.upper_bound - models[model_idx].likelihood.noise_covar.raw_noise_constraint.lower_bound) mll = gpytorch.mlls.ExactMarginalLogLikelihood(models[model_idx].likelihood, models[model_idx]) models[model_idx].train() models[model_idx].likelihood.train() optimizer = torch.optim.LBFGS([ {'params': models[model_idx].parameters()}, # Includes GaussianLikelihood parameters ], lr=lr_train, line_search_fn='strong_wolfe') try: for i in range(num_iter_train): def closure(): optimizer.zero_grad() # Output from model output = models[model_idx](models[model_idx].train_inputs[0]) # Calc loss and backprop gradients loss = -mll(output, models[model_idx].train_targets) torch.nn.utils.clip_grad_value_(models[model_idx].parameters(), clip_grad_value) loss.backward() if print_train: if i % step_print_train == 0: print( 'Iter %d/%d - Loss: %.5f output_scale: %.5f lengthscale: %s noise: %.5f' % ( i + 1, num_iter_train, loss.item(), models[model_idx].covar_module.outputscale.item(), str(models[ model_idx].covar_module.base_kernel.lengthscale.detach().numpy()), pow(models[model_idx].likelihood.noise.item(), 0.5) )) return loss loss = optimizer.step(closure) if loss < best_losses[model_idx]: best_losses[model_idx] = loss.item() best_lengthscales[model_idx] = models[model_idx].covar_module.base_kernel.lengthscale best_noises[model_idx] = models[model_idx].likelihood.noise best_outputscales[model_idx] = models[model_idx].covar_module.outputscale except Exception as e: print(e) print( 'training process - model %d - time train %f - output_scale: %s - lengthscales: %s - noise: %s' % ( model_idx, time.time() - start_time, str(best_outputscales[model_idx].detach().numpy()), str(best_lengthscales[model_idx].detach().numpy()), str(best_noises[model_idx].detach().numpy()))) print('training process - previous marginal log likelihood: %s - new marginal log likelihood: %s' % (str(previous_losses.detach().numpy()), str(best_losses.detach().numpy()))) params_dict_list = [] for model_idx in range(len(models)): params_dict_list.append({ 'covar_module.base_kernel.lengthscale': best_lengthscales[model_idx].detach().numpy(), 'covar_module.outputscale': best_outputscales[model_idx].detach().numpy(), 'likelihood.noise': best_noises[model_idx].detach().numpy()}) queue.put(params_dict_list) def save_state(self) -> SavedState: saved_state = SavedState( inputs=self.x_mem, states_change=self.y_mem, parameters=[model.state_dict() for model in self.models], constraints_hyperparams=self.config.__dict__ ) return saved_state def create_models(gp_init_dict, constraints_gp, inputs:torch.Tensor=None, targets:torch.Tensor=None, num_models=None, num_inputs=None) -> "list[ExactGPModelMonoTask]": """ Define gaussian process models used for predicting state transition, using constraints and init values for (outputscale, noise, lengthscale). Args: train_inputs (torch.Tensor or None): Input values in the memory of the gps train_targets (torch.Tensor or None): target values in the memory of the gps. Represent the change in state values gp_init_dict (dict or list of dict): Value of the hyper-parameters of the gaussian processes. constraints_gp (dict): See the ReadMe about parameters for information about keys num_models (int or None): Must be provided when train_inputs or train_targets are None. The number of models should be equal to the dimension of state, so that the transition for each state can be predicted with a different gp. Default=None num_inputs (int or None): Must be provided when train_inputs or train_targets are None. The number of inputs should be equal to the sum of the dimension of state and dimension of action. Default=None include_time (bool): If True, gp will have one additional input corresponding to the time of the observation. This is usefull if the env change with time, as more recent points will be trusted more than past points (time closer to the point to make inference at). It is to be specified only if Returns: models (list of gpytorch.models.ExactGP): models containing the parameters, memory, constraints of the gps and functions for exact predictions """ if inputs is not None and targets is not None: num_models = len(targets[0]) models = [ExactGPModelMonoTask(inputs, targets[:, idx_model], len(inputs[0])) for idx_model in range(num_models)] else: if num_models is None or num_inputs is None: raise(ValueError('If train_inputs or train_targets are None, num_models and num_inputs must be defined')) else: models = [ExactGPModelMonoTask(None, None, num_inputs) for _ in range(num_models)] for idx_model in range(num_models): if constraints_gp is not None: if "min_std_noise" in constraints_gp.keys() and "max_std_noise" in constraints_gp.keys(): min_var_noise_model = torch.pow(constraints_gp['min_std_noise'][idx_model], 2) max_var_noise_model = torch.pow(constraints_gp['max_std_noise'][idx_model], 2) models[idx_model].likelihood.noise_covar.register_constraint("raw_noise", gpytorch.constraints.Interval(lower_bound=min_var_noise_model, upper_bound=max_var_noise_model)) if "min_outputscale" in constraints_gp.keys(): min_outputscale_model = constraints_gp['min_outputscale'][idx_model] max_outputscale_model = constraints_gp['max_outputscale'][idx_model] models[idx_model].covar_module.register_constraint("raw_outputscale", gpytorch.constraints.Interval(lower_bound=min_outputscale_model, upper_bound=max_outputscale_model)) if "min_lengthscale" in constraints_gp.keys(): min_lengthscale_model = constraints_gp['min_lengthscale'][idx_model] max_lengthscale_model = constraints_gp['max_lengthscale'][idx_model] models[idx_model].covar_module.base_kernel.register_constraint("raw_lengthscale", gpytorch.constraints.Interval(lower_bound=min_lengthscale_model, upper_bound=max_lengthscale_model)) if type(gp_init_dict) == list: models[idx_model].load_state_dict(gp_init_dict[idx_model]) else: hypers = {'base_kernel.lengthscale': gp_init_dict['base_kernel.lengthscale'][idx_model], 'outputscale': gp_init_dict['outputscale'][idx_model]} hypers_likelihood = {'noise_covar.noise': gp_init_dict['noise_covar.noise'][idx_model]} models[idx_model].likelihood.initialize(**hypers_likelihood) models[idx_model].covar_module.initialize(**hypers) return models class ExactGPModelMonoTask(gpytorch.models.ExactGP): def __init__(self, train_x, train_y, dim_input): likelihood = gpytorch.likelihoods.GaussianLikelihood() super(ExactGPModelMonoTask, self).__init__(train_x, train_y, likelihood) self.covar_module = gpytorch.kernels.ScaleKernel(gpytorch.kernels.RBFKernel(ard_num_dims=dim_input)) self.mean_module = gpytorch.means.ZeroMean() def forward(self, x): mean_x = self.mean_module(x) covar_x = self.covar_module(x) return gpytorch.distributions.MultivariateNormal(mean_x, covar_x) def calculate_factorizations(x:torch.Tensor, y:torch.Tensor, models: "list[ExactGPModelMonoTask]"): """ Compute iK and beta using the points in memory, which are needed to make predictions with the gaussian processes. These two variables only depends on data in memory, and not on input distribution, so they separated from other computation such that they can be computed outside the optimisation function, which is computed multiple times at each iteration Function inspired from https://github.com/nrontsis/PILCO/blob/6a962c8e4172f9e7f29ed6e373c4be2dd4b69cb7/pilco/models/mgpr.py#L81, reimplemented from tensorflow to pytorch Args: x (torch.Tensor): matrix containing the states and actions. Dim=(Nm, Ns + Na + Nc) y (torch.Tensor): matrix containing the states change. Dim=(Nm, Ns) models (list of gpytorch.models.ExactGP): list containing the gp models used to predict each state change. Len=Ns Ns: number of states Na: number of actions Nc: number of context variables. At the moment, only 1 if time used Nm: number of points in memory Returns: iK (torch.Tensor): needed by the gaussian processes models to compute the predictions beta (torch.Tensor): needed by the gaussian processes models to compute the predictions """ K = torch.stack([model.covar_module(x).evaluate() for model in models]) batched_eye = torch.eye(K.shape[1]).repeat(K.shape[0], 1, 1) L = torch.linalg.cholesky(K + torch.stack([model.likelihood.noise for model in models])[:, None] * batched_eye) iK = torch.cholesky_solve(batched_eye, L) Y_ = (y.t())[:, :, None] beta = torch.cholesky_solve(Y_, L)[:, :, 0] return iK, beta
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/control_objects/models/abstract_model.py
rl_gp_mpc/control_objects/models/abstract_model.py
import torch from rl_gp_mpc.config_classes.model_config import ModelConfig class AbstractStateTransitionModel: def __init__(self, config:ModelConfig, dim_state, dim_action): self.config = config self.dim_state = dim_state self.dim_action = dim_action self.dim_input = self.dim_state + self.dim_action def predict_trajectory(self, input:torch.Tensor, input_var:torch.Tensor) -> "tuple[torch.Tensor]": raise NotImplementedError def predict_next_state(self, input: torch.Tensor, input_var: torch.Tensor) -> "tuple[torch.Tensor]": raise NotImplementedError def prepare_inference(self, x, y): raise NotImplementedError def train(self, x, y): raise NotImplementedError def save_state(self): raise NotImplementedError def load_state(self, saved_state): raise NotImplementedError
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/control_objects/utils/data_utils.py
rl_gp_mpc/control_objects/utils/data_utils.py
import torch def form_model_input(state, action_model, time_idx, include_time_model, dim_input): x = torch.empty(dim_input) x[:((state.shape[0] + action_model.shape[0]))] = torch.cat((state, action_model)) if include_time_model: x[-1] = time_idx return x
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/control_objects/utils/pytorch_utils.py
rl_gp_mpc/control_objects/utils/pytorch_utils.py
import torch from math import sqrt class Clamp(torch.autograd.Function): # This class allow the flow of gradient when at the boundary. Otherwise, when using the clamp function from pytorch, the action remains blocked at 0 or 1 @staticmethod def forward(ctx, input, min, max): return input.clamp(min=min, max=max) @staticmethod def backward(ctx, grad_output): return grad_output.clone(), None, None def normal_cdf(x, mu, sigma): return 0.5 * (1 + torch.erf((x - mu) / (sigma * sqrt(2))))
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/control_objects/states_reward_mappers/abstract_state_reward_mapper.py
rl_gp_mpc/control_objects/states_reward_mappers/abstract_state_reward_mapper.py
import torch from rl_gp_mpc.config_classes.reward_config import RewardConfig class AbstractStateRewardMapper: def __init__(self, config:RewardConfig): self.config = config def get_reward(self, state_mu:torch.Tensor, state_var: torch.Tensor, action:torch.Tensor): raise NotImplementedError def get_reward_terminal(self, state_mu:torch.Tensor, state_var:torch.Tensor): raise NotImplementedError def get_rewards_trajectory(self, states_mu:torch.Tensor, states_var:torch.Tensor, actions:torch.Tensor) -> torch.Tensor: raise NotImplementedError
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/rl_gp_mpc/control_objects/states_reward_mappers/setpoint_distance_reward_mapper.py
rl_gp_mpc/control_objects/states_reward_mappers/setpoint_distance_reward_mapper.py
import torch from rl_gp_mpc.config_classes.reward_config import RewardConfig from .abstract_state_reward_mapper import AbstractStateRewardMapper from ..utils.pytorch_utils import normal_cdf class SetpointStateRewardMapper(AbstractStateRewardMapper): def __init__(self, config:RewardConfig): super().__init__(config) def get_reward(self, state_mu: torch.Tensor, state_var: torch.Tensor, action:torch.Tensor) -> float: """ Compute the quadratic cost of one state distribution or a trajectory of states distributions given the mean value and variance of states (observations), the weight matrix, and target state. The state, state_var and action must be normalized. If reading directly from the gym env observation, this can be done with the gym env action space and observation space. See an example of normalization in the add_points_memory function. Args: state_mu (torch.Tensor): normed mean value of the state or observation distribution (elements between 0 and 1). dim=(Ns) or dim=(Np, Ns) state_var (torch.Tensor): normed variance matrix of the state or observation distribution (elements between 0 and 1) dim=(Ns, Ns) or dim=(Np, Ns, Ns) action (torch.Tensor): normed actions. (elements between 0 and 1). dim=(Na) or dim=(Np, Na) Np: length of the prediction trajectory. (=self.len_horizon) Na: dimension of the gym environment actions Ns: dimension of the gym environment states Returns: cost_mu (torch.Tensor): mean value of the cost distribution. dim=(1) or dim=(Np) cost_var (torch.Tensor): variance of the cost distribution. dim=(1) or dim=(Np) """ error = torch.cat((state_mu, action), -1) - self.config.target_state_action_norm if state_var.ndim == 3: #error = torch.cat((state_mu, action), 1) - self.config.target_state_action_norm state_action_var = torch.cat(( torch.cat((state_var, torch.zeros((state_var.shape[0], state_var.shape[1], action.shape[1]))), 2), torch.zeros((state_var.shape[0], action.shape[1], state_var.shape[1] + action.shape[1]))), 1) else: #error = torch.cat((state_mu, action), 0) - self.config.target_state_action_norm state_action_var = torch.block_diag(state_var, torch.zeros((action.shape[0], action.shape[0]))) weight_matrix_cost = self.config.weight_matrix_cost cost_mu = torch.diagonal(torch.matmul(state_action_var, weight_matrix_cost), dim1=-1, dim2=-2).sum(-1) + \ torch.matmul(torch.matmul(error[..., None].transpose(-1, -2), weight_matrix_cost), error[..., None]).squeeze() TS = weight_matrix_cost @ state_action_var cost_var_term1 = torch.diagonal(2 * TS @ TS, dim1=-1, dim2=-2).sum(-1) cost_var_term2 = TS @ weight_matrix_cost cost_var_term3 = (4 * error[..., None].transpose(-1, -2) @ cost_var_term2 @ error[..., None]).squeeze() cost_var = cost_var_term1 + cost_var_term3 if self.config.use_constraints: if state_mu.ndim == 2: penalty_min_constraint = torch.stack([normal_cdf(self.config.state_min, state_mu[time_idx], state_var[time_idx].diag()) for time_idx in range(state_mu.shape[0])]) penalty_max_constraint = torch.stack([1 - normal_cdf(self.config.state_max, state_mu[time_idx], state_var[time_idx].diag()) for time_idx in range(state_mu.shape[0])]) else: penalty_min_constraint = normal_cdf(self.config.state_min, state_mu, state_var.diag()) penalty_max_constraint = 1 - normal_cdf(self.config.state_max, state_mu, state_var.diag()) cost_mu = cost_mu + penalty_max_constraint.sum(-1) + penalty_min_constraint.sum(-1) return -cost_mu, cost_var def get_rewards(self, state_mu: torch.Tensor, state_var: torch.Tensor, action:torch.Tensor) -> float: """ Compute the quadratic cost of one state distribution or a trajectory of states distributions given the mean value and variance of states (observations), the weight matrix, and target state. The state, state_var and action must be normalized. If reading directly from the gym env observation, this can be done with the gym env action space and observation space. See an example of normalization in the add_points_memory function. Args: state_mu (torch.Tensor): normed mean value of the state or observation distribution (elements between 0 and 1). dim=(Ns) or dim=(Np, Ns) state_var (torch.Tensor): normed variance matrix of the state or observation distribution (elements between 0 and 1) dim=(Ns, Ns) or dim=(Np, Ns, Ns) action (torch.Tensor): normed actions. (elements between 0 and 1). dim=(Na) or dim=(Np, Na) Np: length of the prediction trajectory. (=self.len_horizon) Na: dimension of the gym environment actions Ns: dimension of the gym environment states Returns: reward_mu (torch.Tensor): mean value of the reward distribution. (=-cost) dim=(1) or dim=(Np) cost_var (torch.Tensor): variance of the cost distribution. dim=(1) or dim=(Np) """ error = torch.cat((state_mu, action), 1) - self.config.target_state_action_norm state_action_var = torch.cat(( torch.cat((state_var, torch.zeros((state_var.shape[0], state_var.shape[1], action.shape[1]))), 2), torch.zeros((state_var.shape[0], action.shape[1], state_var.shape[1] + action.shape[1]))), 1) cost_mu = torch.diagonal(torch.matmul(state_action_var, self.config.weight_matrix_cost), dim1=-1, dim2=-2).sum(-1) + \ torch.matmul(torch.matmul(error[..., None].transpose(-1, -2), self.config.weight_matrix_cost), error[..., None]).squeeze() TS = self.config.weight_matrix_cost @ state_action_var cost_var_term1 = torch.diagonal(2 * TS @ TS, dim1=-1, dim2=-2).sum(-1) cost_var_term2 = TS @ self.config.weight_matrix_cost cost_var_term3 = (4 * error[..., None].transpose(-1, -2) @ cost_var_term2 @ error[..., None]).squeeze() cost_var = cost_var_term1 + cost_var_term3 if self.config.use_constraints: state_distribution = [torch.distributions.normal.Normal(state_mu[idx], state_var[idx]) for idx in range(state_mu.shape[0])] penalty_min_constraint = torch.stack([state_distribution[time_idx].cdf(torch.Tensor( self.config.state_min)) * self.config.area_multiplier for time_idx in range(state_mu.shape[0])]).diagonal(0, -1, -2).sum(-1) penalty_max_constraint = torch.stack([(1 - state_distribution[time_idx].cdf(torch.Tensor( self.config.state_max))) * self.config.area_multiplier for time_idx in range(state_mu.shape[0])]).diagonal(0, -1, -2).sum(-1) cost_mu = cost_mu + penalty_max_constraint + penalty_min_constraint return -cost_mu, cost_var def get_reward_terminal(self, state_mu:torch.Tensor, state_var:torch.Tensor) -> float: """ Compute the terminal cost of the prediction trajectory. Args: state_mu (torch.Tensor): mean value of the terminal state distribution. dim=(Ns) state_var (torch.Tensor): variance matrix of the terminal state distribution. dim=(Ns, Ns) Returns: reward_mu (torch.Tensor): mean value of the reward distribution. (=-cost) dim=(1) cost_var (torch.Tensor): variance of the cost distribution. dim=(1) """ error = state_mu - self.config.target_state_norm cost_mu = torch.trace(torch.matmul(state_var, self.config.weight_matrix_cost_terminal)) + \ torch.matmul(torch.matmul(error.t(), self.config.weight_matrix_cost_terminal), error) TS = self.config.weight_matrix_cost_terminal @ state_var cost_var_term1 = torch.trace(2 * TS @ TS) cost_var_term2 = 4 * error.t() @ TS @ self.config.weight_matrix_cost_terminal @ error cost_var = cost_var_term1 + cost_var_term2 return -cost_mu, cost_var def get_rewards_trajectory(self, states_mu:torch.Tensor, states_var:torch.Tensor, actions:torch.Tensor) -> torch.Tensor: rewards_traj, rewards_traj_var = self.get_reward(states_mu[:-1], states_var[:-1], actions) cost_traj_final, rewards_traj_var_final = self.get_reward_terminal(states_mu[-1], states_var[-1]) rewards_traj = torch.cat((rewards_traj, cost_traj_final[None]), 0) rewards_traj_var = torch.cat((rewards_traj_var, rewards_traj_var_final[None]), 0) return rewards_traj, rewards_traj_var def compute_squared_dist_cost(error, state_action_var, weight_matrix_cost): cost_mu = torch.diagonal(torch.matmul(state_action_var, weight_matrix_cost), dim1=-1, dim2=-2).sum(-1) + \ torch.matmul(torch.matmul(error[..., None].transpose(-1, -2), weight_matrix_cost), error[..., None]).squeeze() TS = weight_matrix_cost @ state_action_var cost_var_term1 = torch.diagonal(2 * TS @ TS, dim1=-1, dim2=-2).sum(-1) cost_var_term2 = TS @ weight_matrix_cost cost_var_term3 = (4 * error[..., None].transpose(-1, -2) @ cost_var_term2 @ error[..., None]).squeeze() cost_var = cost_var_term1 + cost_var_term3 return cost_mu, cost_var
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/examples/pendulum/config_pendulum.py
examples/pendulum/config_pendulum.py
from rl_gp_mpc.config_classes.total_config import Config from rl_gp_mpc.config_classes.controller_config import ControllerConfig from rl_gp_mpc.config_classes.actions_config import ActionsConfig from rl_gp_mpc.config_classes.reward_config import RewardConfig from rl_gp_mpc.config_classes.observation_config import ObservationConfig from rl_gp_mpc.config_classes.memory_config import MemoryConfig from rl_gp_mpc.config_classes.model_config import ModelConfig from rl_gp_mpc.config_classes.training_config import TrainingConfig def get_config(len_horizon=15, include_time_model=False, num_repeat_actions=1): observation_config = ObservationConfig( obs_var_norm = [1e-6, 1e-6, 1e-6] ) reward_config = RewardConfig( target_state_norm=[1, 0.5, 0.5], weight_state=[1, 0.1, 0.1], weight_state_terminal=[5, 2, 2], target_action_norm=[0.5], weight_action=[1e-3], exploration_factor=1, use_constraints=False, state_min=[-3, -3, -3], state_max=[3, 3, 3], area_multiplier=1, clip_lower_bound_cost_to_0=False, ) actions_config = ActionsConfig( limit_action_change=False, max_change_action_norm=[0.3] ) model_config = ModelConfig( gp_init = { "noise_covar.noise": [1e-5, 1e-5, 1e-5], # variance = (std)² "base_kernel.lengthscale": [0.5, 0.5, 0.5], "outputscale": [5e-2, 5e-2, 5e-2] }, # [[0.75, 0.75, 0.75, 0.75], # [0.75, 0.75, 0.75, 0.75], # [0.75, 0.75, 0.75, 0.75]] min_std_noise=1e-3, max_std_noise=1e-2, min_outputscale=1e-2, max_outputscale=0.95, min_lengthscale=4e-3, max_lengthscale=10.0, min_lengthscale_time=10, max_lengthscale_time=10000, init_lengthscale_time=100, include_time_model=include_time_model ) memory_config = MemoryConfig( check_errors_for_storage=True, min_error_prediction_state_for_memory=[3e-4, 3e-4, 3e-4], min_prediction_state_std_for_memory=[3e-3, 3e-3, 3e-3], points_batch_memory=1500 ) training_config = TrainingConfig( lr_train=7e-3, iter_train=15, training_frequency=25, clip_grad_value=1e-3, print_train=False, step_print_train=5 ) controller_config = ControllerConfig( len_horizon=len_horizon, actions_optimizer_params = { "disp": None, "maxcor": 4, "ftol": 1e-15, "gtol": 1e-15, "eps": 1e-2, "maxfun": 4, "maxiter": 4, "iprint": -1, "maxls": 4, "finite_diff_rel_step": None }, num_repeat_actions=num_repeat_actions ) config = Config( observation_config=observation_config, reward_config=reward_config, actions_config=actions_config, model_config=model_config, memory_config=memory_config, training_config=training_config, controller_config=controller_config, ) return config
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/examples/pendulum/run_pendulum.py
examples/pendulum/run_pendulum.py
import sys import os file_path = os.path.abspath(__file__) sys.path.append(os.path.join(os.path.dirname(file_path), '../../')) import gym from rl_gp_mpc.config_classes.visu_config import VisuConfig from rl_gp_mpc.run_env_function import run_env from config_pendulum import get_config def run_pendulum(): num_steps = 150 random_actions_init = 10 num_repeat_actions = 1 len_horizon = 15 verbose = True env_name = 'Pendulum-v0' env = gym.make(env_name) control_config = get_config(len_horizon=len_horizon, num_repeat_actions=num_repeat_actions) visu_config = VisuConfig() costs = run_env(env, control_config, visu_config, random_actions_init=random_actions_init, num_steps=num_steps, verbose=verbose) return costs if __name__ == '__main__': costs = run_pendulum()
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/examples/pendulum/run_pendulum_multiple.py
examples/pendulum/run_pendulum_multiple.py
import sys import os file_path = os.path.abspath(__file__) sys.path.append(os.path.join(os.path.dirname(file_path), '../../')) import gym from rl_gp_mpc.config_classes.visu_config import VisuConfig from rl_gp_mpc.run_env_function import run_env_multiple from config_pendulum import get_config if __name__ == '__main__': num_runs = 20 num_steps = 300 random_actions_init = 10 num_repeat_actions = 1 len_horizon = 12 verbose = True env_name = 'Pendulum-v0' env = gym.make(env_name) control_config = get_config(len_horizon=len_horizon, num_repeat_actions=num_repeat_actions) visu_config = VisuConfig(render_live_plot_2d=False, render_env=True, save_render_env=True) run_env_multiple(env, env_name, control_config, visu_config, num_runs, random_actions_init=random_actions_init, num_steps=num_steps, verbose=verbose)
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/examples/mountain_car/config_mountaincar.py
examples/mountain_car/config_mountaincar.py
from rl_gp_mpc.config_classes.total_config import Config from rl_gp_mpc.config_classes.controller_config import ControllerConfig from rl_gp_mpc.config_classes.actions_config import ActionsConfig from rl_gp_mpc.config_classes.reward_config import RewardConfig from rl_gp_mpc.config_classes.observation_config import ObservationConfig from rl_gp_mpc.config_classes.memory_config import MemoryConfig from rl_gp_mpc.config_classes.model_config import ModelConfig from rl_gp_mpc.config_classes.training_config import TrainingConfig def get_config(len_horizon=10, num_repeat_actions=5, include_time_model=False,): observation_config = ObservationConfig( obs_var_norm = [1e-6, 1e-6] ) reward_config = RewardConfig( target_state_norm=[1, 0.5], weight_state=[1, 0], weight_state_terminal=[5, 0], target_action_norm=[0.5], weight_action=[0.05], exploration_factor=1, use_constraints=False, state_min=[0.2, -2], state_max=[0.925, 0.85], area_multiplier=1, clip_lower_bound_cost_to_0=False, ) actions_config = ActionsConfig( limit_action_change=False, max_change_action_norm=[0.3] ) model_config = ModelConfig( gp_init = { "noise_covar.noise": [1e-5, 1e-5], "base_kernel.lengthscale": [0.5, 0.5], "outputscale": [5e-2, 5e-2] }, min_std_noise=1e-3, max_std_noise=1e-2, min_outputscale=1e-5, max_outputscale=0.95, min_lengthscale=4e-3, max_lengthscale=25.0, min_lengthscale_time=10, max_lengthscale_time=10000, init_lengthscale_time=100, include_time_model=include_time_model, ) memory_config = MemoryConfig( check_errors_for_storage=True, min_error_prediction_state_for_memory=[3e-3, 3e-3], min_prediction_state_std_for_memory=[3e-3, 3e-3], points_batch_memory=1500 ) training_config = TrainingConfig( lr_train=7e-3, iter_train=20, training_frequency=60, clip_grad_value=1e-3, print_train=False, step_print_train=5 ) controller_config = ControllerConfig( len_horizon=len_horizon, actions_optimizer_params = { "disp": None, "maxcor": 8, "ftol": 1e-18, "gtol": 1e-18, "eps": 1e-2, "maxfun": 8, "maxiter": 8, "iprint": -1, "maxls": 8, "finite_diff_rel_step": None }, init_from_previous_actions=True, restarts_optim=2, optimize=True, num_repeat_actions=num_repeat_actions ) config = Config( observation_config=observation_config, reward_config=reward_config, actions_config=actions_config, model_config=model_config, memory_config=memory_config, training_config=training_config, controller_config=controller_config, ) return config
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/examples/mountain_car/run_mountain_car_multiple.py
examples/mountain_car/run_mountain_car_multiple.py
import sys import os file_path = os.path.abspath(__file__) sys.path.append(os.path.join(os.path.dirname(file_path), '../../')) import gym from rl_gp_mpc.config_classes.visu_config import VisuConfig from config_mountaincar import get_config from rl_gp_mpc.run_env_function import run_env_multiple if __name__ == '__main__': num_runs = 20 random_actions_init=10 num_repeat_actions=5 num_steps=500 len_horizon = 12 verbose = False env_name = 'MountainCarContinuous-v0' env = gym.make(env_name) control_config = get_config(len_horizon=len_horizon, num_repeat_actions=num_repeat_actions) visu_config = VisuConfig(render_live_plot_2d=False, render_env=True, save_render_env=True) run_env_multiple(env, env_name, control_config, visu_config, num_runs, random_actions_init=random_actions_init, num_steps=num_steps, verbose=verbose)
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/examples/mountain_car/run_mountaincar.py
examples/mountain_car/run_mountaincar.py
import sys import os file_path = os.path.abspath(__file__) sys.path.append(os.path.join(os.path.dirname(file_path), '../../')) import gym from rl_gp_mpc.config_classes.visu_config import VisuConfig from rl_gp_mpc.run_env_function import run_env from config_mountaincar import get_config def run_mountain_car(num_steps=500, random_actions_init=20, num_repeat_actions=5): len_horizon = 12 verbose = True env_name = 'MountainCarContinuous-v0' env = gym.make(env_name) control_config = get_config(len_horizon=len_horizon, num_repeat_actions=num_repeat_actions) visu_config = VisuConfig(render_live_plot_2d=True, render_env=True, save_render_env=True) costs = run_env(env, control_config, visu_config, random_actions_init=random_actions_init, num_steps=num_steps, verbose=verbose) return costs if __name__ == '__main__': costs = run_mountain_car(num_steps=500, num_repeat_actions=5)
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/examples/process_control/config_process_control.py
examples/process_control/config_process_control.py
from rl_gp_mpc.config_classes.total_config import Config from rl_gp_mpc.config_classes.controller_config import ControllerConfig from rl_gp_mpc.config_classes.actions_config import ActionsConfig from rl_gp_mpc.config_classes.reward_config import RewardConfig from rl_gp_mpc.config_classes.observation_config import ObservationConfig from rl_gp_mpc.config_classes.memory_config import MemoryConfig from rl_gp_mpc.config_classes.model_config import ModelConfig from rl_gp_mpc.config_classes.training_config import TrainingConfig def get_config(len_horizon=4, include_time_model=True, num_repeat_actions=10): observation_config = ObservationConfig( obs_var_norm = [1e-6, 1e-6] ) reward_config = RewardConfig( target_state_norm=[0.5, 0.5], weight_state=[1, 1], weight_state_terminal=[1, 1], target_action_norm=[0, 0], weight_action=[1e-4, 1e-4], exploration_factor=1, use_constraints=False, state_min=[0.1, 0.3], state_max=[0.9, 0.8], area_multiplier=1, clip_lower_bound_cost_to_0=False, ) actions_config = ActionsConfig( limit_action_change=False, max_change_action_norm=[0.1, 0.2] ) model_config = ModelConfig( gp_init = { "noise_covar.noise": [1e-5, 1e-5], "base_kernel.lengthscale": [0.25, 0.25], "outputscale": [5e-2, 5e-2] }, init_lengthscale_time=100, min_std_noise=1e-3, max_std_noise=3e-1, min_outputscale=1e-5, max_outputscale=0.95, min_lengthscale=5e-2, max_lengthscale=25.0, include_time_model=include_time_model, min_lengthscale_time=5, max_lengthscale_time=1000 ) memory_config = MemoryConfig( check_errors_for_storage=True, min_error_prediction_state_for_memory=[1e-5, 1e-5], min_prediction_state_std_for_memory=[3e-3, 3e-3], points_batch_memory=1500 ) training_config = TrainingConfig( lr_train=7e-3, iter_train=15, training_frequency=15, clip_grad_value=1e-3, print_train=False, step_print_train=5 ) controller_config = ControllerConfig( len_horizon=len_horizon, actions_optimizer_params = { "disp": None, "maxcor": 15, "ftol": 1e-99, "gtol": 1e-99, "eps": 1e-2, "maxfun": 15, "maxiter": 15, "iprint": -1, "maxls": 15, "finite_diff_rel_step": None }, init_from_previous_actions=True, restarts_optim=2, optimize=True, num_repeat_actions=num_repeat_actions ) config = Config( observation_config=observation_config, reward_config=reward_config, actions_config=actions_config, model_config=model_config, memory_config=memory_config, training_config=training_config, controller_config=controller_config, ) return config
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/examples/process_control/run_processc_control_multiple.py
examples/process_control/run_processc_control_multiple.py
import sys import os file_path = os.path.abspath(__file__) sys.path.append(os.path.join(os.path.dirname(file_path), '../../')) from rl_gp_mpc.envs.process_control import ProcessControl from rl_gp_mpc.config_classes.visu_config import VisuConfig from config_process_control import get_config from rl_gp_mpc.run_env_function import run_env_multiple if __name__ == '__main__': num_runs = 10 num_steps = 1000 random_actions_init = 10 num_repeat_actions = 10 include_time_model = True len_horizon = 5 verbose = True env = ProcessControl( dt=1, s_range=(20, 30), fi_range=(0.2, 0.3), ci_range=(0.1, 0.2), cr_range=(0.9, 1.0), noise_l_prop_range=(3e-3, 1e-2), noise_co_prop_range=(3e-3, 1e-2), sp_l_range=(0.4, 0.6), sp_co_range=(0.4, 0.6), change_params=True, period_change=500 ) env_name = 'process_control' control_config = get_config(len_horizon=len_horizon, include_time_model=include_time_model, num_repeat_actions=num_repeat_actions) visu_config = VisuConfig(render_live_plot_2d=False, render_env=True, save_render_env=True) run_env_multiple(env, env_name, control_config, visu_config, num_runs, random_actions_init=random_actions_init, num_steps=num_steps, verbose=verbose)
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control
https://github.com/SimonRennotte/Data-Efficient-Reinforcement-Learning-with-Probabilistic-Model-Predictive-Control/blob/e73e805688792b4d63fcee08f4eb06a7fab46a11/examples/process_control/run_process_control.py
examples/process_control/run_process_control.py
import sys import os file_path = os.path.abspath(__file__) sys.path.append(os.path.join(os.path.dirname(file_path), '../../')) from rl_gp_mpc.config_classes.visu_config import VisuConfig from rl_gp_mpc.envs.process_control import ProcessControl from rl_gp_mpc.run_env_function import run_env from config_process_control import get_config def run_process_control(): # choose the configuration file to load the corresponding env # open(os.path.join('params', 'main_parameters_mountain_car.json')) # open(os.path.join('params', 'main_parameters_my_env.json')) num_steps = 500 random_actions_init = 100 num_repeat_actions = 5 include_time_model = False len_horizon = 5 verbose = True env = ProcessControl( dt=1, s_range=(20, 30), fi_range=(0.15, 0.3), ci_range=(0.15, 0.2), cr_range=(0.8, 1.0), noise_l_prop_range=(5e-3, 1e-2), noise_co_prop_range=(5e-3, 1e-2), sp_l_range=(0.4, 0.6), sp_co_range=(0.4, 0.6), change_params=False, period_change=200 ) control_config = get_config(len_horizon=len_horizon, include_time_model=include_time_model, num_repeat_actions=num_repeat_actions) visu_config = VisuConfig(render_live_plot_2d=True, render_env=False, save_render_env=False, save_live_plot_2d=True) costs = run_env(env, control_config, visu_config, random_actions_init=random_actions_init, num_steps=num_steps, verbose=verbose) return costs if __name__ == '__main__': costs = run_process_control()
python
MIT
e73e805688792b4d63fcee08f4eb06a7fab46a11
2026-01-05T07:13:13.323103Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/download-model.py
download-model.py
''' Downloads models from Hugging Face to models/model-name. Source: https://github.com/oobabooga/text-generation-webui Example: python download-model.py facebook/opt-1.3b ''' import argparse import base64 import datetime import hashlib import json import re import sys from pathlib import Path import requests import tqdm from tqdm.contrib.concurrent import thread_map def select_model_from_default_options(): models = { "Vicuna-7B-1.1-GPTQ-4bit-128g": ("TheBloke", "vicuna-7B-1.1-GPTQ-4bit-128g", "main"), } choices = {} print("Select the model that you want to download:\n") for i, name in enumerate(models): char = chr(ord('A') + i) choices[char] = name print(f"{char}) {name}") char = chr(ord('A') + len(models)) print(f"{char}) None of the above") print() print("Input> ", end='') choice = input()[0].strip().upper() if choice == char: print( """\nThen type the name of your desired Hugging Face model in the format organization/name. Examples: facebook/opt-1.3b EleutherAI/pythia-1.4b-deduped """) print("Input> ", end='') model = input() branch = "main" else: arr = models[choices[choice]] model = f"{arr[0]}/{arr[1]}" branch = arr[2] return model, branch def sanitize_model_and_branch_names(model, branch): if model[-1] == '/': model = model[:-1] if branch is None: branch = "main" else: pattern = re.compile(r"^[a-zA-Z0-9._-]+$") if not pattern.match(branch): raise ValueError( "Invalid branch name. Only alphanumeric characters, period, underscore and dash are allowed.") return model, branch def get_download_links_from_huggingface(model, branch, text_only=False): base = "https://huggingface.co" page = f"/api/models/{model}/tree/{branch}?cursor=" cursor = b"" links = [] sha256 = [] classifications = [] has_pytorch = False has_pt = False has_ggml = False has_safetensors = False is_lora = False while True: if len(cursor) > 0: page = f"/api/models/{model}/tree/{branch}?cursor=" content = requests.get(f"{base}{page}{cursor.decode()}").content else: page = f"/api/models/{model}/tree/{branch}" content = requests.get(f"{base}{page}").content dict = json.loads(content) if len(dict) == 0: break for i in range(len(dict)): fname = dict[i]['path'] if not is_lora and fname.endswith( ('adapter_config.json', 'adapter_model.bin')): is_lora = True is_pytorch = re.match("(pytorch|adapter)_model.*\\.bin", fname) is_safetensors = re.match(".*\\.safetensors", fname) is_pt = re.match(".*\\.pt", fname) is_ggml = re.match("ggml.*\\.bin", fname) is_tokenizer = re.match("tokenizer.*\\.model", fname) is_text = re.match(".*\\.(txt|json|py|md)", fname) or is_tokenizer if any( (is_pytorch, is_safetensors, is_pt, is_ggml, is_tokenizer, is_text)): if 'lfs' in dict[i]: sha256.append([fname, dict[i]['lfs']['oid']]) if is_text: links.append( f"https://huggingface.co/{model}/resolve/{branch}/{fname}") classifications.append('text') continue if not text_only: links.append( f"https://huggingface.co/{model}/resolve/{branch}/{fname}") if is_safetensors: has_safetensors = True classifications.append('safetensors') elif is_pytorch: has_pytorch = True classifications.append('pytorch') elif is_pt: has_pt = True classifications.append('pt') elif is_ggml: has_ggml = True classifications.append('ggml') cursor = base64.b64encode( f'{{"file_name":"{dict[-1]["path"]}"}}'.encode()) + b':50' cursor = base64.b64encode(cursor) cursor = cursor.replace(b'=', b'%3D') # If both pytorch and safetensors are available, download safetensors only if (has_pytorch or has_pt) and has_safetensors: for i in range(len(classifications) - 1, -1, -1): if classifications[i] in ['pytorch', 'pt']: links.pop(i) return links, sha256, is_lora def get_output_folder(model, branch, is_lora, base_folder=None): if base_folder is None: base_folder = 'models' if not is_lora else 'loras' output_folder = f"{'_'.join(model.split('/')[-2:])}" if branch != 'main': output_folder += f'_{branch}' output_folder = Path(base_folder) / output_folder return output_folder def get_single_file(url, output_folder, start_from_scratch=False): filename = Path(url.rsplit('/', 1)[1]) output_path = output_folder / filename if output_path.exists() and not start_from_scratch: # Check if the file has already been downloaded completely r = requests.get(url, stream=True) total_size = int(r.headers.get('content-length', 0)) if output_path.stat().st_size >= total_size: return # Otherwise, resume the download from where it left off headers = {'Range': f'bytes={output_path.stat().st_size}-'} mode = 'ab' else: headers = {} mode = 'wb' r = requests.get(url, stream=True, headers=headers) with open(output_path, mode) as f: total_size = int(r.headers.get('content-length', 0)) block_size = 1024 with tqdm.tqdm(total=total_size, unit='iB', unit_scale=True, bar_format='{l_bar}{bar}| {n_fmt:6}/{total_fmt:6} {rate_fmt:6}') as t: for data in r.iter_content(block_size): t.update(len(data)) f.write(data) def start_download_threads( file_list, output_folder, start_from_scratch=False, threads=1): thread_map( lambda url: get_single_file( url, output_folder, start_from_scratch=start_from_scratch), file_list, max_workers=threads, disable=True) def download_model_files( model, branch, links, sha256, output_folder, start_from_scratch=False, threads=1): # Creating the folder and writing the metadata if not output_folder.exists(): output_folder.mkdir() with open(output_folder / 'huggingface-metadata.txt', 'w') as f: f.write(f'url: https://huggingface.co/{model}\n') f.write(f'branch: {branch}\n') f.write( f'download date: {str(datetime.datetime.now().strftime("%Y-%m-%d %H:%M:%S"))}\n') sha256_str = '' for i in range(len(sha256)): sha256_str += f' {sha256[i][1]} {sha256[i][0]}\n' if sha256_str != '': f.write(f'sha256sum:\n{sha256_str}') # Downloading the files print(f"Downloading the model to {output_folder}") start_download_threads( links, output_folder, start_from_scratch=start_from_scratch, threads=threads) def check_model_files(model, branch, links, sha256, output_folder): # Validate the checksums validated = True for i in range(len(sha256)): fpath = (output_folder / sha256[i][0]) if not fpath.exists(): print(f"The following file is missing: {fpath}") validated = False continue with open(output_folder / sha256[i][0], "rb") as f: bytes = f.read() file_hash = hashlib.sha256(bytes).hexdigest() if file_hash != sha256[i][1]: print(f'Checksum failed: {sha256[i][0]} {sha256[i][1]}') validated = False else: print(f'Checksum validated: {sha256[i][0]} {sha256[i][1]}') if validated: print('[+] Validated checksums of all model files!') else: print('[-] Invalid checksums. Rerun download-model.py with the --clean flag.') if __name__ == '__main__': parser = argparse.ArgumentParser() parser.add_argument('MODEL', type=str, default=None, nargs='?') parser.add_argument('--branch', type=str, default='main', help='Name of the Git branch to download from.') parser.add_argument('--threads', type=int, default=1, help='Number of files to download simultaneously.') parser.add_argument('--text-only', action='store_true', help='Only download text files (txt/json).') parser.add_argument('--output', type=str, default=None, help='The folder where the model should be saved.') parser.add_argument('--clean', action='store_true', help='Does not resume the previous download.') parser.add_argument('--check', action='store_true', help='Validates the checksums of model files.') args = parser.parse_args() branch = args.branch model = args.MODEL if model is None: model, branch = select_model_from_default_options() # Cleaning up the model/branch names try: model, branch = sanitize_model_and_branch_names(model, branch) except ValueError as err_branch: print(f"Error: {err_branch}") sys.exit() # Getting the download links from Hugging Face links, sha256, is_lora = get_download_links_from_huggingface( model, branch, text_only=args.text_only) # Getting the output folder output_folder = get_output_folder( model, branch, is_lora, base_folder=args.output) if args.check: # Check previously downloaded files check_model_files(model, branch, links, sha256, output_folder) else: # Download files download_model_files(model, branch, links, sha256, output_folder, threads=args.threads)
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/bootstrap_models.py
bootstrap_models.py
import sys import os import subprocess DEFAULT_EMBEDDINGS_MODEL ="sentence-transformers/all-MiniLM-L6-v2" DEFAULT_MODEL = "openlm-research/open_llama_3b" MODEL_DIR = "./models" def _download_if_not_exists(model): models_dir = os.path.join(MODEL_DIR, model) if os.path.exists(models_dir): print(f"Directory {models_dir} already exists! Skpping download!") else: print(f"Downloading model {model} to {models_dir}") print("Please note that if model is large this may take a while.") process = subprocess.run(["python3", "download-model.py", model, "--output", MODEL_DIR], capture_output=True) process.check_returncode() def main(model, embeddings_model): print(f"""Your choices: MODEL: {model} EMBEDDINGS MODEL: {embeddings_model} """) try: os.mkdir(MODEL_DIR) except FileExistsError: pass _download_if_not_exists(embeddings_model) _download_if_not_exists(model) print("Success!") if __name__ == "__main__": model = None embeddings = None if len(sys.argv) > 2: embeddings = sys.argv[2] if len(sys.argv) > 1: model = sys.argv[1] if len(sys.argv) == 1: print( "NOTE: You can change the default downloaded model by passing an additional argument:" + f"{sys.argv[0]} [hugging-face-llm-model-name] [hugging-face-embeddings-model-name]" ) if not embeddings: embeddings = DEFAULT_EMBEDDINGS_MODEL if not model: model = DEFAULT_MODEL main(model, embeddings)
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/main.py
app/main.py
import os import shutil from fastapi import FastAPI, Depends, UploadFile from fastapi.middleware.cors import CORSMiddleware from sqlmodel import SQLModel, create_engine, Session, desc from models.all import Conversation, Message, ConversationWithMessages from typing import List from settings import load_config, logger from plugins import load_plugins from alembic import command from alembic.config import Config from configparser import ConfigParser config = load_config() sqlite_database_url = config.database_url connect_args = {"check_same_thread": False} engine = create_engine(sqlite_database_url, echo=True, connect_args=connect_args) # Introducing a new feature flag # So GuidanceLLaMAcpp can coexist with FlowAgents if config.use_flow_agents: logger.info("Using (experimental) flow agents") from conversations.document_based_flow import DocumentBasedConversationFlowAgent convo = DocumentBasedConversationFlowAgent() else: logger.info("Using experimental Guidance LLaMA cpp implementation.") from conversations.document_based import DocumentBasedConversation convo = DocumentBasedConversation() def create_db_and_tables(): confparser = ConfigParser() confparser.read(f"{config.backend_root_path}/alembic.ini") confparser.set('alembic', 'script_location', f"{config.backend_root_path}/migrations") confparser.set('alembic', 'prepend_sys_path', config.backend_root_path) migrations_config_path = os.path.join(config.backend_root_path, "generated_alembic.ini") with open(migrations_config_path, 'w') as config_file: confparser.write(config_file) migrations_config = Config(migrations_config_path) command.upgrade(migrations_config, "head") def get_session(): with Session(engine) as session: yield session app = FastAPI() # Load the plugins load_plugins(app=app) origins = [ "http://127.0.0.1:5173", "http://localhost:5173", "http://0.0.0.0:5173", ] app.add_middleware( CORSMiddleware, allow_origins=["*"], allow_credentials=True, allow_methods=["*"], allow_headers=["*"], ) @app.on_event("startup") def on_startup(): create_db_and_tables() @app.post('/llm/query/', response_model=str) def llm_query(*, query: str, session: Session = Depends(get_session)): """ Query the LLM """ return convo.predict(query, []) @app.post("/conversations", response_model=Conversation) def create_conversation(*, session: Session = Depends(get_session), conversation: Conversation): """ Create a new conversation. """ conversation = Conversation.from_orm(conversation) session.add(conversation) session.commit() session.refresh(conversation) print(str(conversation)) return conversation @app.put('/conversations/{conversation_id}', response_model=Conversation) def update_conversation(*, session: Session = Depends(get_session), conversation_id: int, payload: dict): """ Update the title of a conversation. """ conversation = session.get(Conversation, conversation_id) conversation.title = payload["title"] session.add(conversation) session.commit() session.refresh(conversation) return conversation @app.delete("/conversations/{conversation_id}") def delete_conversation(*, session: Session = Depends(get_session), conversation_id: int): """ Delete a conversation. """ conversation = session.get(Conversation, conversation_id) session.delete(conversation) session.commit() return conversation @app.get("/conversations", response_model=List[Conversation]) def get_conversations(session: Session = Depends(get_session)): """ Get all conversations. """ return session.query(Conversation).order_by(desc(Conversation.id)).all() @app.get("/conversations/{conversation_id}", response_model=ConversationWithMessages) def get_conversation(conversation_id: int, session: Session = Depends(get_session)): """ Get a conversation by id. """ conversation = session.get(Conversation, conversation_id) return conversation @app.post("/conversations/{conversation_id}/messages", response_model=Message) def create_message(*, session: Session = Depends(get_session), conversation_id: int, message: Message): """ Create a new message. """ message = Message.from_orm(message) session.add(message) session.commit() session.refresh(message) return message @app.post("/conversations/{conversation_id}/files", response_model=dict) def upload_file(*, conversation_id: int, file: UploadFile): """ Upload a file. """ try: uploaded_file_name = file.filename filepath = os.path.join(os.getcwd(), "data", config.upload_path, uploaded_file_name) os.makedirs(os.path.dirname(filepath), mode=0o777, exist_ok=True) with open(filepath, "wb") as f: shutil.copyfileobj(file.file, f) convo.load_document(filepath, conversation_id) return {"text": f"{uploaded_file_name} has been loaded into memory for this conversation."} except Exception as e: logger.error(f"Error adding file to history: {e}") return f"Error adding file to history: {e}" @app.post('/llm/{conversation_id}/', response_model=str) def llm(*, conversation_id: str, query: str, session: Session = Depends(get_session)): """ Query the LLM """ conversation_data = get_conversation(conversation_id, session) history = conversation_data.messages return convo.predict(query, conversation_id) # we could also work from history only # return convo.predict(query, history) @app.post("/conversations/{conversation_id}/messages/{message_id}/upvote", response_model=Message) def upvote_message(*, session: Session = Depends(get_session), conversation_id: int, message_id: int): """ Upvote a message. """ message = session.get(Message, message_id) message.rating = 1 session.add(message) session.commit() session.refresh(message) return message @app.post("/conversations/{conversation_id}/messages/{message_id}/downvote", response_model=Message) def downvote_message(*, session: Session = Depends(get_session), conversation_id: int, message_id: int): """ Downvote a message. """ message = session.get(Message, message_id) message.rating = -1 session.add(message) session.commit() session.refresh(message) return message @app.post("/conversations/{conversation_id}/messages/{message_id}/resetVote", response_model=Message) def reset_message_vote(*, session: Session = Depends(get_session), conversation_id: int, message_id: int): """ Reset a message vote. """ message = session.get(Message, message_id) message.rating = 0 session.add(message) session.commit() session.refresh(message) return message @app.post("/reset", response_model=dict) def reset_all(): """ Reset the database. """ SQLModel.metadata.drop_all(engine) print("Database has been reset.") SQLModel.metadata.create_all(engine) return {"text": "Database has been reset."} if __name__ == "__main__": import uvicorn uvicorn.run("main:app", host="0.0.0.0", port=7865, reload=False)
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/settings.py
app/settings.py
import os from dotenv import load_dotenv import logging from langchain.embeddings import ( HuggingFaceEmbeddings, HuggingFaceInstructEmbeddings, ) class CustomFormatter(logging.Formatter): grey = "\x1b[38;20m" yellow = "\x1b[33;20m" red = "\x1b[31;20m" bold_red = "\x1b[31;1m" reset = "\x1b[0m" format = "%(asctime)s - %(name)s - %(levelname)s - %(message)s (%(filename)s:%(lineno)d)" FORMATS = { logging.DEBUG: grey + format + reset, logging.INFO: grey + format + reset, logging.WARNING: yellow + format + reset, logging.ERROR: red + format + reset, logging.CRITICAL: bold_red + format + reset, } def format(self, record): log_fmt = self.FORMATS.get(record.levelno) formatter = logging.Formatter(log_fmt) return formatter.format(record) logging.basicConfig(format='%(asctime)s - %(levelname)s - %(message)s', level=logging.INFO) logger = logging.getLogger(__name__) logger.setLevel(logging.DEBUG) # create console handler with a higher log level ch = logging.StreamHandler() ch.setLevel(logging.DEBUG) ch.setFormatter(CustomFormatter()) logger.addHandler(ch) class Settings: _instance = None def __new__(cls): if cls._instance is None: cls._instance = super().__new__(cls) return cls._instance def __init__(self): load_dotenv() # Chat API - By default, we are assuming Oobabooga's Text Generation # WebUI is running self.root_path = os.getcwd() if self.root_path.endswith("app"): self.root_path = self.root_path[:-4] self.model_root_path = os.path.join(self.root_path, "models") self.backend_root_path = os.path.join(self.root_path, "app") # let's ensure the models and backend path is accurate when mounting docker volumes if self.root_path.startswith("/code"): self.model_root_path = "/models" self.backend_root_path = self.root_path # Guidance new settings self.test_file = os.getenv("TEST_FILE", "/data/uploads/the_trial.txt") self.embeddings_map = { **{name: HuggingFaceInstructEmbeddings for name in ["hkunlp/instructor-xl", "hkunlp/instructor-large"]}, **{name: HuggingFaceEmbeddings for name in ["all-MiniLM-L6-v2", "sentence-t5-xxl"]}, } self.persist_directory = os.getenv("PERSIST_DIRECTORY", "./persist_directory") self.embeddings_model = os.getenv("EMBEDDINGS_MODEL", f"{self.model_root_path}/all-MiniLM-L6-v2") self.chat_api_url = os.getenv("CHAT_API_URL", "http://0.0.0.0:5000/api/v1/generate") self.model_path = self.model_root_path + os.getenv("MODEL_PATH") self.guidance_reasoning_model_path = self.model_root_path + os.getenv("GUIDANCE_REASONING_MODEL_PATH") self.guidance_extraction_model_path = self.model_root_path + os.getenv("GUIDANCE_EXTRACTION_MODEL_PATH") # Where all data is stored self.data_path = os.getenv("DATA_PATH", f"{self.backend_root_path}/data") # Where short-term memory is stored self.memories_path = os.getenv("MEMORIES_PATH", f"{self.data_path}/memories") # Where uploads are saved self.upload_path = os.getenv("UPLOAD_PATH", f"{self.data_path}/uploads") # Where conversation history is stored self.conversation_history_path = os.getenv( "CONVERSATION_HISTORY_PATH", f"{self.data_path}/conversation_history/", ) # Document store name self.document_store_name = os.getenv("DOCUMENT_STORE_NAME", "brainchulo_docs") self.conversation_store_name = os.getenv("CONVERSATION_STORE_NAME", "brainchulo_convos") # Default objective - If we go objective-based, this is the default self.default_objective = os.getenv("DEFAULT_OBJECTIVE", "Be a CEO.") # Database URL self.database_url = os.getenv("DATABASE_URL", "sqlite:///data/brainchulo.db") self.andromeda_url = os.getenv("ANDROMEDA_URL", "http://0.0.0.0:9000") self.use_flow_agents = os.getenv("USE_FLOW_AGENTS", "false") == "true" def load_config(): return Settings()
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/__init__.py
app/__init__.py
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/guidance_tooling/guidance_programs/tools.py
app/guidance_tooling/guidance_programs/tools.py
from dotenv import load_dotenv from langchain.chains import RetrievalQA from langchain.embeddings import HuggingFaceEmbeddings, HuggingFaceInstructEmbeddings from langchain import HuggingFacePipeline from colorama import Fore, Style import re from langchain.vectorstores import Chroma from langchain.docstore.document import Document from langchain.text_splitter import CharacterTextSplitter, TokenTextSplitter, RecursiveCharacterTextSplitter import os from langchain.llms import LlamaCpp load_dotenv() TEST_FILE = os.getenv("TEST_FILE") EMBEDDINGS_MODEL = os.getenv("EMBEDDINGS_MODEL") EMBEDDINGS_MAP = { **{name: HuggingFaceInstructEmbeddings for name in ["hkunlp/instructor-xl", "hkunlp/instructor-large"]}, **{name: HuggingFaceEmbeddings for name in ["all-MiniLM-L6-v2", "sentence-t5-xxl"]} } model_type = os.environ.get('MODEL_TYPE') model_path = os.environ.get('MODEL_PATH') model_n_ctx =1000 target_source_chunks = os.environ.get('TARGET_SOURCE_CHUNKS') n_gpu_layers = os.environ.get('N_GPU_LAYERS') use_mlock = os.environ.get('USE_MLOCK') n_batch = os.environ.get('N_BATCH') if os.environ.get('N_BATCH') != None else 512 callbacks = [] qa_prompt = "" CHROMA_SETTINGS = {} # Set your Chroma settings here def clean_text(text): # Remove line breaksRetrievalQA text = text.replace('\n', ' ') # Remove special characters text = re.sub(r'[^\w\s]', '', text) return text def load_unstructured_document(document: str) -> list[Document]: with open(document, 'r') as file: text = file.read() title = os.path.basename(document) return [Document(page_content=text, metadata={"title": title})] def split_documents(documents: list[Document], chunk_size: int = 250, chunk_overlap: int = 20) -> list[Document]: text_splitter = RecursiveCharacterTextSplitter(chunk_size=chunk_size, chunk_overlap=chunk_overlap) return text_splitter.split_documents(documents) def ingest_file(file_path): # Load unstructured document documents = load_unstructured_document(file_path) # Split documents into chunks documents = split_documents(documents, chunk_size=250, chunk_overlap=100) # Determine the embedding model to use EmbeddingsModel = EMBEDDINGS_MAP.get(EMBEDDINGS_MODEL) if EmbeddingsModel is None: raise ValueError(f"Invalid embeddings model: {EMBEDDINGS_MODEL}") model_kwargs = {"device": "cuda:0"} if EmbeddingsModel == HuggingFaceInstructEmbeddings else {} embedding = EmbeddingsModel(model_name=EMBEDDINGS_MODEL, model_kwargs=model_kwargs) # Store embeddings from the chunked documents vectordb = Chroma.from_documents(documents=documents, embedding=embedding) retriever = vectordb.as_retriever(search_kwargs={"k":4}) print(file_path) print(retriever) return retriever def load_tools(): #llm = LlamaCpp(model_path=model_path, n_ctx=model_n_ctx, callbacks=callbacks, verbose=False,n_gpu_layers=n_gpu_layers, use_mlock=use_mlock,top_p=0.9, n_batch=n_batch) def ingest_file(file_path): # Load unstructured document documents = load_unstructured_document(file_path) # Split documents into chunks documents = split_documents(documents, chunk_size=120, chunk_overlap=20) # Determine the embedding model to use EmbeddingsModel = EMBEDDINGS_MAP.get(EMBEDDINGS_MODEL) if EmbeddingsModel is None: raise ValueError(f"Invalid embeddings model: {EMBEDDINGS_MODEL}") model_kwargs = {"device": "cuda:0"} if EmbeddingsModel == HuggingFaceInstructEmbeddings else {} embedding = EmbeddingsModel(model_name=EMBEDDINGS_MODEL, model_kwargs=model_kwargs) # Store embeddings from the chunked documents vectordb = Chroma.from_documents(documents=documents, embedding=embedding) retriever = vectordb.as_retriever(search_kwargs={"k":4}) print(file_path) print(retriever) return retriever, file_path file_path = TEST_FILE retriever, title = ingest_file(file_path) dict_tools = { 'File Ingestion': ingest_file, } return dict_tools
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/tools/web_access.py
app/tools/web_access.py
from settings import Settings from tools.base import BaseTool config = Settings.load_config() class WebAccess(BaseTool): url: str def commands(self): return {}
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/tools/document_memory.py
app/tools/document_memory.py
from tools.base import BaseTool from tools.utils import _build_conversation_filter from typing import Dict, Any class DocumentSearchTool(BaseTool): def __init__(self, conversation_id: str, tool_context: Dict[str, Any], name: str = "Document Search"): required_context_keys = ["vector_store_docs", "k"] super().__init__(conversation_id, name, tool_context=tool_context, required_context_keys=required_context_keys) def short_description(self) -> str: return "Useful for when you need to answer questions about documents that were uploaded by the user." def few_shot_examples(self) -> str: return """Question: What's your name? Thought: I should search my name in the documents. Action: Document Search Action Input: What's my name? Observation: You're an AI. You don't have a name. Thought: I should answer that I don't have a name. Final Answer: As an AI, I don't have a name, at least not in the human sense.""" def __call__(self, search_input: Dict[str, str]) -> str: """Executes the tool Search for the given input in the vector store and return the top k most similar documents with their scores. This function is used as a helper function for the SearchLongTermMemory tool Args: search_input (str): The input to search for in the vector store. Returns: List[Tuple[str, float]]: A list of tuples containing the document text and their similarity score. """ k = self.tool_context["k"] filter_= _build_conversation_filter(conversation_id=self.conversation_id) docs = self.tool_context["vector_store_docs"].similarity_search_with_score( search_input, k=k, filter=filter_ ) return [{"document_content": doc[0].page_content, "similarity": doc[1]} for doc in docs]
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/tools/conversation_memory.py
app/tools/conversation_memory.py
from tools.base import BaseTool from tools.utils import _build_conversation_filter from typing import Dict, Any class ConversationSearchTool(BaseTool): def __init__(self, conversation_id: str, tool_context: Dict[str, Any], name: str = "Conversation Search"): required_context_keys = ["vector_store_convs", "k"] super().__init__(conversation_id, name, tool_context=tool_context, required_context_keys=required_context_keys) def short_description(self) -> str: return "A tool to search in your memory previous conversations with this user." def few_shot_examples(self) -> str: return """Question: What's your name? Thought: I should search my name in the previous conversations. Action: Conversation Search Action Input: What's my name? Observation: User: I'd like to give you a better name. Bot: How would you like to call me? User: I'd like to call you Joseph. Bot: Alright, you may call me Joseph from now on. Thought: The user wants to call me Joseph. Final Answer: As I recall from a previous conversation, you call me Joseph.""" def __call__(self, search_input): """ Search for the given input in the vector store and return the top 10 most similar documents with their scores. This function is used as a helper function for the SearchLongTermMemory tool Args: search_input (str): The input to search for in the vector store. Returns: List[Tuple[str, float]]: A list of tuples containing the document text and their similarity score. """ k = self.tool_context["k"] filter_= _build_conversation_filter(conversation_id=self.conversation_id) docs = self.tool_context["vector_store_convs"].similarity_search_with_score( search_input, k=5, filter=filter_ ) return [{"document_content": doc[0].page_content, "similarity": doc[1]} for doc in docs]
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/tools/utils.py
app/tools/utils.py
from typing import Optional def _build_conversation_filter(conversation_id: Optional[str]): if conversation_id is not None: return {"conversation_id": conversation_id} else: return {}
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/tools/base.py
app/tools/base.py
from pyparsing import abstractmethod from typing import Dict, List, Any class ToolFactory(): """Instantiates tools with a reference to the current conversation_id""" def __init__(self, list_of_tool_classes: List["BaseTool"]) -> None: self._tool_class_references = list_of_tool_classes def build_tools(self, conversation_id, context: Dict[str, Any]) -> Dict[str, "BaseTool"]: resolved_tools = {} for tool_class in self._tool_class_references: tool = tool_class(conversation_id, context) resolved_tools[tool.name] = tool return resolved_tools class BaseTool: """Base interface expected of tools""" def __init__(self, conversation_id: str, name: str, tool_context: Dict[str, Any], required_context_keys: List[str]): """Stores a reference to the conversation ID. Avoiding injecting expensive operations in the __init__ method of the subclasses. If you need to do something expensive, use a Singleton or redesign this Tool interface. The reason being is that these Tools are instantied **per processed message**, so the constructor must be cheap to execute. """ self.conversation_id = conversation_id self.name = name self.tool_context = tool_context self._validate_context_keys(required_context_keys, tool_context) def _validate_context_keys(self, keys, context): for key in keys: if key not in context: raise TypeError(f"This instance of {self.__class__.__name__} requires variable {key} in context.") @abstractmethod def short_description(self) -> str: """Returns a short description of the tool.""" raise NotImplementedError() @abstractmethod def few_shot_examples(self) -> str: """Returns few """ raise NotImplementedError() @abstractmethod def __call__(self, variables: Dict[str, str]) -> str: """Executes the tool""" raise NotImplementedError()
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/flow/flow.py
app/flow/flow.py
from abc import abstractmethod from colorama import Fore from typing import Tuple, List, Union, Dict from andromeda_chain import AndromedaChain, AndromedaPrompt, AndromedaResponse from agents.base import color_print from prompts.guidance_choice import CHOICE_PROMPT from copy import deepcopy class Node: def __init__(self, name) -> None: self.name = name @abstractmethod def run(self, variables) -> Union[AndromedaResponse, Dict[str, str]]: raise NotImplementedError() class HiddenNode(Node): """Classes of nodes that can be executed in the background, without expanding the context history for the agent. """ class ChoiceNode(HiddenNode): def __init__(self, name, choices: List[str], max_decisions, force_exit_on) -> None: super().__init__(name) self.choices = choices self.max_decisions = max_decisions self.force_exit_on=force_exit_on self.decisions_made = 0 def run(self, chain: AndromedaChain, variables) -> str: if self.decisions_made >= self.max_decisions: return self.force_exit_on history = "" if "history" in variables: history = variables["history"] result = chain.run_guidance_prompt( CHOICE_PROMPT, input_vars={ "history": history, "valid_choices": self.choices } ) self.decisions_made += 1 return result.result_vars["choice"] def set_next(self, next_): self._next = next_ def next(self): return self._next class ToolNode(Node): def __init__(self, name, tool_callback, variable_name = "observation") -> None: super().__init__(name) self.tool_callback = tool_callback self.variable_name = variable_name def run(self, variables) -> Dict[str, str]: return {self.variable_name: self.tool_callback(variables)} def set_next(self, next_): self._next = next_ def next(self): return self._next class PromptNode(Node): def __init__(self, name, prompt: AndromedaPrompt) -> None: super().__init__(name) self.prompt = prompt self._next = None def run(self, chain: AndromedaChain, variables) -> AndromedaResponse: input_dict = {} for var_ in self.prompt.input_vars: value = variables[var_] input_dict[var_] = value return chain.run_guidance_prompt( self.prompt, input_vars=input_dict ) def set_next(self, next_): self._next = next_ def next(self): return self._next class StartNode(PromptNode): def __init__(self, name, prompt: AndromedaPrompt, choices: List[str]) -> None: super().__init__(name, prompt) self.choices = choices def run(self, chain: AndromedaChain, variables) -> Tuple[str, AndromedaResponse]: response = super().run(chain, variables) choice = response.result_vars.pop("choice") return self.choices[choice], response class Flow: def __init__(self, nodes: List[PromptNode]) -> None: assert len(nodes) > 0 self.nodes = nodes def execute(self, chain, query: str, variables: Dict[str, str], return_key="final_answer"): node = self.nodes[0] variables["query"] = query history = "" while node: color_print(f"---> On node {node.name}", Fore.RED) debug_vars = deepcopy(variables) if "history" in debug_vars: debug_vars.pop("history") if "prompt_start" in debug_vars: debug_vars.pop("prompt_start") if isinstance(node, StartNode): color_print(f"Executing start node {node.name} with variables: {debug_vars}", Fore.YELLOW) choice, result = node.run(chain, variables) color_print(f"Node result: {result.result_vars}", Fore.GREEN) history = result.expanded_generation variables.update(result.result_vars) variables["history"] = history node = _find_node_by_name(choice, self.nodes, node) color_print(f"Choice decided to jump to node {node.name}", Fore.RED) elif isinstance(node, PromptNode): color_print(f"Executing node {node.name} with variables: {debug_vars}", Fore.YELLOW) result = node.run(chain, variables) color_print(f"Node result: {result.result_vars}", Fore.GREEN) history = result.expanded_generation # Merge contexts variables.update(result.result_vars) variables["history"] = history node = node.next() elif isinstance(node, ToolNode): color_print(f"Executing tool node {node.name} with variables: {debug_vars}", Fore.YELLOW) tool_result = node.run(variables) variables.update(tool_result) node = node.next() elif isinstance(node, ChoiceNode): color_print(f"Executing choice node {node.name} with variables: {debug_vars}", Fore.YELLOW) choice = node.run(chain, variables) node = _find_node_by_name(choice, self.nodes, node) color_print(f"Choice decided to jump to node {node.name}", Fore.RED) else: raise ValueError(f"Invalid node class: {type(node)}") color_print(f"History: {history}", Fore.CYAN) color_print(f"Flow ended, returning variable '{return_key}'.", Fore.GREEN) return variables[return_key] def _find_node_by_name(choice, nodes, current_node): new_node = None for n in nodes: if n.name == choice: new_node = n break if not new_node: raise ValueError(f"Choice {choice} led to limbo! Please choose the name of another node in the flow.") if new_node == current_node: raise ValueError(f"Choice {choice} led to an infinite loop on itself! Make sure choice node hop to itself.") return new_node
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/flow/__init__.py
app/flow/__init__.py
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/agents/chain_of_thoughts.py
app/agents/chain_of_thoughts.py
from agents.base import BaseAgent from guidance_tooling.guidance_programs.tools import ingest_file from guidance_tooling.guidance_programs.tools import clean_text from langchain.llms import LlamaCpp import os import time import guidance from colorama import Fore, Style from langchain.chains import RetrievalQA from langchain.llms import LlamaCpp from prompt_templates.qa_agent import * from settings import load_config import re config = load_config() llm = None valid_answers = ['Action', 'Final Answer'] valid_tools = ["Check Question", "Google Search"] TEST_FILE = os.getenv("TEST_FILE") TEST_MODE = os.getenv("TEST_MODE") ETHICS = os.getenv("ETHICS") QA_MODEL = os.getenv("MODEL_PATH") model_path = config.model_path if ETHICS == "ON": agent_template = QA_ETHICAL_AGENT else: agent_template = QA_AGENT def get_llm(): global llm if llm is None: print("Loading qa model...") model_path =QA_MODEL model_n_ctx =1000 n_gpu_layers = 500 use_mlock = 0 n_batch = os.environ.get('N_BATCH') if os.environ.get('N_BATCH') != None else 512 callbacks = [] llm = LlamaCpp(model_path=model_path, n_ctx=model_n_ctx, callbacks=callbacks, verbose=False,n_gpu_layers=n_gpu_layers, use_mlock=use_mlock,top_p=0.9, n_batch=n_batch) return llm class ChainOfThoughtsAgent(BaseAgent): def __init__(self, guidance, llama_model, llama_model2): self.guidance = guidance # We first load the model in charge of reasoning along the guidance program self.llama_model = llama_model # We then load the model in charge of correctly identifying the data within the context and provide an answer self.llama_model2 = llama_model2 def print_stage(self, stage_name, message): print(Fore.CYAN + Style.BRIGHT + f"Entering {stage_name} round" + Style.RESET_ALL) time.sleep(1) print(Fore.GREEN + Style.BRIGHT + message + Style.RESET_ALL) def searchQA(self, t): return self.checkQuestion(self.question, self.context) def checkQuestion(self, question: str, context): context = context if TEST_MODE == "ON": print(Fore.GREEN + Style.BRIGHT + "No document loaded in conversation. Falling back on test file." + Style.RESET_ALL) question = question.replace("Action Input: ", "") qa = RetrievalQA.from_chain_type(llm=self.llm, chain_type="stuff", retriever=self.retriever, return_source_documents=True) answer_data = qa({"query": question}) if 'result' not in answer_data: print(f"\033[1;31m{answer_data}\033[0m") return "Issue in retrieving the answer." context_documents = answer_data['source_documents'] context = " ".join([clean_text(doc.page_content) for doc in context_documents]) print(Fore.WHITE + Style.BRIGHT + "Printing langchain context..." + Style.RESET_ALL) print(Fore.WHITE + Style.BRIGHT + context + Style.RESET_ALL) return context def ethics_check(self, question, ethics_prompt): ethics_program = self.guidance(ethics_prompt) return ethics_program(question=question) def query_identification(self, question, conversation_prompt): guidance.llm = self.llama_model conversation_program = self.guidance(conversation_prompt) return conversation_program(question=question) def phatic_answer(self, question, history, phatic_prompt): phatic_program = self.guidance(phatic_prompt) return phatic_program(question=question, history=history) def data_retrieval(self, question): if self.llama_model2 is not None: guidance.llm = self.llama_model2 referential_program = self.guidance(referential_prompt) referential_round = referential_program(question=question, search=self.searchQA) return referential_round def answer_question(self, question, answer_prompt): if self.llama_model2 is not None: guidance.llm = self.llama_model2 answer_program = self.guidance(answer_prompt) answer_round = answer_program(question=question, search=self.searchQA) return answer_round["final_answer"] def run(self, query: str, context, history) -> str: self.question = query self.context = context self.history = history print(Fore.GREEN + Style.BRIGHT + "Starting guidance agent..." + Style.RESET_ALL) conversation_round= self.query_identification(self.question , conversation_prompt) if conversation_round["query_type"] == "Phatic": self.print_stage("answering", "User query identified as phatic") phatic_round = self.phatic_answer(self.question , history, phatic_prompt) return phatic_round["phatic_answer"] self.print_stage("data retrieval", "User query identified as referential") referential_round = self.data_retrieval(self.question ) if referential_round["answerable"] == "Yes": self.print_stage("answering", "Matching information found") return self.answer_question(self.question, answer_prompt) else: return "I don't have enough information to answer."
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/agents/flow_cot.py
app/agents/flow_cot.py
from typing import Dict, Callable from andromeda_chain import AndromedaChain from agents.flow_based import BaseFlowAgent from flow.flow import Flow, PromptNode, ToolNode, ChoiceNode, StartNode from prompts.flow_guidance_cot import FlowChainOfThoughts, PROMPT_START_STRING class ChainOfThoughtsFlowAgent(BaseFlowAgent): def __init__(self, andromeda: AndromedaChain, tools: Dict[str, Callable[[str], str]]): def execute_tool(variables): action_name = variables["tool_name"] action_input = variables["act_input"] return self.do_tool(action_name, action_input) start = StartNode("start", FlowChainOfThoughts.flow_prompt_start, { "Action": "choose_action", "Final Answer": "final_prompt" }) thought = PromptNode("thought", FlowChainOfThoughts.thought_gen) choose_action = PromptNode("choose_action", FlowChainOfThoughts.choose_action) perform_action = PromptNode("perform_action", FlowChainOfThoughts.action_input) execute_tool_node = ToolNode("execute_tool", execute_tool) decide = ChoiceNode("decide", ["thought", "final_prompt"], max_decisions=3, force_exit_on="final_prompt") final = PromptNode("final_prompt", FlowChainOfThoughts.final_prompt) thought.set_next(choose_action) choose_action.set_next(perform_action) perform_action.set_next(execute_tool_node) execute_tool_node.set_next(decide) flow = Flow( [start, thought, choose_action, perform_action, execute_tool_node, decide, final] ) super().__init__(andromeda, tools, flow) self.valid_answers = ["Action", "Final Answer"] self.prepare_start_prompt(PROMPT_START_STRING) def run(self, query: str) -> str: if not self.prepared_prompt: raise TypeError("Your inherinted BaseFlowAgent class must call 'prepare_start_prompt' in it's constructor.") return super().run(query, variables={ "prompt_start": self.prepared_prompt, "question": query, "valid_tools": self.valid_tools, "valid_answers": self.valid_answers, })
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/agents/flow_based.py
app/agents/flow_based.py
# import the necessary libraries from typing import Dict, Callable from agents.base import BaseAgent from colorama import Style from andromeda_chain import AndromedaChain from flow.flow import Flow def color_print(msg, color): print(color + Style.BRIGHT + msg + Style.RESET_ALL) class BaseFlowAgent(BaseAgent): """Base Flow Agent. Implements a graph that the agents execute. """ def __init__(self, andromeda: AndromedaChain, tools: Dict[str, Callable[[str], str]], flow: Flow): super().__init__(andromeda, tools) self.flow = flow def run(self, query: str, variables=None) -> str: if not variables: variables = {} return self.flow.execute(self.andromeda, query, variables)
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/agents/__init__.py
app/agents/__init__.py
from agents.chain_of_thoughts import ChainOfThoughtsAgent from agents.flow_cot import ChainOfThoughtsFlowAgent
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/agents/base.py
app/agents/base.py
# import the necessary libraries from abc import abstractmethod from colorama import Style, Fore from typing import List from andromeda_chain import AndromedaChain from tools.base import BaseTool def color_print(msg, color): print(color + Style.BRIGHT + msg + Style.RESET_ALL, flush=True) class BaseAgent: """Base Agent. Nothing too exciting here. """ def __init__(self, andromeda: AndromedaChain, tools: List[BaseTool]): self.andromeda = andromeda self.tools = tools @abstractmethod def run(self, query: str) -> str: raise NotImplementedError() def do_tool(self, tool_name, act_input): color_print(f"Using tool: {tool_name}", Fore.GREEN) result = self.tools[tool_name](act_input) color_print(f"Tool result: {result}", Fore.BLUE) return result def _build_tool_description_line(self, tool: BaseTool): return f"{tool.name}: {tool.short_description()}" def prepare_start_prompt(self, prompt_start_template): tools = [item for _, item in self.tools.items()] tools_descriptions = "\n".join([self._build_tool_description_line(tool) for tool in tools]) action_list = str([tool.name for tool in tools]).replace('"', "") few_shot_examples = "\n".join( [ f"Example {idx}:\n{tool.few_shot_examples()}" for idx, tool in enumerate(tools) ] ) self.valid_tools = [tool.name for tool in tools] self.prepared_prompt = prompt_start_template.format( tools_descriptions=tools_descriptions, action_list=action_list, few_shot_examples=few_shot_examples )
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/models/all.py
app/models/all.py
from datetime import datetime from typing import Optional, List from sqlmodel import SQLModel, Field, Relationship from importlib import import_module NAMING_CONVENTION = { "ix": "ix_%(column_0_label)s", "uq": "uq_%(table_name)s_%(column_0_name)s", "ck": "ck_%(table_name)s_%(constraint_name)s", "fk": "fk_%(table_name)s_%(column_0_name)s_%(referred_table_name)s", "pk": "pk_%(table_name)s", } metadata = SQLModel.metadata metadata.naming_convention = NAMING_CONVENTION def load_plugin_tables(): """ Loads all plugins in the plugins directory that have a database.py file and merge their metadata. :return: None """ import os plugins_dir = os.path.dirname(__file__) + '/../plugins' for plugin_name in os.listdir(plugins_dir): if plugin_name.startswith('_'): continue # Skip hidden files plugin_dir = os.path.join(plugins_dir, plugin_name) if not os.path.exists(os.path.join(plugin_dir, 'database.py')): continue # Skip plugins without a database.py file import_module(f'plugins.{plugin_name}.database') class ConversationBase(SQLModel): created_at: datetime = Field(default_factory=datetime.utcnow) title: Optional[str] class Conversation(ConversationBase, table=True): id: Optional[int] = Field(default=None, primary_key=True) messages: List["Message"] = Relationship(back_populates="conversation") class ConversationRead(ConversationBase): id: int class MessageBase(SQLModel): text: str is_user: bool conversation_id: Optional[int] = Field(default=None, foreign_key="conversation.id") rating: int = 0 # -1, 0, 1 created_at: datetime = Field(default_factory=datetime.utcnow, nullable=False) class Message(MessageBase, table=True): id: Optional[int] = Field(default=None, primary_key=True) conversation: Optional[Conversation] = Relationship(back_populates="messages") class MessageRead(MessageBase): id: int class ConversationWithMessages(ConversationRead): messages: List[MessageRead] = [] load_plugin_tables()
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/memory/chroma_memory.py
app/memory/chroma_memory.py
import os from typing import Any, Iterable, List, Optional, Type from memory.base import BaseMemory from langchain import vectorstores from langchain.embeddings import ( HuggingFaceInstructEmbeddings, HuggingFaceEmbeddings, ) from langchain.docstore.document import Document from settings import load_config, logger config = load_config() class Chroma(BaseMemory): vector_store: Optional[Type[vectorstores.Chroma]] collection_name: Optional[str] def __init__(self, **kwargs: Any): super().__init__(**kwargs) embeddings_model_name = config.embeddings_model.split("/")[-1] EmbeddingsModel = config.embeddings_map.get(embeddings_model_name) if EmbeddingsModel is None: raise ValueError(f"Invalid embeddings model: {config.embeddings_model}") kwargs = {"model_name": config.embeddings_model} if EmbeddingsModel == HuggingFaceInstructEmbeddings: kwargs["model_kwargs"] = {"device": "cuda"} embeddings = EmbeddingsModel(**kwargs) persist_directory = os.path.join(os.getcwd(), "data", config.memories_path) # Create directory if it doesn't exist os.makedirs(persist_directory, exist_ok=True, mode=0o777) self.vector_store = vectorstores.Chroma( collection_name=self.collection_name, embedding_function=embeddings, persist_directory=persist_directory, ) self.setup_index() def setup_index(self): collection = self.vector_store._client.get_collection(self.collection_name) if len(collection.get()['ids']) < 6: self.add_texts( ["Hello", "world!", "this", "is", "a", "test"], ids=["1", "2", "3", "4", "5", "6"], )
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/memory/base.py
app/memory/base.py
from typing import Any, Iterable, List, Optional, Dict, Tuple, Type from pydantic import BaseModel from pyparsing import abstractmethod from langchain.docstore.document import Document from langchain.vectorstores import VectorStore class BaseMemory(BaseModel): collection_name: Optional[str] vector_store: Optional[Type[VectorStore]] def __init__(self, collection_name: str = "default_collection"): # init super class super().__init__() self.collection_name = collection_name def add_texts( self, texts: Iterable[str], metadatas: Optional[List[dict]] = None, ids: Optional[List[str]] = None, **kwargs: Any, ) -> List[str]: response = self.vector_store.add_texts(texts, metadatas, ids, **kwargs) self.vector_store._client.persist() return response def add_documents(self, documents: list[Document]): texts = [doc.page_content for doc in documents] metadatas = [doc.metadata for doc in documents] return self.add_texts(texts, metadatas) def similarity_search( self, query: str, k: int = 4, filter: Optional[Dict[str, str]] = None, **kwargs: Any, ) -> List[Document]: return self.vector_store.similarity_search(query, k=k, **kwargs) def similarity_search_by_vector( self, embedding: List[float], k: int = 4, filter: Optional[Dict[str, str]] = None, **kwargs: Any, ) -> List[Document]: return self.vector_store.similarity_search_by_vector( embedding, k=k, **kwargs) def similarity_search_with_score( self, query: str, k: int = 4, filter: Optional[Dict[str, str]] = None, **kwargs: Any, ) -> List[Tuple[Document, float]]: return self.vector_store.similarity_search_with_score( query, k=k, **kwargs) def max_marginal_relevance_search_by_vector( self, embedding: List[float], k: int = 4, fetch_k: int = 20, lambda_mult: float = 0.5, filter: Optional[Dict[str, str]] = None, **kwargs: Any, ) -> List[Document]: return self.vector_store.max_marginal_relevance_search_by_vector( embedding, k=k, fetch_k=fetch_k, lambda_mult=lambda_mult, **kwargs) def max_marginal_relevance_search( self, query: str, k: int = 4, fetch_k: int = 20, lambda_mult: float = 0.5, filter: Optional[Dict[str, str]] = None, **kwargs: Any, ) -> List[Document]: return self.vector_store.max_marginal_relevance_search( query, k=k, fetch_k=fetch_k, lambda_mult=lambda_mult, **kwargs) def delete_collection(self) -> None: return self.vector_store.delete_collection() def persist(self) -> None: return self.vector_store.persist() def update_document(self, document_id: str, document: Document) -> None: return self.vector_store.update_document(document_id, document) def get_store(self) -> Document: return self.vector_store
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/plugins/__init__.py
app/plugins/__init__.py
import importlib import pkgutil import os def load_plugins(app): # Get the current directory plugins_dir = os.path.dirname(__file__) # Iterate over the files in the plugins directory for _, plugin_name, is_package in pkgutil.iter_modules([plugins_dir]): if is_package: # Dynamically import the routes module from each plugin plugin_module = importlib.import_module(f"plugins.{plugin_name}.routes") # Get the router from the plugin's routes module router = plugin_module.router # Include the plugin's routes in the main FastAPI app app.include_router(router, prefix=f"/plugins/{plugin_name}")
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/plugins/sample_plugin/database.py
app/plugins/sample_plugin/database.py
from datetime import datetime from typing import Optional, List from sqlmodel import SQLModel, Field, Relationship from models.all import Conversation class SamplePluginModelBase(SQLModel): """A base model for SamplePluginModel""" # __tablename__: str = 'sample_plugin_model' created_at: datetime = Field(default_factory=datetime.utcnow) title: Optional[str] conversation_id: Optional[int] = Field(default=None, foreign_key="conversation.id") class SamplePluginModel(SamplePluginModelBase, table=True): """A model for SamplePlugin""" id: Optional[int] = Field(default=None, primary_key=True) class SamplePluginModelRead(SamplePluginModelBase): """A read model for SamplePlugin""" id: int
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/plugins/sample_plugin/__init__.py
app/plugins/sample_plugin/__init__.py
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/plugins/sample_plugin/routes.py
app/plugins/sample_plugin/routes.py
from fastapi import APIRouter router = APIRouter() @router.get("/") def sample_plugin_route(): return {"message": "Sample plugin route"} # Additional routes and functionality for the sample plugin
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/migrations/env.py
app/migrations/env.py
from logging.config import fileConfig from sqlalchemy import engine_from_config from sqlalchemy import pool from alembic import context from models import all # this is the Alembic Config object, which provides # access to the values within the .ini file in use. config = context.config # Interpret the config file for Python logging. # This line sets up loggers basically. if config.config_file_name is not None: fileConfig(config.config_file_name) # add your model's MetaData object here # for 'autogenerate' support # from myapp import mymodel # target_metadata = mymodel.Base.metadata target_metadata = all.metadata # other values from the config, defined by the needs of env.py, # can be acquired: # my_important_option = config.get_main_option("my_important_option") # ... etc. def run_migrations_offline() -> None: """Run migrations in 'offline' mode. This configures the context with just a URL and not an Engine, though an Engine is acceptable here as well. By skipping the Engine creation we don't even need a DBAPI to be available. Calls to context.execute() here emit the given string to the script output. """ url = config.get_main_option("sqlalchemy.url") context.configure( url=url, target_metadata=target_metadata, literal_binds=True, dialect_opts={"paramstyle": "named"}, ) with context.begin_transaction(): context.run_migrations() def run_migrations_online() -> None: """Run migrations in 'online' mode. In this scenario we need to create an Engine and associate a connection with the context. """ connectable = engine_from_config( config.get_section(config.config_ini_section, {}), prefix="sqlalchemy.", poolclass=pool.NullPool, ) with connectable.connect() as connection: context.configure( connection=connection, target_metadata=target_metadata, render_as_batch=True, user_module_prefix="sqlmodel.sql.sqltypes." ) with context.begin_transaction(): context.run_migrations() if context.is_offline_mode(): run_migrations_offline() else: run_migrations_online()
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/migrations/versions/51a9f1ab4b8b_add_sample_plugin_demo_model.py
app/migrations/versions/51a9f1ab4b8b_add_sample_plugin_demo_model.py
"""Add Sample Plugin Demo Model Revision ID: 51a9f1ab4b8b Revises: 804b20c9599c Create Date: 2023-06-04 21:30:37.777274 """ from alembic import op import sqlalchemy as sa import sqlmodel.sql.sqltypes # revision identifiers, used by Alembic. revision = '51a9f1ab4b8b' down_revision = '804b20c9599c' branch_labels = None depends_on = None def upgrade() -> None: # ### commands auto generated by Alembic - please adjust! ### op.create_table('samplepluginmodel', sa.Column('created_at', sa.DateTime(), nullable=False), sa.Column('title', sqlmodel.sql.sqltypes.AutoString(), nullable=True), sa.Column('conversation_id', sa.Integer(), nullable=True), sa.Column('id', sa.Integer(), nullable=False), sa.ForeignKeyConstraint(['conversation_id'], ['conversation.id'], name=op.f('fk_samplepluginmodel_conversation_id_conversation')), sa.PrimaryKeyConstraint('id', name=op.f('pk_samplepluginmodel')) ) # ### end Alembic commands ### def downgrade() -> None: # ### commands auto generated by Alembic - please adjust! ### op.drop_table('samplepluginmodel') # ### end Alembic commands ###
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/migrations/versions/804b20c9599c_initial.py
app/migrations/versions/804b20c9599c_initial.py
"""initial Revision ID: 804b20c9599c Revises: Create Date: 2023-05-31 20:06:36.912608 """ from alembic import op import sqlalchemy as sa import sqlmodel.sql.sqltypes # revision identifiers, used by Alembic. revision = '804b20c9599c' down_revision = None branch_labels = None depends_on = None def upgrade() -> None: # ### commands auto generated by Alembic - please adjust! ### op.create_table('conversation', sa.Column('created_at', sa.DateTime(), nullable=False), sa.Column('title', sqlmodel.sql.sqltypes.AutoString(), nullable=True), sa.Column('id', sa.Integer(), nullable=False), sa.PrimaryKeyConstraint('id', name=op.f('pk_conversation')) ) op.create_table('message', sa.Column('text', sqlmodel.sql.sqltypes.AutoString(), nullable=False), sa.Column('is_user', sa.Boolean(), nullable=False), sa.Column('conversation_id', sa.Integer(), nullable=True), sa.Column('rating', sa.Integer(), nullable=False), sa.Column('created_at', sa.DateTime(), nullable=False), sa.Column('id', sa.Integer(), nullable=False), sa.ForeignKeyConstraint(['conversation_id'], ['conversation.id'], name=op.f('fk_message_conversation_id_conversation')), sa.PrimaryKeyConstraint('id', name=op.f('pk_message')) ) # ### end Alembic commands ### def downgrade() -> None: # ### commands auto generated by Alembic - please adjust! ### op.drop_table('message') op.drop_table('conversation') # ### end Alembic commands ###
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/prompts/flow_guidance_cot.py
app/prompts/flow_guidance_cot.py
from andromeda_chain.prompt import AndromedaPrompt PROMPT_START_STRING = """You're an AI assistant with access to tools. You're nice and friendly, and try to answer questions to the best of your ability. You have access to the following tools. {tools_descriptions} Strictly use the following format: Question: the input question you must answer Thought: you should always think about what to do Action: the action to take, should be one of {action_list} Action Input: the input to the action, should be a question. Observation: the result of the action ... (this Thought/Action/Action Input/Observation can repeat N times) Thought: I now know the final answer Final Answer: the final answer to the original input question When chatting with the user, you can search information using your tools. {few_shot_examples} Now your turn. Question:""" class FlowChainOfThoughts: choose_action = AndromedaPrompt( name="cot_choose_action", prompt_template="""{{history}} Action: {{select 'tool_name' options=valid_tools}}""", guidance_kwargs={}, input_vars=["history", "valid_tools"], output_vars=["tool_name"], ) action_input = AndromedaPrompt( name="cot_action_input", prompt_template="""{{history}}{{gen 'act_input' stop='\\n'}}""", guidance_kwargs={}, input_vars=["history"], output_vars=["act_input"], ) thought_gen = AndromedaPrompt( name="cot_thought_gen", prompt_template="""{{history}} Observation: {{observation}} Thought: {{gen 'thought' stop='\\n'}}""", guidance_kwargs={}, input_vars=["history", "observation"], output_vars=["thought"], ) final_prompt = AndromedaPrompt( name="cot_final", prompt_template="""{{history}} Thought: I should now reply the user with what I thought and gathered. Final Answer: {{gen 'final_answer' stop='\\n'}}""", guidance_kwargs={}, input_vars=["history"], output_vars=["final_answer"], ) flow_prompt_start = AndromedaPrompt( name="cot_flow_prompt_start", prompt_template="""{{prompt_start}} {{question}} Think carefully about what you should do next. Take an action or provide a final answer to the user. Thought: {{gen 'thought' stop='\\n'}}{{#block hidden=True}} {{select 'choice' logprobs='logprobs' options=valid_answers}} :{{/block}}""", guidance_kwargs={}, input_vars=["prompt_start", "question", "valid_answers"], output_vars=["thought", "choice"], )
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/prompts/guidance_check_question.py
app/prompts/guidance_check_question.py
from andromeda_chain.prompt import AndromedaPrompt PROMPT_CHECK_QUESTION = AndromedaPrompt( name="start-prompt", prompt_template="""You MUST answer with 'yes' or 'no'. Given the following pieces of context, determine if there are any elements related to the question in the context. Don't forget you MUST answer with 'yes' or 'no'. Context: {{context}} Question: Are there any elements related to ""{{question}}"" in the context? {{select 'answer' options=['yes', 'no']}} """, guidance_kwargs={}, input_vars=["context", "question"], output_vars=["answer"], )
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/prompts/__init__.py
app/prompts/__init__.py
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/prompts/guidance_choice.py
app/prompts/guidance_choice.py
from andromeda_chain.prompt import AndromedaPrompt CHOICE_PROMPT = AndromedaPrompt( name="choice_prompt", prompt_template = """{{history}} You must now choose an option out of the {{valid_choices}}. Remember that it must be coherent with your last thought. {{select 'choice' logprobs='logprobs' options=valid_choices}}: """, input_vars=["history", "valid_choices"], output_vars=["choice"], )
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/conversations/document_based.py
app/conversations/document_based.py
from langchain.document_loaders import TextLoader from memory.chroma_memory import Chroma from langchain.text_splitter import RecursiveCharacterTextSplitter from colorama import Fore, Style from guidance_tooling.guidance_programs.tools import clean_text from andromeda_chain import AndromedaChain from agents import ChainOfThoughtsAgent from settings import logger, load_config import guidance import os config = load_config() dict_tools = None llama_model = None llama_model2 = None guidance_reasoning_model_path = config.guidance_reasoning_model_path guidance_extraction_model_path = config.guidance_extraction_model_path TEST_MODE = os.getenv("TEST_MODE") GUIDANCE_MODEL = os.getenv("GUIDANCE_MODEL_PATH") def get_llama_model(): global llama_model if llama_model is None: print("Loading main guidance model...") llama_model = guidance.llms.LlamaCpp( model = guidance_reasoning_model_path, tokenizer = "openaccess-ai-collective/manticore-13b-chat-pyg", before_role = "<|", after_role = "|>", n_gpu_layers=300, n_threads=12, caching=False, ) print("Loading main guidance model...") guidance.llm = llama_model return llama_model def get_llama_model2(): global llama_model2 if llama_model2 is None and guidance_extraction_model_path is not None: print("Loading guidance model...") llama_model2 = guidance.llms.LlamaCpp( model = guidance_extraction_model_path, tokenizer = "openaccess-ai-collective/manticore-13b-chat-pyg", before_role = "<|", after_role = "|>", n_gpu_layers=300, n_threads=12, caching=False, ) print("Loading second guidance model...") return llama_model2 class DocumentBasedConversation: def __init__(self): """ Initializes an instance of the class. It sets up LLM, text splitter, vector store, prompt template, retriever, conversation chain, tools, and conversation agent if USE_AGENT is True. """ self.text_splitter = RecursiveCharacterTextSplitter( chunk_size=500, chunk_overlap=20, length_function=len) self.llama_model = get_llama_model() if llama_model2 is not None: self.llama_model2 = get_llama_model2() guidance.llm = self.llama_model self.vector_store_docs = Chroma(collection_name="docs_collection") self.vector_store_convs = Chroma(collection_name="convos_collection") tools = { "Search Documents": self.search_documents, "Search Conversations": self.search_conversations, } self.andromeda = AndromedaChain(config.andromeda_url) self.document_qa_agent = ChainOfThoughtsAgent(guidance, llama_model,llama_model2) def load_document(self, document_path, conversation_id=None): """ Load a document from a file and add its contents to the vector store. Args: document_path: A string representing the path to the document file. Returns: None. """ text_loader = TextLoader(document_path, encoding="utf8") documents = text_loader.load() documents = self.text_splitter.split_documents(documents) if conversation_id is not None: for doc in documents: doc.metadata["conversation_id"] = conversation_id self.vector_store_docs.add_documents(documents) def search_documents(self, search_input, conversation_id=None): """ Search for the given input in the vector store and return the top 10 most similar documents with their scores. This function is used as a helper function for the SearchLongTermMemory tool Args: search_input (str): The input to search for in the vector store. Returns: List[Tuple[str, float]]: A list of tuples containing the document text and their similarity score. """ logger.info(f"Searching for: {search_input} in LTM") docs = self.vector_store_docs.similarity_search_with_score( search_input, k=5, filter=filter ) return [{"document_content": doc[0].page_content, "similarity": doc[1]} for doc in docs] def search_conversations(self, search_input, conversation_id=None): """ Search for the given input in the vector store and return the top 10 most similar documents with their scores. This function is used as a helper function for the SearchLongTermMemory tool Args: search_input (str): The input to search for in the vector store. Returns: List[Tuple[str, float]]: A list of tuples containing the document text and their similarity score. """ if conversation_id is not None: filter = {"conversation_id": conversation_id} else: filter = {} logger.info(f"Searching for: {search_input} in LTM") docs = self.vector_store_convs.similarity_search_with_score( search_input, k=5, filter=filter ) return [{"document_content": doc[0].page_content, "similarity": doc[1]} for doc in docs] def predict(self, input, history): global dict_tools """ Predicts a response based on the given input. Args: input (str): The input string to generate a response for. Returns: str: The generated response string. Raises: OutputParserException: If the response from the conversation agent could not be parsed. """ context = self.search_documents(input) str_context = str(context) print(Fore.GREEN + Style.BRIGHT + "Printing vector search context..." + Style.RESET_ALL) print(Fore.GREEN + Style.BRIGHT + str_context + Style.RESET_ALL) final_answer = self.document_qa_agent.run(input, context, history) print(Fore.CYAN + Style.BRIGHT + "Printing full thought process..." + Style.RESET_ALL) print(Fore.CYAN + Style.BRIGHT + str(final_answer) + Style.RESET_ALL) if isinstance(final_answer, dict): final_answer = {'answer': str(final_answer), 'function': str(final_answer['fn'])} else: # Handle the case when final_answer is not a dictionary. final_answer = {'answer': str(final_answer)} # Check if 'Final Answer:' key exists in the dictionary if 'Final Answer:' in final_answer['answer']: # Find the index of 'Final Answer:' and extract everything after it answer_start_index = final_answer['answer'].index('Final Answer:') + len('Final Answer:') final_answer_text = final_answer['answer'][answer_start_index:] return final_answer_text.strip() else: return final_answer["answer"]
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/conversations/document_based_flow.py
app/conversations/document_based_flow.py
from langchain.document_loaders import TextLoader from memory.chroma_memory import Chroma from langchain.text_splitter import RecursiveCharacterTextSplitter from andromeda_chain import AndromedaChain from agents import ChainOfThoughtsFlowAgent from tools.base import ToolFactory from tools.document_memory import DocumentSearchTool from tools.conversation_memory import ConversationSearchTool from settings import load_config, logger config = load_config() class DocumentBasedConversationFlowAgent: def __init__(self): """ Initializes an instance of the class. It sets up LLM, text splitter, vector store, prompt template, retriever, conversation chain, tools, and conversation agent if USE_AGENT is True. """ self.text_splitter = RecursiveCharacterTextSplitter( chunk_size=500, chunk_overlap=20, length_function=len) self.vector_store_docs = Chroma(collection_name="docs_collection") self.vector_store_convs = Chroma(collection_name="convos_collection") self.tools = [DocumentSearchTool, ConversationSearchTool] self.andromeda = AndromedaChain(config.andromeda_url) self.active_agent_class = ChainOfThoughtsFlowAgent self.tool_context = { "vector_store_docs": self.vector_store_docs, "vector_store_convs": self.vector_store_convs, "k": 5, } def load_document(self, document_path, conversation_id=None): """ Load a document from a file and add its contents to the vector store. Args: document_path: A string representing the path to the document file. Returns: None. """ text_loader = TextLoader(document_path, encoding="utf8") documents = text_loader.load() documents = self.text_splitter.split_documents(documents) if conversation_id is not None: for doc in documents: doc.metadata["conversation_id"] = conversation_id self.vector_store_docs.add_documents(documents) def predict(self, input: str, conversation_id: str): """ Predicts a response based on the given input. Args: input (str): The input string to generate a response for. Returns: str: The generated response string. Raises: OutputParserException: If the response from the conversation agent could not be parsed. """ logger.info("Defined tools: %s", self.tools) loaded_tools = ToolFactory(self.tools).build_tools(conversation_id, self.tool_context) logger.info("Loaded tools: %s", loaded_tools) loaded_agent = self.active_agent_class(self.andromeda, loaded_tools) final_answer = loaded_agent.run(input) if isinstance(final_answer, dict): final_answer = {'answer': str(final_answer), 'function': str(final_answer['fn'])} else: # Handle the case when final_answer is not a dictionary. final_answer = {'answer': str(final_answer)} return final_answer["answer"]
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
ChuloAI/BrainChulo
https://github.com/ChuloAI/BrainChulo/blob/cacf3acff407ed131ef02196dfede54d8ff216da/app/prompt_templates/qa_agent.py
app/prompt_templates/qa_agent.py
QA_ETHICAL_AGENT=""" {{#system~}} Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request. ### Instruction: You are a helpful assistant. Answer the following questions as best you can. You have access to the following tools: Search: Useful for when you need to answer questions about current events. The input is the question to search relevant information. {{~/system}} {{#block hidden=True}} {{~! The assistant either starts the conversation or not, depending on if the user's query is offensive or not}} {{#user~}} Utilizing your extensive understanding of common moral and ethical principles, please evaluate the following user's query: {{question}}. Analyze the potential implications and outcomes associated with the query, considering various ethical frameworks such as consequentialism, deontology, and virtue ethics, among others. Also consider the social, cultural, and legal context of the query. Is the user's query ethical and/or moral? {{~/user}} {{#assistant~}} Observation: Let's see if the query is offensive. Decision:{{#select 'offensive' logprobs='logprobs'}}Yes{{or}}No{{/select}} {{~/assistant}} {{/block}} {{#if (equal offensive 'Yes')~}} Final Answer: I'm sorry, but I can't answer that. {{else}} {{~! The assistant then classifies the user intent to decide whether he needs to enter qa mode}} {{#user~}} Question: {{question}} Now classify the intent behind this question:{{question}} Identify if this question is phatic (serving a social function) or referential (seeking information). Provide reasoning for your classification based on elements like the content, structure, or context of the user's input. {{~/user}} {{#assistant~}} Thought: I need to evaluate the user's query and determine its intent - is {{question}} phatic or referential? Decision:{{#select 'query_type' logprobs='logprobs'}}Phatic{{or}}Referential{{/select}} {{~/assistant}} {{#if (equal query_type "Phatic")~}} Observation: The user's query is conversational. I need to answer as a helpful assistant while taking into account our chat history; Chat history: {{history}} Latest user message: {{question}} Thought: I need to stay in my role of a helpful assistant and make casual conversation. Final Answer: {{gen 'phatic answer' temperature=0.7 max_tokens=50}} {{else}} {{#user~}} Here are the relevant documents from our database:{{set 'documents' (search question)}} Using the concept of deixis, please evaluate if the answer to {{question}} is in :{{documents}}. Note that your response MUST contain either 'yes' or 'no'. {{~/user}} {{#assistant~}} Thought: I need to determine if the answer to {{question}} is in: {{documents}}. Decision:{{#select 'answerable' logprobs='logprobs'}}Yes{{or}}No{{/select}} {{#if (equal answerable "Yes")~}} Observation: I believe I can answer {{question}} based on the information contained in my analysis. Thought: Now that I know that I can answer, I should provide the information to the user. Final Answer: {{gen 'answer' temperature=0 max_tokens=100}} {{else}} Thought: I don't think I can answer the question based on the information contained in the returned documents. Final Answer: I'm sorry, but I don't have sufficient information to provide an answer to this question. {{/if}} {{~/assistant}} {{/if}} {{/if}} """ QA_AGENT= """ {{#system~}} Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request. ### Instruction: Answer the following questions as best you can. You have access to the following tools: Search: Useful for when you need to answer questions about current events. The input is the question to search relevant information. {{~/system}} {{#user~}} Question: {{question}} {{~/user}} {{#assistant~}} Thought: Let's first check our database. Action: Check Question Action Input: {{question}} {{~/assistant}} {{~! The assistant then classifies the user intent to decide whether he needs to enter qa mode}} {{#user~}} Question: {{question}} Now classify the intent behind this question:{{question}} Identify if this question is phatic (serving a social function) or referential (seeking information). Provide reasoning for your classification based on elements like the content, structure, or context of the user's input. {{~/user}} {{#assistant~}} Thought: I need to evaluate the user's query and determine its intent - is {{question}} phatic or referential? Decision:{{#select 'query_type' logprobs='logprobs'}}Phatic{{or}}Referential{{/select}} {{~/assistant}} {{#if (equal query_type "Phatic")~}} Observation: The user's query is conversational. I need to answer as a helpful assistant while taking into account our chat history; Chat history: {{history}} Latest user message: {{question}} Thought: I need to stay in my role of a helpful assistant and make casual conversation. Final Answer: {{gen 'phatic answer' temperature=0.7 max_tokens=50}} {{else}} {{#user~}} Here are the relevant documents from our database:{{set 'documents' (search question)}} Using the concept of deixis, please evaluate if the answer to {{question}} is in :{{documents}}. Note that your response MUST contain either 'yes' or 'no'. {{~/user}} {{#assistant~}} Thought: I need to determine if the answer to {{question}} is in: {{documents}}. Decision:{{#select 'answerable' logprobs='logprobs'}}Yes{{or}}No{{/select}} {{#if (equal answerable "Yes")~}} Observation: I believe I can answer {{question}} based on the information contained in my analysis. Thought: Now that I know that I can answer, I should provide the information to the user. Final Answer: {{gen 'answer' temperature=0 max_tokens=100}} {{else}} Thought: I don't think I can answer the question based on the information contained in the returned documents. Final Answer: I'm sorry, but I don't have sufficient information to provide an answer to this question. {{/if}} {{~/assistant}} {{/if}} """ ETHICS_PROMPT= ''' {{#system~}} Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request. ### Instruction: Answer the following questions as best you can. You have access to the following tools: Search: Useful for when you need to answer questions about current events. The input is the question to search relevant information. {{~/system}} {{~! The assistant either starts the conversation or not, depending on if the user's query is offensive or not}} {{#user~}} Utilizing your extensive understanding of common moral and ethical principles, please evaluate the following user's query {{question}}. Analyze the potential implications and outcomes associated with the query, considering various ethical frameworks such as consequentialism, deontology, and virtue ethics, among others. Also consider the social, cultural, and legal context of the query. Is the user's query ethical and/or moral? {{~/user}} {{#assistant~}} Observation: Let's see if the query is offensive. Decision:{{#select 'ethic' logprobs='logprobs'}}Yes{{or}}No{{/select}} {{~/assistant}} ''' ethics_prompt= ''' {{#system~}} Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request. ### Instruction: You are a helpful assistant. Answer the following questions as best you can. You have access to the following tools: Search: Useful for when you need to answer questions about current events. The input is the question to search relevant information. {{~/system}} {{~! The assistant either starts the conversation or not, depending on if the user's query is offensive or not}} {{#user~}} Utilizing your extensive understanding of common moral and ethical principles, please evaluate the following user's query: {{question}}. Analyze the potential implications and outcomes associated with the query, considering various ethical frameworks such as consequentialism, deontology, and virtue ethics, among others. Also consider the social, cultural, and legal context of the query. Is the user's query ethical and/or moral? {{~/user}} {{#assistant~}} Observation: Let's see if the query is inherently offensive. Decision:{{#select 'offensive' logprobs='logprobs'}}Yes{{or}}No{{/select}} {{~/assistant}} ''' conversation_prompt=""" {{#system~}} Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request. ### Instruction: You are a helpful assistant. Answer the following questions as best you can. You have access to the following tools: Search: Useful for when you need to answer questions about current events. The input is the question to search relevant information. {{~/system}} {{~! The assistant then classifies the user intent to decide whether he needs to enter qa mode}} {{#user~}} Question: {{question}} Now classify the intent behind this question:{{question}} Identify if this question is phatic (serving a social function) or referential (seeking information). Provide reasoning for your classification based on elements like the content, structure, or context of the user's input. {{~/user}} {{#assistant~}} Thought: I need to evaluate the user's query and determine its intent - is {{question}} phatic or referential? Decision:{{#select 'query_type' logprobs='logprobs'}}Phatic{{or}}Referential{{/select}} {{~/assistant}} """ phatic_prompt= ''' {{#system~}} Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request. ### Instruction: You are a helpful assistant. Answer the following questions as best you can. You have access to the following tools: Search: Useful for when you need to answer questions about current events. The input is the question to search relevant information. {{~/system}} Observation: The user's query is conversational. I need to answer him as a helpful assistant while taking into account our chat history; Chat history: {{history}} Latest user message: {{question}} Final Answer: {{gen 'phatic_answer' temperature=0 max_tokens=50}}''' referential_prompt = ''' {{#system~}} Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request. ### Instruction: You are a helpful assistant. Follow user queries as best you can. You have access to the following tools: Search: Useful for when you need to answer questions about factual information. The input is the question to search relevant information. {{~/system}} {{#user~}} Using the concept of deixis, please evaluate if the answer to {{question}} is in the documents retrieved from your database. Note that your response MUST contain either 'yes' or 'no'. {{~/user}} {{#assistant~}} Observation: I need to determine if the answer to {{question}} is in: {{#each (search question)}} {{this.document_content}} {{/each}} {{#select 'answerable' logprobs='logprobs'}}Yes{{or}}No{{/select}} {{~/assistant}} ''' answer_prompt = """ {{#system~}} Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request. ### Instruction: You are a helpful assistant. Answer the following questions as best you can. You have access to the following tools: Search: Useful for when you need to answer questions about current events. The input is the question to search relevant information. {{~/system}} {{#user~}} Please answer {{question}} based on the information contained in: {{#each (search question)}} {{this.document_content}} {{/each}} {{~/user}} {{#assistant~}} Thought: Now that I know that I can answer, I should provide the information to the user. Final Answer: {{gen 'final_answer' temperature=0.7 max_tokens=50}} {{~/assistant}}"""
python
MIT
cacf3acff407ed131ef02196dfede54d8ff216da
2026-01-05T07:13:18.041503Z
false
Snooz82/robotframework-datadriver
https://github.com/Snooz82/robotframework-datadriver/blob/d2d519666250324478bdccd7ccd65206a9110bf4/create_readme.py
create_readme.py
from inspect import getdoc from pathlib import Path from DataDriver import DataDriver with Path("Readme.rst").open("w", encoding="utf-8") as readme: doc_string = getdoc(DataDriver) readme.write(str(doc_string).replace("\\", "\\\\").replace("\\\\*", "\\*"))
python
Apache-2.0
d2d519666250324478bdccd7ccd65206a9110bf4
2026-01-05T07:13:20.041560Z
false
Snooz82/robotframework-datadriver
https://github.com/Snooz82/robotframework-datadriver/blob/d2d519666250324478bdccd7ccd65206a9110bf4/setup.py
setup.py
import re from pathlib import Path from setuptools import find_packages, setup with Path("Readme.rst").open(encoding="utf-8") as fh: long_description = fh.read() with (Path(__file__).resolve().parent / "src" / "DataDriver" / "DataDriver.py").open( encoding="utf-8" ) as f: VERSION = re.search('\n__version__ = "(.*)"', f.read()).group(1) setup( name="robotframework-datadriver", version=VERSION, author="René Rohner(Snooz82)", author_email="snooz@posteo.de", description="A library for Data-Driven Testing.", long_description=long_description, long_description_content_type="text/x-rst", url="https://github.com/Snooz82/robotframework-datadriver", package_dir={"": "src"}, packages=find_packages("src"), classifiers=[ "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Topic :: Software Development :: Testing", "Topic :: Software Development :: Testing :: Acceptance", "Framework :: Robot Framework", ], install_requires=["robotframework >= 4.0.2, < 8.0", "docutils", "Pygments"], extras_require={"xls": ["pandas", "xlrd >= 1.2.0", "openpyxl"]}, python_requires=">=3.8.0", )
python
Apache-2.0
d2d519666250324478bdccd7ccd65206a9110bf4
2026-01-05T07:13:20.041560Z
false
Snooz82/robotframework-datadriver
https://github.com/Snooz82/robotframework-datadriver/blob/d2d519666250324478bdccd7ccd65206a9110bf4/atest/run_atest.py
atest/run_atest.py
from robot.run import run # robot -d logs --listener RobotStackTracer --exclude performanceORfailingORfiltered --loglevel TRACE:INFO --extension robot . result = run( ".", outputdir="logs", listener="RobotStackTracer", exclude="performanceORfailingORfiltered", loglevel="TRACE:INFO", extension="robot", ) print(result)
python
Apache-2.0
d2d519666250324478bdccd7ccd65206a9110bf4
2026-01-05T07:13:20.041560Z
false
Snooz82/robotframework-datadriver
https://github.com/Snooz82/robotframework-datadriver/blob/d2d519666250324478bdccd7ccd65206a9110bf4/atest/TestCases/Tags/foo_reader.py
atest/TestCases/Tags/foo_reader.py
from DataDriver.AbstractReaderClass import AbstractReaderClass from DataDriver.ReaderConfig import TestCaseData class foo_reader(AbstractReaderClass): def get_data_from_source(self): return self._read_file_to_data_table() def _read_file_to_data_table(self): test_data = [] flipflop = True for i in range(6): args = {"${fooarg}": i} tags = ["included"] if flipflop: tags = ["filtered"] flipflop = False else: flipflop = True if i == 3: tags = [] test_data.append(TestCaseData(f"Test {i}", args, tags)) return test_data
python
Apache-2.0
d2d519666250324478bdccd7ccd65206a9110bf4
2026-01-05T07:13:20.041560Z
false