Update config for requirement_check

#12
requirement_check/alora/gpt-oss-20b/io.yaml CHANGED
@@ -1,8 +1,9 @@
1
- model: requirement
 
2
  # JSON schema of the model's output
3
  response_format: |
4
  {
5
- "type": "boolean",
6
  }
7
  transformations:
8
  - type: likelihood
@@ -10,9 +11,14 @@ transformations:
10
  true: 1.0
11
  false: 0.0
12
  input_path: []
 
 
 
 
13
  instruction: |
14
  Check if this requirement was satisfied:
15
  {requirement}
16
  parameters:
 
17
  max_completion_tokens: 5
18
  sentence_boundaries: ~
 
1
+ # Model name; overrides the model from the original request if present.
2
+ model: ~
3
  # JSON schema of the model's output
4
  response_format: |
5
  {
6
+ "type": "boolean"
7
  }
8
  transformations:
9
  - type: likelihood
 
11
  true: 1.0
12
  false: 0.0
13
  input_path: []
14
+ # Convert scalar value to a record for consistency with other intrinsics
15
+ - type: nest
16
+ input_path: []
17
+ field_name: "requirement_likelihood"
18
  instruction: |
19
  Check if this requirement was satisfied:
20
  {requirement}
21
  parameters:
22
+ # "false" can be up to 5 tokens at high temperatures
23
  max_completion_tokens: 5
24
  sentence_boundaries: ~
requirement_check/alora/granite-3.3-2b-instruct/io.yaml CHANGED
@@ -1,8 +1,9 @@
1
- model: requirement
 
2
  # JSON schema of the model's output
3
  response_format: |
4
  {
5
- "type": "boolean",
6
  }
7
  transformations:
8
  - type: likelihood
@@ -10,9 +11,14 @@ transformations:
10
  true: 1.0
11
  false: 0.0
12
  input_path: []
 
 
 
 
13
  instruction: |
14
  Check if this requirement was satisfied:
15
  {requirement}
16
  parameters:
 
17
  max_completion_tokens: 5
18
  sentence_boundaries: ~
 
1
+ # Model name; overrides the model from the original request if present.
2
+ model: ~
3
  # JSON schema of the model's output
4
  response_format: |
5
  {
6
+ "type": "boolean"
7
  }
8
  transformations:
9
  - type: likelihood
 
11
  true: 1.0
12
  false: 0.0
13
  input_path: []
14
+ # Convert scalar value to a record for consistency with other intrinsics
15
+ - type: nest
16
+ input_path: []
17
+ field_name: "requirement_likelihood"
18
  instruction: |
19
  Check if this requirement was satisfied:
20
  {requirement}
21
  parameters:
22
+ # "false" can be up to 5 tokens at high temperatures
23
  max_completion_tokens: 5
24
  sentence_boundaries: ~
requirement_check/alora/granite-3.3-8b-instruct/io.yaml CHANGED
@@ -1,8 +1,9 @@
1
- model: requirement
 
2
  # JSON schema of the model's output
3
  response_format: |
4
  {
5
- "type": "boolean",
6
  }
7
  transformations:
8
  - type: likelihood
@@ -10,9 +11,14 @@ transformations:
10
  true: 1.0
11
  false: 0.0
12
  input_path: []
 
 
 
 
13
  instruction: |
14
  Check if this requirement was satisfied:
15
  {requirement}
16
  parameters:
 
17
  max_completion_tokens: 5
18
  sentence_boundaries: ~
 
1
+ # Model name; overrides the model from the original request if present.
2
+ model: ~
3
  # JSON schema of the model's output
4
  response_format: |
5
  {
6
+ "type": "boolean"
7
  }
8
  transformations:
9
  - type: likelihood
 
11
  true: 1.0
12
  false: 0.0
13
  input_path: []
14
+ # Convert scalar value to a record for consistency with other intrinsics
15
+ - type: nest
16
+ input_path: []
17
+ field_name: "requirement_likelihood"
18
  instruction: |
19
  Check if this requirement was satisfied:
20
  {requirement}
21
  parameters:
22
+ # "false" can be up to 5 tokens at high temperatures
23
  max_completion_tokens: 5
24
  sentence_boundaries: ~
requirement_check/lora/gpt-oss-20b/io.yaml CHANGED
@@ -1,8 +1,9 @@
1
- model: requirement
 
2
  # JSON schema of the model's output
3
  response_format: |
4
  {
5
- "type": "boolean",
6
  }
7
  transformations:
8
  - type: likelihood
@@ -10,9 +11,14 @@ transformations:
10
  true: 1.0
11
  false: 0.0
12
  input_path: []
 
 
 
 
13
  instruction: |
14
  Check if this requirement was satisfied:
15
  {requirement}
16
  parameters:
 
17
  max_completion_tokens: 5
18
  sentence_boundaries: ~
 
1
+ # Model name; overrides the model from the original request if present.
2
+ model: ~
3
  # JSON schema of the model's output
4
  response_format: |
5
  {
6
+ "type": "boolean"
7
  }
8
  transformations:
9
  - type: likelihood
 
11
  true: 1.0
12
  false: 0.0
13
  input_path: []
14
+ # Convert scalar value to a record for consistency with other intrinsics
15
+ - type: nest
16
+ input_path: []
17
+ field_name: "requirement_likelihood"
18
  instruction: |
19
  Check if this requirement was satisfied:
20
  {requirement}
21
  parameters:
22
+ # "false" can be up to 5 tokens at high temperatures
23
  max_completion_tokens: 5
24
  sentence_boundaries: ~
requirement_check/lora/granite-3.3-2b-instruct/io.yaml CHANGED
@@ -1,8 +1,9 @@
1
- model: requirement
 
2
  # JSON schema of the model's output
3
  response_format: |
4
  {
5
- "type": "boolean",
6
  }
7
  transformations:
8
  - type: likelihood
@@ -10,9 +11,14 @@ transformations:
10
  true: 1.0
11
  false: 0.0
12
  input_path: []
 
 
 
 
13
  instruction: |
14
  Check if this requirement was satisfied:
15
  {requirement}
16
  parameters:
 
17
  max_completion_tokens: 5
18
  sentence_boundaries: ~
 
1
+ # Model name; overrides the model from the original request if present.
2
+ model: ~
3
  # JSON schema of the model's output
4
  response_format: |
5
  {
6
+ "type": "boolean"
7
  }
8
  transformations:
9
  - type: likelihood
 
11
  true: 1.0
12
  false: 0.0
13
  input_path: []
14
+ # Convert scalar value to a record for consistency with other intrinsics
15
+ - type: nest
16
+ input_path: []
17
+ field_name: "requirement_likelihood"
18
  instruction: |
19
  Check if this requirement was satisfied:
20
  {requirement}
21
  parameters:
22
+ # "false" can be up to 5 tokens at high temperatures
23
  max_completion_tokens: 5
24
  sentence_boundaries: ~
requirement_check/lora/granite-3.3-8b-instruct/io.yaml CHANGED
@@ -1,8 +1,9 @@
1
- model: requirement
 
2
  # JSON schema of the model's output
3
  response_format: |
4
  {
5
- "type": "boolean",
6
  }
7
  transformations:
8
  - type: likelihood
@@ -10,9 +11,14 @@ transformations:
10
  true: 1.0
11
  false: 0.0
12
  input_path: []
 
 
 
 
13
  instruction: |
14
  Check if this requirement was satisfied:
15
  {requirement}
16
  parameters:
 
17
  max_completion_tokens: 5
18
  sentence_boundaries: ~
 
1
+ # Model name; overrides the model from the original request if present.
2
+ model: ~
3
  # JSON schema of the model's output
4
  response_format: |
5
  {
6
+ "type": "boolean"
7
  }
8
  transformations:
9
  - type: likelihood
 
11
  true: 1.0
12
  false: 0.0
13
  input_path: []
14
+ # Convert scalar value to a record for consistency with other intrinsics
15
+ - type: nest
16
+ input_path: []
17
+ field_name: "requirement_likelihood"
18
  instruction: |
19
  Check if this requirement was satisfied:
20
  {requirement}
21
  parameters:
22
+ # "false" can be up to 5 tokens at high temperatures
23
  max_completion_tokens: 5
24
  sentence_boundaries: ~