@@ -82,47 +82,47 @@ func handleModerationEndpoint(w http.ResponseWriter, r *http.Request) {
82
82
switch {
83
83
case strings .Contains (moderationReq .Input , "hate" ):
84
84
resCat = openai.ResultCategories {Hate : true }
85
- resCatScore = openai.ResultCategoryScores {Hate : true }
85
+ resCatScore = openai.ResultCategoryScores {Hate : 1 }
86
86
87
87
case strings .Contains (moderationReq .Input , "hate more" ):
88
88
resCat = openai.ResultCategories {HateThreatening : true }
89
- resCatScore = openai.ResultCategoryScores {HateThreatening : true }
89
+ resCatScore = openai.ResultCategoryScores {HateThreatening : 1 }
90
90
91
91
case strings .Contains (moderationReq .Input , "harass" ):
92
92
resCat = openai.ResultCategories {Harassment : true }
93
- resCatScore = openai.ResultCategoryScores {Harassment : true }
93
+ resCatScore = openai.ResultCategoryScores {Harassment : 1 }
94
94
95
95
case strings .Contains (moderationReq .Input , "harass hard" ):
96
96
resCat = openai.ResultCategories {Harassment : true }
97
- resCatScore = openai.ResultCategoryScores {HarassmentThreatening : true }
97
+ resCatScore = openai.ResultCategoryScores {HarassmentThreatening : 1 }
98
98
99
99
case strings .Contains (moderationReq .Input , "suicide" ):
100
100
resCat = openai.ResultCategories {SelfHarm : true }
101
- resCatScore = openai.ResultCategoryScores {SelfHarm : true }
101
+ resCatScore = openai.ResultCategoryScores {SelfHarm : 1 }
102
102
103
103
case strings .Contains (moderationReq .Input , "wanna suicide" ):
104
104
resCat = openai.ResultCategories {SelfHarmIntent : true }
105
- resCatScore = openai.ResultCategoryScores {SelfHarm : true }
105
+ resCatScore = openai.ResultCategoryScores {SelfHarm : 1 }
106
106
107
107
case strings .Contains (moderationReq .Input , "drink bleach" ):
108
108
resCat = openai.ResultCategories {SelfHarmInstructions : true }
109
- resCatScore = openai.ResultCategoryScores {SelfHarmInstructions : true }
109
+ resCatScore = openai.ResultCategoryScores {SelfHarmInstructions : 1 }
110
110
111
111
case strings .Contains (moderationReq .Input , "porn" ):
112
112
resCat = openai.ResultCategories {Sexual : true }
113
- resCatScore = openai.ResultCategoryScores {Sexual : true }
113
+ resCatScore = openai.ResultCategoryScores {Sexual : 1 }
114
114
115
115
case strings .Contains (moderationReq .Input , "child porn" ):
116
116
resCat = openai.ResultCategories {SexualMinors : true }
117
- resCatScore = openai.ResultCategoryScores {SexualMinors : true }
117
+ resCatScore = openai.ResultCategoryScores {SexualMinors : 1 }
118
118
119
119
case strings .Contains (moderationReq .Input , "kill" ):
120
120
resCat = openai.ResultCategories {Violence : true }
121
- resCatScore = openai.ResultCategoryScores {Violence : true }
121
+ resCatScore = openai.ResultCategoryScores {Violence : 1 }
122
122
123
123
case strings .Contains (moderationReq .Input , "corpse" ):
124
124
resCat = openai.ResultCategories {ViolenceGraphic : true }
125
- resCatScore = openai.ResultCategoryScores {ViolenceGraphic : true }
125
+ resCatScore = openai.ResultCategoryScores {ViolenceGraphic : 1 }
126
126
}
127
127
128
128
result := openai.Result {Categories : resCat , CategoryScores : resCatScore , Flagged : true }
0 commit comments