Automate moderation by getting toxicity of messages.
HTML form for interacting with the function.
Returns toxicity score of the provided text, as determined by Google's Perspective API
| Name | Description | Location | Type | Sample Value |
|---|---|---|---|---|
| Content-Type | Content type of the request | Header | application/json |
N/A |
| text | Text to analyze | Body | String | Goodbye, have a great day! |
Response
Sample 200 Response:
{
"ok": true,
"score": 0.1
}Sample 400 Response:
{
"ok": false,
"error": "Missing required field: text"
}Sample 500 Response:
{
"ok": false,
"error": "Error fetching from perspective API"
}| Setting | Value |
|---|---|
| Runtime | Node (18.0) |
| Entrypoint | src/main.js |
| Build Commands | npm install |
| Permissions | any |
| Timeout (Seconds) | 15 |
Google Perspective API key. It authenticates your function, allowing it to interact with the API.
| Question | Answer |
|---|---|
| Required | Yes |
| Sample Value | AIzaS...fk-fuM |
| Documentation | Setup Perspective API |