@Generated(value="com.amazonaws:aws-java-sdk-code-generator") public class DetectToxicContentResult extends AmazonWebServiceResult<ResponseMetadata> implements Serializable, Cloneable
| Constructor and Description |
|---|
DetectToxicContentResult() |
| Modifier and Type | Method and Description |
|---|---|
DetectToxicContentResult |
clone() |
boolean |
equals(Object obj) |
List<ToxicLabels> |
getResultList()
Results of the content moderation analysis.
|
int |
hashCode() |
void |
setResultList(Collection<ToxicLabels> resultList)
Results of the content moderation analysis.
|
String |
toString()
Returns a string representation of this object.
|
DetectToxicContentResult |
withResultList(Collection<ToxicLabels> resultList)
Results of the content moderation analysis.
|
DetectToxicContentResult |
withResultList(ToxicLabels... resultList)
Results of the content moderation analysis.
|
getSdkHttpMetadata, getSdkResponseMetadata, setSdkHttpMetadata, setSdkResponseMetadatapublic List<ToxicLabels> getResultList()
Results of the content moderation analysis. Each entry in the results list contains a list of toxic content types identified in the text, along with a confidence score for each content type. The results list also includes a toxicity score for each entry in the results list.
public void setResultList(Collection<ToxicLabels> resultList)
Results of the content moderation analysis. Each entry in the results list contains a list of toxic content types identified in the text, along with a confidence score for each content type. The results list also includes a toxicity score for each entry in the results list.
resultList - Results of the content moderation analysis. Each entry in the results list contains a list of toxic
content types identified in the text, along with a confidence score for each content type. The results
list also includes a toxicity score for each entry in the results list.public DetectToxicContentResult withResultList(ToxicLabels... resultList)
Results of the content moderation analysis. Each entry in the results list contains a list of toxic content types identified in the text, along with a confidence score for each content type. The results list also includes a toxicity score for each entry in the results list.
NOTE: This method appends the values to the existing list (if any). Use
setResultList(java.util.Collection) or withResultList(java.util.Collection) if you want to
override the existing values.
resultList - Results of the content moderation analysis. Each entry in the results list contains a list of toxic
content types identified in the text, along with a confidence score for each content type. The results
list also includes a toxicity score for each entry in the results list.public DetectToxicContentResult withResultList(Collection<ToxicLabels> resultList)
Results of the content moderation analysis. Each entry in the results list contains a list of toxic content types identified in the text, along with a confidence score for each content type. The results list also includes a toxicity score for each entry in the results list.
resultList - Results of the content moderation analysis. Each entry in the results list contains a list of toxic
content types identified in the text, along with a confidence score for each content type. The results
list also includes a toxicity score for each entry in the results list.public String toString()
toString in class ObjectObject.toString()public DetectToxicContentResult clone()