All Products
Search
Document Center

Intelligent Media Services:QuerySmarttagJob

Last Updated:Dec 16, 2024

Queries the information about a smart tagging job.

Debugging

You can run this interface directly in OpenAPI Explorer, saving you the trouble of calculating signatures. After running successfully, OpenAPI Explorer can automatically generate SDK code samples.

Authorization information

The following table shows the authorization information corresponding to the API. The authorization information can be used in the Action policy element to grant a RAM user or RAM role the permissions to call this API operation. Description:

  • Operation: the value that you can use in the Action element to specify the operation on a resource.
  • Access level: the access level of each operation. The levels are read, write, and list.
  • Resource type: the type of the resource on which you can authorize the RAM user or the RAM role to perform the operation. Take note of the following items:
    • The required resource types are displayed in bold characters.
    • If the permissions cannot be granted at the resource level, All Resources is used in the Resource type column of the operation.
  • Condition Key: the condition key that is defined by the cloud service.
  • Associated operation: other operations that the RAM user or the RAM role must have permissions to perform to complete the operation. To complete the operation, the RAM user or the RAM role must have the permissions to perform the associated operations.
OperationAccess levelResource typeCondition keyAssociated operation
ice:QuerySmarttagJob
*All Resources
*
    none
none

Request parameters

ParameterTypeRequiredDescriptionExample
JobIdstringYes

The ID of the smart tagging job that you want to query. You can obtain the job ID from the response parameters of the SubmitSmarttagJob operation.

88c6ca184c0e47098a5b665e2****
ParamsstringNo

The extra parameters that you want to query in the request. The value is a JSON string. Example: {"labelResultType":"auto"}. The value of labelResultType is of the STRING type. Valid values:

  • auto: machine tagging
  • hmi: tagging by human and machine
{"labelResultType":"auto"}

Response parameters

ParameterTypeDescriptionExample
object
JobStatusstring

The status of the job. Valid values:

  • Success: The job was successful.
  • Fail: The job failed.
  • Processing: The job is in progress.
  • Submitted: The job is submitted and waiting to be processed.
Success
RequestIdstring

The request ID.

******11-DB8D-4A9A-875B-275798******
UserDatastring

The content of callback messages that are sent to Simple Message Queue (SMQ) when the information of the smart tagging job changes. For more information about the parameters contained in the callback message, see the "Callback parameters" section of this topic.

{"userId":"123432412831"}
Resultsarray<object>

The analysis results of the smart tagging job. The value is an array.

Resultobject
Typestring

The type of the analysis result.

  • The type of the analysis result based on Smart tagging V1.0. Valid values:
  1. TextLabel: the text tag.
  2. VideoLabel: the video tag.
  3. ASR: the original result of automatic speech recognition (ASR). By default, this type of result is not returned.
  4. OCR: the original result of optical character recognition (OCR). By default, this type of result is not returned.
  5. NLP: the natural language processing (NLP)-based result. By default, this type of result is not returned.
  • The type of the analysis result based on Smart tagging V2.0. Valid values:
  1. CPVLabel
  2. Meta: the information about the video file, such as the title of the video. By default, this type of information is not returned.
  • The type of the analysis result based on Smart tagging V2.0-custom. Valid values:
  1. CPVLabel
  2. Meta: the information about the video file, such as the title of the video. By default, this type of information is not returned.
Meta
Datastring

The details of the analysis result. The value is a JSON string. For more information about the parameters of different result types, see the "Parameters of different result types" section of this topic.

{"title":"example-title-****"}

Callback parameters When the status of the smart tagging job changes, ApsaraVideo Media Processing (MPS) sends a message to the specified SMQ queue. For more information about how to specify an SMQ queue for receiving callbacks, see the UpdatePipeline topic. The callback message is a JSON string that contains the parameters described in the following table.

ParameterTypeDescription
TypeStringThe type of the job. The fixed value is smarttag, which indicates a smart tagging job.
JobIdStringThe unique ID of the job.
StateStringThe current status of the job. The value is the same as that of the JobStatus response parameter of the QuerySmarttagJob operation.
StateStringThe current status of the job. The value is the same as that of the JobStatus response parameter of the QuerySmarttagJob operation.
UserDataStringThe UserData information passed for the SubmitSmarttagJob operation.
UserDataStringThe UserData information passed for the SubmitSmarttagJob operation.

Parameters of different result types

Parameters of VideoLabel

ParameterTypeDescription
personsJSONArrayThe information of figures identified by the smart tagging job.
persons.nameStringThe name of the identified figure.
persons.categoryStringThe type of the identified figure. Valid values: celebrity, politician, sensitive, and unknown. A figure is identified as unknown based on the custom figure library. In this case, the ID of the custom figure is returned.
persons.ratiodoubleThe appearance rate of the figure. Valid values: 0 to 1.
persons.occurrencesJSONArrayThe details of the appearances of the figure.
persons.occurrences.scoredoubleThe score for the confidence level.
persons.occurrences.fromdoubleThe point in time when the figure appears. Unit: seconds.
persons.occurrences.todoubleThe point in time when the figure disappears. Unit: seconds.
persons.occurrences.positionJSONObjectThe face coordinates of the figure.
persons.occurrences.position.leftTopint[]The x and y coordinates of the upper-left corner.
persons.occurrences.position.rightBottomint[]The x and y coordinates of the lower-right corner.
persons.occurrences.timestampdoubleThe timestamp of the face coordinates. Unit: seconds.
persons.occurrences.sceneStringThe camera shot of the figure. Valid values: closeUp, medium-closeUp, medium, and medium-long.
tagsJSONArrayThe tags of the detected objects. For more information, see the following table.
tags.mainTagNameStringThe main tag.
tags.subTagNameStringThe subtag.
tags.ratiodoubleThe appearance rate of the tag. Valid values: 0 to 1.
tags.occurrencesJSONArrayThe details of the appearances of the tag.
tags.occurrences.scoredoubleThe score for the confidence level.
tags.occurrences.fromdoubleThe point in time when the tag appears. Unit: seconds.
tags.occurrences.todoubleThe point in time when the tag disappears. Unit: seconds.
classificationsJSONArrayThe category of the video.
classifications.scoredoubleThe score for the confidence level.
classifications.category1StringThe level-1 category, such as daily life activity, animation, and automobile.
classifications.category2StringThe level-2 category, such as health and home under the level-1 category daily life activity.

Examples of video tags

mainTagNamesubTagName
ProgramExamples: Dad Where Are We Going and Top Funny Comedian.
Figure roleExamples: doctor, nurse, and teacher.
ObjectExamples: piano, cup, table, scrambled eggs with tomatoes, car, and cosmetics.
TV channel logoExamples: CCTV-1, CCTV-2, YOUKU, and Dragon TV.
ActionExamples: dancing, kissing, hugging, meeting, singing, telephoning, horseback riding, and fighting.
LocationExamples: Tian'anmen Square, the Statue of Liberty, Leshan Giant Buddha, China, and America.
SceneExamples: bedroom, subway station, terraced field, beach, and desert.

Parameters of ImageLabel

ParameterTypeDescription
personsJSONArrayThe information of figures identified by the smart tagging job.
persons.nameStringThe name of the identified figure.
persons.categoryStringThe type of the identified figure. Valid values: celebrity, politician, and sensitive.
persons.scoredoubleThe score for the confidence level of the identified figure.
persons.positionJSONObjectThe face coordinates of the figure.
persons.position.leftTopint[]The x and y coordinates of the upper-left corner.
persons.position.rightBottomint[]The x and y coordinates of the lower-right corner.
persons.sceneStringThe camera shot of the figure. Valid values: closeUp, medium-closeUp, medium, and medium-long.
tagsJSONArrayThe tags of the detected objects. For more information, see the following table.
tags.mainTagNameStringThe main tag.
tags.subTagNameStringThe subtag.
tags.scoredoubleThe score for the confidence level.

Examples of image tags

mainTagNamesubTagName
Figure roleExamples: doctor, nurse, and teacher.
LocationExamples: Tian'anmen Square, the Statue of Liberty, Leshan Giant Buddha, China, and America.
ActionExample: talking.
TV channel logoExamples: CCTV-1, CCTV-2, YOUKU, and Dragon TV.
ActionExamples: dancing, kissing, hugging, meeting, singing, telephoning, horseback riding, and fighting.
ObjectExamples: piano, cup, table, scrambled eggs with tomatoes, car, and cosmetics.
SceneExamples: bedroom, subway station, terraced field, beach, and desert.

Parameters of TextLabel (from ASR and OCR)

ParameterTypeDescription
tagsJSONArrayThe text tags. For more information, see the following table.
tags.nameStringThe type of the tag.
tags.valueStringThe values of the tag. Multiple tag values are separated by commas (,).

Examples of text tags

namevalue
LocationExamples: Tian'anmen Square, the Statue of Liberty, Leshan Giant Buddha, China, and America.
OrganizationExamples: China Wildlife Conservation Association and China Media Group (CMG).
Brand nameExamples: Nike and Li-Ning.
KeywordExample: backbone force.

Parameters of CPVLabel

  • cates: the tagging category, including level-1 category, level-2 category, and level-3 category.
  • entities: the properties of the tagging category, including the knowledge graph information.
  • hotwords: the hotwords to which you pay attention.
  • freeTags: keywords.
ParameterTypeExampleDescription
typeStringhmiThe type of the result. Valid values: hmi and autp. A value of hmi indicates the results of tagging by human and machine. A value of autp indicates the results of machine tagging.
catesJSONArray-The information about the category of the tagging result.
cates.labelLevel1StringTourismThe level-1 tag.
cates.labelLevel2StringTourist landscapeThe level-2 tag.
cates.labelString""The name of the tag. An empty value may be returned by the algorithm.
cates.appearanceProbabilitydouble0.96The appearance rate of the tag.
cates.detailInfoJSONArray--
cates.detailInfo.scoredouble0.9The score for the confidence level.
cates.detailInfo.startTimedouble0.021The point in time when the object appears in the video.
cates.detailInfo.endTimedouble29.021The point in time when the object disappears in the video.
entitiesJSONArray--
entities.labelLevel1StringLocationThe level-1 tag.
entities.labelLevel2StringLandmarkThe level-2 tag.
entities.labelStringHuangguoshu WaterfallThe name of the tag.
entities.appearanceProbabilitydouble0.067The appearance rate of the tag.
entities.knowledgeInfoString{"name": "Huangguoshu Waterfall", "nameEn": "Huangguoshu Waterfall", "description": "One of the four largest waterfalls in Asia"}The knowledge graph information. The fields contained in the knowledge graph information are provided in Appendix, such as fields related to the intellectual property (IP) that is featured in films and television shows, music, figures, landmarks, and objects.
entities.detailInfoJSONArray--
entities.detailInfo.scoredouble0.33292606472969055The score for the confidence level.
entities.detailInfo.startTimedouble6.021The point in time when the object appears in the video.
entities.detailInfo.endTimedouble8.021The point in time when the object disappears in the video.
entities.detailInfo.trackDataJSONArray-The structured information about the tag of the object.
entities.detailInfo.trackData.scoredouble0.32The score for the confidence level.
entities.detailInfo.trackData.bboxinteger[]23, 43, 45, 67The coordinates of the object.
entities.detailInfo.trackData.timestampdouble7.9The timestamp of the coordinates. Unit: seconds.
hotwordsJSONArray--
hotwords.labelLevel1StringThe information about the hotword.The level-1 tag.
hotwords.labelLevel1StringHotwordThe level-2 tag.
hotwords.labelLevel2String""The level-2 tag.
hotwords.labelStringChina Meteorological AdministrationThe content of the hotword.
hotwords.appearanceProbabilitydouble0.96The appearance rate of the hotword.
hotwords.detailInfoJSONArray
hotwords.detailInfo.scoredouble1.0The score for the confidence level.
hotwords.detailInfo.startTimedouble0.021The point in time when the hotword appears in the video.
hotwords.detailInfo.endTimedouble29.021The point in time when the hotword disappears in the video.
freeTagsJSONArray
freeTags.labelLevel1StringKeywordThe level-1 tag.
freeTags.labelLevel2String""The level-2 tag.
freeTags.labelStringCentral Meteorological ObservatoryThe content of the keyword.
freeTags.appearanceProbabilitydouble0.96The appearance rate of the keyword.
freeTags.detailInfoJSONArray
freeTags.detailInfo.scoredouble0.9The score for the confidence level.
freeTags.detailInfo.startTimedouble0.021The point in time when the keyword appears in the video.
freeTags.detailInfo.endTimedouble29.021The point in time when the keyword disappears in the video.

Parameters of the ASR result

ParameterTypeDescription
detailsJSONArrayThe details of the result.
details.fromdoubleThe start timestamp of the recognition. Unit: seconds.
details.todoubleThe end timestamp of the recognition. Unit: seconds.
details.contentStringThe recognized text.

Parameters of the OCR result

ParameterTypeDescription
detailsJSONArrayThe details of the result.
details.timestampdoubleThe timestamp information. Unit: seconds.
details.infoJSONArrayThe details of the recognized text at the specified timestamp.
details.info.scoredoubleThe score for the confidence level.
details.info.positionJSONObjectThe coordinates of the text.
details.info.position.leftTopint[]The x and y coordinates of the upper-left corner.
details.info.position.rightBottomint[]The x and y coordinates of the lower-right corner.
details.info.contentStringThe recognized text.

Parameter of returned metadata

Note If you do not use tagging by human and machine and you specify the needMetaData parameter when you call the SubmitSmarttagJob operation, the original title of the video is returned in the result.

ParameterTypeDescription
titleStringThe title of the video.

Parameters of the extracted caption

ParameterTypeDescription
detailsJSONArrayThe details of the result.
details.allResultUrlStringThe URL of the file that contains all captions. The URL is valid for half a year after the job is complete.
details.chResultUrlStringThe URL of the file that contains only Chinese captions. The URL is valid for half a year after the job is complete.
details.engResultUrlStringThe URL of the file that contains only English captions. The URL is valid for half a year after the job is complete.

Note The content of the caption file is in the Serial number + Time range + Caption content format. Each line in the file contains a sentence.

Parameters of the NLP-based result

ParameterTypeDescription
transcriptionobjectThe speech-to-text result.
autoChaptersobjectThe chapter overview.
summarizationobjectThe large model summary.
meetingAssistanceobjectThe intelligent minutes.
translationobjectThe text translation result.

Parameters of transcription

ParameterTypeDescription
transcriptionobjectThe speech-to-text result.
transcription.paragraphslist[]A list of paragraphs that contain the speech-to-text result.
transcription.paragraphs[i].paragraphIdstringThe paragraph ID.
transcription.paragraphs[i].speakerIdstringThe speaker ID.
transcription.paragraphs[i].wordslist[]The words contained in the paragraph.
transcription.paragraphs[i].words[i].idintThe word ID. You do not need to pay attention to it.
transcription.paragraphs[i].words[i].sentenceIdintThe sentence ID. The words that have the same sentence ID can be assembled into a sentence.
transcription.paragraphs[i].words[i].startlongThe start time of the word. The value is a timestamp representing the number of milliseconds that have elapsed since the audio starts.
transcription.paragraphs[i].words[i].endlongThe end time of the word. The value is a timestamp representing the number of milliseconds that have elapsed since the audio starts.
transcription.paragraphs[i].words[i].textstringThe word.

Parameters of summarization

ParameterTypeDescription
summarizationobjectThe summary results. The results may be empty or of different summary types.
summarization.paragraphSummarystringThe summary of the full text.
summarization.conversationalSummarylist[]A list of summary results for a conversation.
summarization.conversationalSummary[i].speakerIdstringThe speaker ID.
summarization.conversationalSummary[i].speakerNamestringThe name of the speaker.
summarization.conversationalSummary[i].summarystringThe summary corresponding to the speaker.
summarization.questionsAnsweringSummarylist[]A list of summary results for an Q&A.
summarization.questionsAnsweringSummary[i].questionstringThe question.
summarization.questionsAnsweringSummary[i].sentenceIdsOfQuestionlist[]A list of IDs of the sentences that are generated based on the original speech corresponding to the question.
summarization.questionsAnsweringSummary[i].answerstringThe answer to the question.
summarization.questionsAnsweringSummary[i].sentenceIdsOfAnswerlist[]A list of IDs of the sentences that are generated based on the original speech corresponding to the answer.
summarization.mindMapSummarylist[object]The mind map of the summary results. The mind map may contain the summary of each topic and the relationship between topics.
summarization.mindMapSummary[i].titlestringThe title of the topic.
summarization.mindMapSummary[i].topiclist[object]An array that contains each topic and its subtopics.
summarization.mindMapSummary[i].topic[i].titlestringThe title of the topic.
summarization.mindMapSummary[i].topic[i].topiclist[object]An array that contains the subtopics of the topic. The array can be empty.

Parameters of translation

ParameterTypeDescription
translationobjectThe translation result.
translation.paragraphslist[]A list of paragraphs that contain the translation result, which corresponds to the ASR result.
translation.paragraphs.paragraphIdstringThe paragraph ID, which corresponds to the paragraph ID in the ASR result.
translation.paragraphs.sentenceslist[]A list of translated text sentences.
translation.paragraphs.sentences[i].sentenctIdlongThe sentence ID.
translation.paragraphs.sentences[i].startlongThe start time of the sentence. The value is a timestamp representing the number of milliseconds that have elapsed since the audio starts.
translation.paragraphs.sentences[i].endlongThe end time of the sentence. The value is a timestamp representing the number of milliseconds that have elapsed since the audio starts.
translation.paragraphs.sentences[i].textstringThe translated text, which corresponds to the ASR result.

Parameters of autoChapters

ParameterTypeDescription
autoChapterslist[]The chapter overview result that may contain the overview of zero, one, or multiple chapters.
autoChapters[i].idintThe serial number of the chapter.
autoChapters[i].startlongThe start time of the chapter. The value is a timestamp representing the number of milliseconds that have elapsed since the audio starts.
autoChapters[i].endlongThe end time of the chapter. The value is a timestamp representing the number of milliseconds that have elapsed since the audio starts.
autoChapters[i].headlinestringThe headline of the chapter.
autoChapters[i].summarystringThe chapter overview.

Parameters of meetingAssistance

ParameterTypeDescription
meetingAssistanceobjectThe result of the intelligent minutes, which may be empty or of different types.
meetingAssistance.keywordslist[]A list of extracted keywords.
meetingAssistance.keySentenceslist[]A list of extracted key sentences.
meetingAssistance.keySentences[i].idlongThe serial number of the key sentence.
meetingAssistance.keySentences[i].sentenceIdlongThe ID of the key sentence, which corresponds to the sentence ID in the original ASR result.
meetingAssistance.keySentences[i].startlongThe start time of the key sentence. The value is a timestamp representing the number of milliseconds that have elapsed since the audio starts.
meetingAssistance.keySentences[i].endlongThe end time of the key sentence. The value is a timestamp representing the number of milliseconds that have elapsed since the audio starts.
meetingAssistance.keySentences[i].textstringThe key sentence information.
meetingAssistance.actionslist[]A list of to-do items.
meetingAssistance.actions[i].idlongThe serial number of the to-do item.
meetingAssistance.actions[i].sentenceIdlongThe ID of the key sentence, which corresponds to the sentence ID in the original ASR result.
meetingAssistance.actions[i].startlongThe start time of the key sentence. The value is a timestamp representing the number of milliseconds that have elapsed since the audio starts.
meetingAssistance.actions[i].endlongThe end time of the key sentence. The value is a timestamp representing the number of milliseconds that have elapsed since the audio starts.
meetingAssistance.actions[i].textstringThe content of the to-do item.
meetingAssistance.classificationsobjectThe scenario type. Only three types of scenarios are supported.
meetingAssistance.classifications.interviewfloatThe score for the confidence level of the interview scenario.
meetingAssistance.classifications.lecturefloatThe score for the confidence level of the presentation scenario.
meetingAssistance.classifications.meetingfloatThe score for the confidence level of the meeting scenario.

Examples

Sample success responses

JSONformat

{
  "JobStatus": "Success",
  "RequestId": "******11-DB8D-4A9A-875B-275798******",
  "UserData": "{\"userId\":\"123432412831\"}",
  "Results": {
    "Result": [
      {
        "Type": "Meta",
        "Data": "{\"title\":\"example-title-****\"}\t\n"
      }
    ]
  }
}

Error codes

For a list of error codes, visit the Service error codes.

Change history

Change timeSummary of changesOperation
2022-08-25Add OperationView Change Details