`."}],"requestBody":{"content":{"application/json":{"schema":{"$ref":"#/components/schemas/EmbeddingsRequest"}}},"required":true},"x-nvai-meta":{"name":"Create Text Embedding","description":"Generates an embedding vector from the provided text\\nusing a specified model. The embedding can be returned\\nin either float array or base64-encoded format.\\n","path":"create","templates":[{"title":"Synchronous requests","requestEjs":{"curl":"$7e","python":"$7f","node":"$80"},"response":"$81"}]},"responses":{"200":{"description":"Invocation is fulfilled.","content":{"application/json":{"schema":{"$ref":"#/components/schemas/EmbeddingsResponse"}}}},"402":{"description":"Payment Required.","content":{"application/json":{"schema":{"properties":{"detail":{"type":"string","maxLength":256,"format":"^[a-zA-Z-]{1,64}$","description":"Contains specific information related to the error and why it occurred.","example":"You have reached your limit of credits."}},"type":"object","title":"PaymentRequiredError"}}}},"422":{"description":"Validation failed, provided entity could not be processed.","content":{"application/json":{"schema":{"$ref":"#/components/schemas/Errors"},"example":{"type":"urn:nvcf-worker-service:problem-details:unprocessable-entity","title":"Unprocessable Entity","status":422,"detail":"string","instance":"/v2/nvcf/pexec/functions/bc205f8e-1740-40df-8d32-c4321763498a","requestId":"3fa85f64-5717-4562-b3fc-2c963f66afa6"}}}},"500":{"description":"The invocation ended with an error.","content":{"application/json":{"schema":{"$ref":"#/components/schemas/Errors"},"example":{"type":"urn:nvcf-worker-service:problem-details:internal-server-error","title":"Internal Server Error","status":500,"detail":"string","instance":"/v2/nvcf/pexec/functions/bc205f8e-1740-40df-8d32-c4321763498a","requestId":"3fa85f64-5717-4562-b3fc-2c963f66afa6"}}}}}}}},"security":[{"Token":[]}],"components":{"securitySchemes":{"Token":{"type":"http","scheme":"bearer"}},"schemas":{"EmbeddingsRequest":{"additionalProperties":false,"properties":{"input":{"oneOf":[{"type":"string","title":"string","example":"This is a test.","pattern":"^.*$"},{"type":"array","items":{"type":"string","pattern":"^.*$"},"maxItems":64}],"title":"Input","description":"The list of images or texts that you want to generate embeddings for. Images should be in form of `data:image/{format};base64,{base64encodedimage}`. If the size of an image is more than 200KB, it needs to be uploaded to a presigned S3 bucket using NVCF Asset APIs. Once uploaded you can refer to it using the following format: ``. Accepted formats are `jpg`, `png` and `jpeg`.","examples":[["The quick brown fox jumped over the lazy dog","data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAEAAAABCAIAAACQd1PeAAAAEElEQVR4nGK6HcwNCAAA//8DTgE8HuxwEQAAAABJRU5ErkJggg=="],["The quick brown fox jumped over the lazy dog","data:image/png;asset_id,87b132b0-08f9-43ea-8fab-f107350a5d00"]]},"encoding_format":{"type":"string","enum":["float","base64"],"title":"Encoding Format","description":"The format to return the embeddings in. Can be either `float` or `base64`.","default":"float"},"model":{"type":"string","enum":["nvidia/nvclip"],"title":"Model","description":"ID of the embedding model.","examples":["nvidia/nvclip"]},"dimensions":{"description":"Not implemented, but provided for API compliance. This field is ignored.\\n","type":"integer","minimum":1,"maximum":1,"format":"int32"},"user":{"type":"string","description":"Not implemented, but provided for API compliance. This field is ignored.","pattern":"^.*$","maxLength":204800,"example":"user123"}},"type":"object","required":["input","model"],"title":"EmbeddingsRequest"},"EmbeddingsResponse":{"type":"object","properties":{"object":{"description":"The object type, which is always `list`.","enum":["list"],"title":"Object","type":"string"},"data":{"items":{"$ref":"#/components/schemas/Embedding"},"type":"array","title":"Data","description":"The list of embeddings generated by the model.","maxItems":128,"examples":[[{"embedding":[0.1,0.1,0.1],"index":0,"object":"embedding"},{"embedding":[0.1,0.1,0.1],"index":1,"object":"embedding"}]]},"model":{"type":"string","description":"Model used to generate embeddings.","example":"nvidia/nvclip","title":"Model","const":"nvidia/nvclip"},"usage":{"allOf":[{"$ref":"#/components/schemas/Usage"}],"description":"Usage statistics for the embeddings request.","examples":[{"num_images":1,"prompt_tokens":80,"total_tokens":80}]}},"required":["object","data","usage","model"],"title":"EmbeddingsResponse"},"Embedding":{"type":"object","description":"Represents an embedding vector returned by embedding endpoint.","properties":{"object":{"description":"The object type, which is always `embedding`.","enum":["embedding"],"title":"Object","type":"string"},"index":{"type":"integer","format":"int32","title":"Index","description":"The index of the embedding in the list of embeddings.","minimum":0,"maximum":127},"embedding":{"oneOf":[{"type":"array","description":"The embedding vector as a list of floats, The length of the vector depends on the model.\\n","items":[{"type":"number"}],"maxItems":1024},{"type":"string","description":"The embedding vector as base64 string. The length of the vector depends on the model.\\n","format":"^(?:[A-Za-z0-9+/]{4})*(?:[A-Za-z0-9+/]{4}|[A-Za-z0-9+/]{3}=|[A-Za-z0-9+/]{2}==)$"}]},"title":"Embedding"},"required":["index","object","embedding"],"title":"Embeddings"},"Usage":{"properties":{"num_images":{"type":"integer","format":"int32","title":"Num Images","minimum":0,"maximum":64,"description":"Number of images passed."},"prompt_tokens":{"type":"integer","format":"int32","title":"Prompt Tokens","description":"Number of tokens in the prompt.","minimum":77,"maximum":4928},"total_tokens":{"type":"integer","format":"int32","title":"Total Tokens","description":"Total number of tokens used in the request.","minimum":77,"maximum":4928}},"type":"object","required":["num_images","prompt_tokens","total_tokens"],"title":"Usage"},"Errors":{"properties":{"type":{"type":"string","format":"^.{1, 128}$","maxLength":128,"description":"Error type"},"title":{"type":"string","format":"^.{1, 128}$","maxLength":128,"description":"Error title"},"status":{"type":"integer","format":"int32","minimum":100,"maximum":999,"description":"Error status code"},"detail":{"type":"string","format":"^.{1, 1024}$","maxLength":1024,"description":"Detailed information about the error"},"instance":{"type":"string","format":"^.{1, 256}$","maxLength":256,"description":"Function instance used to invoke the request"},"requestId":{"type":"string","format":"uuid","maxLength":36,"description":"UUID of the request"}},"type":"object","required":["type","title","status","detail","instance","requestId"],"title":"InvokeError"}}}},"namespace":"qc69jvmznzxy","updatedDate":"2025-04-11T18:04:12.307Z","nvcfFunctionId":"3072eebf-b0f0-4318-a5b8-5a45cd035b95","createdDate":"2024-06-13T20:35:52.529Z","attributes":{"apiDocsUrl":"https://docs.api.nvidia.com/nim/reference/nvidia-nvclip","termsOfUse":"GOVERNING TERMS: GOVERNING TERMS: The trial service is governed by the NVIDIA API Trial Service Terms of Use; and the use of this model is governed by the NVIDIA AI Foundation Models Community License.\\n","showUnavailableBanner":false,"cta":{"type":"Apply to Self-Host","url":"https://www.nvidia.com/en-us/ai/nim-notifyme/"},"deploy":[{"label":"Linux with Docker","filename":"linux.md","contents":"$82"},{"label":"Windows on RTX AI PCs (Beta)","filename":"wsl2.md","contents":"$83"}]},"artifactName":"nvclip"},"config":{"name":"nvclip","type":"model"}},{"endpoint":{"artifact":{"artifactType":"ENDPOINT","name":"paddleocr","displayName":"paddleocr","publisher":"baidu","shortDescription":"Model for table extraction that receives an image as input, runs OCR on the image, and returns the text within the image and its bounding boxes.","logo":"https://assets.ngc.nvidia.com/products/api-catalog/images/paddleocr.jpg","labels":["Optical Character Detection","Optical Character Recognition","Table Extraction","data ingestion","extraction","nemo retriever","run on rtx"],"attributes":[{"key":"AVAILABLE","value":"true"},{"key":"PREVIEW","value":"false"}],"canGuestDownload":true,"createdDate":"2024-11-22T17:15:01.232Z","description":"$84","isPublic":true,"isReadOnly":true,"orgName":"qc69jvmznzxy","updatedDate":"2025-03-25T00:02:36.704Z"},"requestStatus":{"statusCode":"SUCCESS","requestId":"55bec8f6-a214-4b32-8255-f32cd42c778b"}},"spec":{"openAPISpec":{"openapi":"3.1.0","info":{"title":"PaddleOCR","description":"PaddleOCR","version":"0.2.1-rc0"},"paths":{"/v1/infer":{"post":{"summary":"Post Infer","operationId":"post_infer_v1_infer_post","requestBody":{"content":{"application/json":{"schema":{"$ref":"#/components/schemas/PaddleOCRRequest"}}},"required":true},"responses":{"200":{"description":"Successful Response","content":{"application/json":{"schema":{"$ref":"#/components/schemas/PaddleOCRResponse"}}}},"422":{"description":"Validation Error","content":{"application/json":{"schema":{"$ref":"#/components/schemas/HTTPValidationError"}}}}},"x-nvai-meta":{"name":"Inference model on image","returns":"Returns a object list.\\n","path":"create","examples":[{"name":"Image from example","input":{"image":"https://assets.ngc.nvidia.com/products/api-catalog/paddle/paddleocr1.png"},"requestJson":"$85","responseJson":"$86"},{"name":"Image from example","input":{"image":"https://assets.ngc.nvidia.com/products/api-catalog/paddle/paddleocr2.png"},"requestJson":"$87","responseJson":"$88"}],"templates":[{"title":"Default","requestEjs":{"python":"import requests, base64\\n\\ninvoke_url = \\"https://ai.api.nvidia.com/v1/cv/baidu/paddleocr\\"\\n\\nwith open(\\"paddleocr1.png\\", \\"rb\\") as f:\\n image_b64 = base64.b64encode(f.read()).decode()\\n\\nassert len(image_b64) < 180_000, \\\\\\n \\"To upload larger images, use the assets API (see docs)\\"\\n\\nheaders = {\\n \\"Authorization\\": \\"Bearer $NVIDIA_API_KEY\\",\\n \\"Accept\\": \\"application/json\\"\\n}\\n\\npayload = {\\n \\"input\\": [\\n {\\n \\"type\\": \\"image_url\\",\\n \\"url\\": f\\"data:image/png;base64,{image_b64}\\"\\n }\\n ]\\n}\\n\\nresponse = requests.post(invoke_url, headers=headers, json=payload)\\n\\nprint(response.json())\\n","node.js":"import axios from \'axios\';\\nimport { readFile } from \'node:fs/promises\';\\n\\nconst invokeUrl = \\"https://ai.api.nvidia.com/v1/cv/baidu/paddleocr\\";\\n\\nconst headers = {\\n \\"Authorization\\": \\"Bearer $NVIDIA_API_KEY\\",\\n \\"Accept\\": \\"application/json\\"\\n};\\n\\nreadFile(\\"paddleocr1.png\\")\\n .then(data => {\\n const imageB64 = Buffer.from(data).toString(\'base64\');\\n if (imageB64.length > 180_000) {\\n throw new Error(\\"To upload larger images, use the assets API (see docs)\\");\\n }\\n\\n const payload = {\\n \\"input\\": [\\n {\\n \\"type\\": \\"image_url\\",\\n \\"url\\": `data:image/png;base64,${imageB64}`\\n }\\n ]\\n };\\n\\n return axios.post(invokeUrl, payload, { headers: headers, responseType: \'json\' });\\n })\\n .then(response => {\\n console.log(JSON.stringify(response.data));\\n })\\n .catch(error => {\\n console.error(error);\\n });\\n","curl":"image_b64=$( base64 -i paddleocr1.png )\\n\\naccept_header=\'Accept: application/json\'\\n\\n# Construct the JSON payload\\necho \'{\\n \\"input\\": [\\n {\\n \\"type\\": \\"image_url\\",\\n \\"url\\": \\"data:image/png;base64,\'\\"${image_b64}\\"\'\\"\\n }\\n ]\\n}\' > payload.json\\n\\ncurl https://ai.api.nvidia.com/v1/cv/baidu/paddleocr \\\\\\n -H \\"Authorization: Bearer $NVIDIA_API_KEY\\" \\\\\\n -H \\"Content-Type: application/json\\" \\\\\\n -H \\"$accept_header\\" \\\\\\n -d @payload.json\\n"},"response":"$89"}]}}},"/v1/health/live":{"get":{"summary":"Health Live","description":"Return service liveness status","operationId":"health_live_v1_health_live_get","responses":{"200":{"description":"Successful Response","content":{"application/json":{"schema":{"$ref":"#/components/schemas/HealthLiveResponse"}}}}}}},"/v1/health/ready":{"get":{"summary":"Health Ready","description":"Return service readiness status","operationId":"health_ready_v1_health_ready_get","responses":{"200":{"description":"Successful Response","content":{"application/json":{"schema":{"$ref":"#/components/schemas/HealthReadyResponse"}}}}}}},"/v1/metrics":{"get":{"summary":"Metrics","description":"Handler for metrics endpoint.","operationId":"metrics_v1_metrics_get","responses":{"200":{"description":"Successful Response","content":{"application/json":{"schema":{"type":"string","title":"Response Metrics V1 Metrics Get"}}}}}}},"/v1/license":{"get":{"summary":"License","description":"Handler for license endpoint.","operationId":"license_v1_license_get","responses":{"200":{"description":"Successful Response","content":{"application/json":{"schema":{"$ref":"#/components/schemas/LicenseEndpointModel"}}}}}}},"/v1/metadata":{"get":{"summary":"Metadata","description":"Handler for metadata endpoint.","operationId":"metadata_v1_metadata_get","responses":{"200":{"description":"Successful Response","content":{"application/json":{"schema":{"$ref":"#/components/schemas/MetadataEndpointModel"}}}}}}},"/v1/manifest":{"get":{"summary":"Manifest","description":"Handler for the manifest endpoint.","operationId":"manifest_v1_manifest_get","responses":{"200":{"description":"Successful Response","content":{"application/json":{"schema":{"$ref":"#/components/schemas/ManifestEndpointModel"}}}}}}}},"components":{"schemas":{"HTTPValidationError":{"properties":{"detail":{"items":{"$ref":"#/components/schemas/ValidationError"},"type":"array","title":"Detail"}},"type":"object","title":"HTTPValidationError"},"HealthLiveResponse":{"properties":{"live":{"type":"boolean","title":"Live"}},"type":"object","required":["live"],"title":"HealthLiveResponse"},"HealthReadyResponse":{"properties":{"ready":{"type":"boolean","title":"Ready"}},"type":"object","required":["ready"],"title":"HealthReadyResponse"},"Image":{"properties":{"type":{"type":"string","enum":["image_url"],"const":"image_url","title":"Type","default":"image_url"},"image_url":{"$ref":"#/components/schemas/ImageUrl"}},"type":"object","required":["image_url"],"title":"Image"},"ImageUrl":{"properties":{"url":{"type":"string","title":"Url","description":"The URL of the image to be processed."}},"type":"object","required":["url"],"title":"ImageUrl"},"LicenseEndpointModel":{"properties":{"name":{"type":"string","title":"Name","description":"The name of the license for the NIM container."},"path":{"type":"string","title":"Path","description":"The filepath within the container containing the license content."},"sha":{"type":"string","title":"Sha","description":"A SHA1 hash of the license contents."},"size":{"type":"integer","title":"Size","description":"The number of characters in the license content."},"url":{"type":"string","title":"Url","description":"The url where this license is hosted externally."},"type":{"type":"string","enum":["file"],"const":"file","title":"Type","description":"The format of the license content."},"content":{"type":"string","title":"Content","description":"The license text."}},"type":"object","required":["name","path","sha","size","url","type","content"],"title":"LicenseEndpointModel","description":"A model representing the license response."},"ManifestEndpointModel":{"properties":{"manifest_file":{"type":"string","title":"Manifest File","description":"The content of the manifest file describing the required model artifacts."}},"type":"object","required":["manifest_file"],"title":"ManifestEndpointModel","description":"A model representing the manifest response."},"Message":{"properties":{"role":{"anyOf":[{"type":"string"},{"type":"null"}],"title":"Role","default":""},"content":{"items":{"$ref":"#/components/schemas/Image"},"type":"array","minItems":1,"title":"Content"}},"type":"object","required":["content"],"title":"Message"},"MetadataEndpointModel":{"properties":{"assetInfo":{"items":{"type":"string"},"type":"array","title":"Assetinfo","description":"A list of required container assets excluding model artifacts"},"licenseInfo":{"$ref":"#/components/schemas/LicenseEndpointModel","description":"The license info."},"modelInfo":{"items":{"$ref":"#/components/schemas/ModelInfo"},"type":"array","title":"Modelinfo","description":"A list of models being served by the NIM."},"version":{"type":"string","title":"Version","description":"The version of the NIM service."}},"type":"object","required":["assetInfo","licenseInfo","modelInfo","version"],"title":"MetadataEndpointModel","description":"A model representing the metadata response."},"ModelInfo":{"properties":{"modelUrl":{"type":"string","title":"Modelurl"},"shortName":{"type":"string","title":"Shortname"}},"type":"object","required":["modelUrl","shortName"],"title":"ModelInfo","description":"A model representing the model response."},"PaddleOCRRequest":{"properties":{"model":{"type":"string","enum":["baidu/paddleocr"],"const":"baidu/paddleocr","title":"Model Name","default":"baidu/paddleocr"},"messages":{"items":{"$ref":"#/components/schemas/Message"},"type":"array","maxItems":1,"minItems":1,"title":"Messages"}},"type":"object","required":["messages"],"title":"PaddleOCRRequest"},"PaddleOCRResponse":{"properties":{"object":{"type":"string","enum":["list"],"const":"list","title":"Object","default":"list"},"data":{"items":{"$ref":"#/components/schemas/PaddleOCRResponseItem"},"type":"array","title":"Data"},"model":{"type":"string","title":"Model"},"usage":{"anyOf":[{"type":"object"},{"type":"null"}],"title":"Usage"}},"type":"object","required":["data","model"],"title":"PaddleOCRResponse"},"PaddleOCRResponseItem":{"properties":{"index":{"type":"integer","title":"Index"},"content":{"type":"string","title":"Content"},"object":{"type":"string","enum":["string"],"const":"string","title":"Object","default":"string"}},"type":"object","required":["index","content"],"title":"PaddleOCRResponseItem"},"ValidationError":{"properties":{"loc":{"items":{"anyOf":[{"type":"string"},{"type":"integer"}]},"type":"array","title":"Location"},"msg":{"type":"string","title":"Message"},"type":{"type":"string","title":"Error Type"}},"type":"object","required":["loc","msg","type"],"title":"ValidationError"}}}},"namespace":"qc69jvmznzxy","updatedDate":"2025-03-25T00:02:37.700Z","nvcfFunctionId":"43ace38a-357d-4440-9385-10a9d4c9a581","createdDate":"2024-11-22T17:15:01.822Z","attributes":{"apiDocsUrl":"https://docs.nvidia.com/nim/ingestion/table-extraction/latest/api-reference.html","termsOfUse":"GOVERNING TERMS: The trial service is governed by the NVIDIA API Trial Terms of Service; and the use of this model is governed by the NVIDIA Evaluation License Agreement. Additional Information: Apache License Version 2.0.\\n","showUnavailableBanner":false,"cta":{"text":"Apply to Self-Host","url":"https://www.nvidia.com/en-us/ai/nim-notifyme/"},"deploy":[{"label":"Linux with Docker","filename":"linux.md","contents":"$8a"},{"label":"Windows on RTX AI PCs (Beta)","filename":"wsl2.md","contents":"$8b"}]},"artifactName":"paddleocr"},"config":{"name":"paddleocr","type":"model"}},{"endpoint":{"artifact":{"artifactType":"ENDPOINT","name":"nv-yolox-page-elements-v1","displayName":"nv-yolox-page-elements-v1","publisher":"nvidia","shortDescription":"Model for object detection, fine-tuned to detect charts, tables, and titles in documents.","logo":"https://assets.ngc.nvidia.com/products/api-catalog/images/nv-yolox-page-elements-v1.jpg","labels":["Chart Detection","Data ingestion","Object Detection","Table Detection","extraction","nemo retriever","run on rtx"],"attributes":[{"key":"AVAILABLE","value":"true"},{"key":"PREVIEW","value":"false"}],"canGuestDownload":true,"createdDate":"2024-11-22T17:15:01.485Z","description":"$8c","isPublic":true,"isReadOnly":true,"orgName":"qc69jvmznzxy","updatedDate":"2025-03-25T00:03:00.623Z"},"requestStatus":{"statusCode":"SUCCESS","requestId":"99bff37b-51ce-4083-a442-738335602d2c"}},"spec":{"openAPISpec":{"openapi":"3.1.0","info":{"title":"NeMo Retriever YOLOX Structured Images v1","description":"You Only Look Once (YOLO) anchor-free object detection of charts and tables in structured images.","version":"0.2.1-rc0"},"paths":{"/v1/infer":{"post":{"summary":"Post V1 Infer","description":"Detect charts and tables in the provided images.","operationId":"post_v1_infer_v1_infer_post","requestBody":{"content":{"application/json":{"schema":{"$ref":"#/components/schemas/YoloxRequest"}}},"required":true},"responses":{"200":{"description":"Successful Response","content":{"application/json":{"schema":{"$ref":"#/components/schemas/YoloxResponse"}}}},"422":{"description":"Validation Error","content":{"application/json":{"schema":{"$ref":"#/components/schemas/HTTPValidationError"}}}}},"x-nvai-meta":{"name":"Inference model on image","returns":"Returns a object list.\\n","path":"create","examples":[{"name":"Image from example","input":{"image":"https://assets.ngc.nvidia.com/products/api-catalog/yolox/yolox1.jpg"},"requestJson":"$8d","responseJson":"$8e"},{"name":"Image from example","input":{"image":"https://assets.ngc.nvidia.com/products/api-catalog/yolox/yolox2.jpg"},"requestJson":"$8f","responseJson":"$90"}],"templates":[{"title":"Default","requestEjs":{"python":"import requests, base64\\n\\ninvoke_url = \\"https://ai.api.nvidia.com/v1/cv/nvidia/nv-yolox-page-elements-v1\\"\\n\\nwith open(\\"yolox1.png\\", \\"rb\\") as f:\\n image_b64 = base64.b64encode(f.read()).decode()\\n\\nassert len(image_b64) < 180_000, \\\\\\n \\"To upload larger images, use the assets API (see docs)\\"\\n\\nheaders = {\\n \\"Authorization\\": \\"Bearer $NVIDIA_API_KEY\\",\\n \\"Accept\\": \\"application/json\\"\\n}\\n\\npayload = {\\n \\"input\\": [\\n {\\n \\"type\\": \\"image_url\\",\\n \\"url\\": f\\"data:image/png;base64,{image_b64}\\"\\n }\\n ]\\n}\\n\\nresponse = requests.post(invoke_url, headers=headers, json=payload)\\n\\nprint(response.json())\\n","node.js":"import axios from \'axios\';\\nimport { readFile } from \'node:fs/promises\';\\n\\nconst invokeUrl = \\"https://ai.api.nvidia.com/v1/cv/nvidia/nv-yolox-page-elements-v1\\";\\n\\nconst headers = {\\n \\"Authorization\\": \\"Bearer $NVIDIA_API_KEY\\",\\n \\"Accept\\": \\"application/json\\"\\n};\\n\\nreadFile(\\"yolox1.png\\")\\n .then(data => {\\n const imageB64 = Buffer.from(data).toString(\'base64\');\\n if (imageB64.length > 180_000) {\\n throw new Error(\\"To upload larger images, use the assets API (see docs)\\");\\n }\\n\\n const payload = {\\n \\"input\\": [\\n {\\n \\"type\\": \\"image_url\\",\\n \\"url\\": `data:image/png;base64,${imageB64}`\\n }\\n ]\\n };\\n\\n return axios.post(invokeUrl, payload, { headers: headers, responseType: \'json\' });\\n })\\n .then(response => {\\n console.log(JSON.stringify(response.data));\\n })\\n .catch(error => {\\n console.error(error);\\n });\\n","curl":"image_b64=$( base64 -i yolox1.png )\\n\\naccept_header=\'Accept: application/json\'\\n\\n# Construct the JSON payload\\necho \'{\\n \\"input\\": [\\n {\\n \\"type\\": \\"image_url\\",\\n \\"url\\": \\"data:image/png;base64,\'\\"${image_b64}\\"\'\\"\\n }\\n ]\\n}\' > payload.json\\n\\ncurl https://ai.api.nvidia.com/v1/cv/nvidia/nv-yolox-page-elements-v1 \\\\\\n -H \\"Authorization: Bearer $NVIDIA_API_KEY\\" \\\\\\n -H \\"Content-Type: application/json\\" \\\\\\n -H \\"$accept_header\\" \\\\\\n -d @payload.json\\n"},"response":"$91"}]}}},"/v1/health/live":{"get":{"summary":"Health Live","description":"Check if the service is running.","operationId":"health_live_v1_health_live_get","responses":{"200":{"description":"Successful Response","content":{"application/json":{"schema":{"$ref":"#/components/schemas/HealthLiveResponse"}}}}}}},"/v1/health/ready":{"get":{"summary":"Health Ready","description":"Check if the service is ready to recieve traffic.","operationId":"health_ready_v1_health_ready_get","responses":{"200":{"description":"Successful Response","content":{"application/json":{"schema":{"$ref":"#/components/schemas/HealthReadyResponse"}}}}}}},"/v1/metrics":{"get":{"summary":"Metrics","description":"Handler for metrics endpoint.","operationId":"metrics_v1_metrics_get","responses":{"200":{"description":"Successful Response","content":{"application/json":{"schema":{"type":"string","title":"Response Metrics V1 Metrics Get"}}}}}}},"/v1/license":{"get":{"summary":"License","description":"Handler for license endpoint.","operationId":"license_v1_license_get","responses":{"200":{"description":"Successful Response","content":{"application/json":{"schema":{"$ref":"#/components/schemas/LicenseEndpointModel"}}}}}}},"/v1/metadata":{"get":{"summary":"Metadata","description":"Handler for metadata endpoint.","operationId":"metadata_v1_metadata_get","responses":{"200":{"description":"Successful Response","content":{"application/json":{"schema":{"$ref":"#/components/schemas/MetadataEndpointModel"}}}}}}},"/v1/manifest":{"get":{"summary":"Manifest","description":"Handler for the manifest endpoint.","operationId":"manifest_v1_manifest_get","responses":{"200":{"description":"Successful Response","content":{"application/json":{"schema":{"$ref":"#/components/schemas/ManifestEndpointModel"}}}}}}}},"components":{"schemas":{"HTTPValidationError":{"properties":{"detail":{"items":{"$ref":"#/components/schemas/ValidationError"},"type":"array","title":"Detail"}},"type":"object","title":"HTTPValidationError"},"HealthLiveResponse":{"properties":{"live":{"type":"boolean","title":"Live"}},"type":"object","required":["live"],"title":"HealthLiveResponse"},"HealthReadyResponse":{"properties":{"ready":{"type":"boolean","title":"Ready"}},"type":"object","required":["ready"],"title":"HealthReadyResponse"},"LicenseEndpointModel":{"properties":{"name":{"type":"string","title":"Name","description":"The name of the license for the NIM container."},"path":{"type":"string","title":"Path","description":"The filepath within the container containing the license content."},"sha":{"type":"string","title":"Sha","description":"A SHA1 hash of the license contents."},"size":{"type":"integer","title":"Size","description":"The number of characters in the license content."},"url":{"type":"string","title":"Url","description":"The url where this license is hosted externally."},"type":{"type":"string","enum":["file"],"const":"file","title":"Type","description":"The format of the license content."},"content":{"type":"string","title":"Content","description":"The license text."}},"type":"object","required":["name","path","sha","size","url","type","content"],"title":"LicenseEndpointModel","description":"A model representing the license response."},"ManifestEndpointModel":{"properties":{"manifest_file":{"type":"string","title":"Manifest File","description":"The content of the manifest file describing the required model artifacts."}},"type":"object","required":["manifest_file"],"title":"ManifestEndpointModel","description":"A model representing the manifest response."},"MetadataEndpointModel":{"properties":{"assetInfo":{"items":{"type":"string"},"type":"array","title":"Assetinfo","description":"A list of required container assets excluding model artifacts"},"licenseInfo":{"$ref":"#/components/schemas/LicenseEndpointModel","description":"The license info."},"modelInfo":{"items":{"$ref":"#/components/schemas/ModelInfo"},"type":"array","title":"Modelinfo","description":"A list of models being served by the NIM."},"version":{"type":"string","title":"Version","description":"The version of the NIM service."}},"type":"object","required":["assetInfo","licenseInfo","modelInfo","version"],"title":"MetadataEndpointModel","description":"A model representing the metadata response."},"ModelInfo":{"properties":{"modelUrl":{"type":"string","title":"Modelurl"},"shortName":{"type":"string","title":"Shortname"}},"type":"object","required":["modelUrl","shortName"],"title":"ModelInfo","description":"A model representing the model response."},"Type":{"type":"string","enum":["chart","table","title"],"title":"Type"},"ValidationError":{"properties":{"loc":{"items":{"anyOf":[{"type":"string"},{"type":"integer"}]},"type":"array","title":"Location"},"msg":{"type":"string","title":"Message"},"type":{"type":"string","title":"Error Type"}},"type":"object","required":["loc","msg","type"],"title":"ValidationError"},"YoloxBoundingBox":{"properties":{"xmin":{"type":"number","title":"Xmin"},"ymin":{"type":"number","title":"Ymin"},"xmax":{"type":"number","title":"Xmax"},"ymax":{"type":"number","title":"Ymax"},"confidence":{"type":"number","title":"Confidence"}},"type":"object","required":["xmin","ymin","xmax","ymax","confidence"],"title":"YoloxBoundingBox"},"YoloxImage":{"properties":{"type":{"type":"string","enum":["image_url"],"const":"image_url","title":"Type","default":"image_url"},"image_url":{"$ref":"#/components/schemas/YoloxImageUrl"}},"type":"object","required":["image_url"],"title":"YoloxImage"},"YoloxImageUrl":{"properties":{"url":{"type":"string","title":"Url","description":"The URL of the image to be processed."}},"type":"object","required":["url"],"title":"YoloxImageUrl"},"YoloxMessage":{"properties":{"role":{"anyOf":[{"type":"string"},{"type":"null"}],"title":"Role","default":""},"content":{"items":{"$ref":"#/components/schemas/YoloxImage"},"type":"array","minItems":1,"title":"Content"}},"type":"object","required":["content"],"title":"YoloxMessage"},"YoloxRequest":{"properties":{"model":{"type":"string","enum":["nvidia/nv-yolox-structured-images-v1"],"const":"nvidia/nv-yolox-structured-images-v1","title":"Model","default":"nvidia/nv-yolox-structured-images-v1"},"messages":{"items":{"$ref":"#/components/schemas/YoloxMessage"},"type":"array","maxItems":1,"minItems":1,"title":"Messages"},"confidence_threshold":{"anyOf":[{"type":"number"},{"type":"null"}],"title":"Confidence Threshold","description":"Threshold for filtering low-confidence predictions","default":0.01},"nms_threshold":{"anyOf":[{"type":"number"},{"type":"null"}],"title":"Nms Threshold","description":"Threshold for filtering overlapping predictions via non-max suppression","default":0.5}},"type":"object","required":["messages"],"title":"YoloxRequest"},"YoloxResponse":{"properties":{"object":{"type":"string","enum":["list"],"const":"list","title":"Object","default":"list"},"data":{"items":{"items":{"$ref":"#/components/schemas/YoloxResponseItem"},"type":"array"},"type":"array","title":"Data"},"model":{"type":"string","title":"Model"},"usage":{"anyOf":[{"type":"object"},{"type":"null"}],"title":"Usage"}},"type":"object","required":["data","model"],"title":"YoloxResponse"},"YoloxResponseItem":{"properties":{"bboxes":{"items":{"$ref":"#/components/schemas/YoloxBoundingBox"},"type":"array","title":"Bboxes"},"type":{"$ref":"#/components/schemas/Type"}},"type":"object","required":["bboxes","type"],"title":"YoloxResponseItem"}}}},"namespace":"qc69jvmznzxy","updatedDate":"2025-03-25T00:03:03.075Z","nvcfFunctionId":"96d723c1-5970-49a2-917a-e799e078242d","createdDate":"2024-11-22T17:15:02.010Z","attributes":{"apiDocsUrl":"https://docs.nvidia.com/nim/ingestion/object-detection/latest/api-reference.html","termsOfUse":"GOVERNING TERMS: The trial service is governed by the NVIDIA API Trial Terms of Service; and the use of this model is governed by the NVIDIA Evaluation License Agreement. Additional Information: Apache License Version 2.0.\\n","showUnavailableBanner":false,"cta":{"text":"Apply to Self-Host","url":"https://www.nvidia.com/en-us/ai/nim-notifyme/"},"deploy":[{"label":"Linux with Docker","filename":"linux.md","contents":"$92"},{"label":"Windows on RTX AI PCs (Beta)","filename":"wsl2.md","contents":"$93"}]},"artifactName":"nv-yolox-page-elements-v1"},"config":{"name":"nv-yolox-page-elements-v1","type":"model"}}],"items":["$94","$95","$96","$97","$98","$99","$9a","$9b","$9c","$9d"],"params":{},"slotTitle":[["$","div",null,{"className":"mb-2 flex items-start gap-2 max-xs:justify-between","children":[["$","h2",null,{"className":"text-ml font-medium leading-body tracking-less text-manitoulinLightWhite mb-0","children":"AI Models for RTX AI PCs and Workstations"}],["$","$L25",null,{"href":"/search?q=run-on-rtx","children":[["$","$L4e",null,{"children":["$","svg","arrow-right:fill",{"data-src":"https://brand-assets.cne.ngc.nvidia.com/assets/icons/3.1.0/fill/arrow-right.svg","height":"1em","width":"1em","display":"inline-block","data-icon-name":"arrow-right","data-cache":"disabled","color":"$undefined","className":"btn-icon"}]}],"View All"],"className":"inline-flex items-center justify-center gap-2 text-center font-sans font-medium leading-text flex-row-reverse btn-tertiary btn-sm btn-pill text-nowrap mt-[3px]"}]]}],["$","p",null,{"className":"text-md font-normal text-manitoulinLightGray mb-0","children":"Spanning language, speech, animation, content generation, and vision capabilities, run NVIDIA NIM microservices on your RTX AI PC."}]," "]}]\n'])
Spanning language, speech, animation, content generation, and vision capabilities, run NVIDIA NIM microservices on your RTX AI PC.
Pre-trained foundation models and blueprints for digital twins, synthetic data generation, and robotic simulation to accelerate Physical AI development.
Blueprints to help you expedite simulation and development with NVIDIA Omniverse.
AI-driven drug discovery and accelerated genomics workflows.