Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
8 changes: 8 additions & 0 deletions docs/_apps/parakeet-wrapper/index.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,8 @@
---
layout: posts
classes: wide
title: parakeet-wrapper
date: 1970-01-01T00:00:00+00:00
---
A CLAMS wrapper for NVIDIA NeMo Parakeet ASR models available on huggingface-hub with support for punctuation, capitalization, and word-level timestamping.
- [v1.0](v1.0) ([`@shel-ho`](https://github.com/shel-ho))
105 changes: 105 additions & 0 deletions docs/_apps/parakeet-wrapper/v1.0/index.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,105 @@
---
layout: posts
classes: wide
title: "Parakeet Wrapper (v1.0)"
date: 2025-07-29T15:07:13+00:00
---
## About this version

- Submitter: [shel-ho](https://github.com/shel-ho)
- Submission Time: 2025-07-29T15:07:13+00:00
- Prebuilt Container Image: [ghcr.io/clamsproject/app-parakeet-wrapper:v1.0](https://github.com/clamsproject/app-parakeet-wrapper/pkgs/container/app-parakeet-wrapper/v1.0)
- Release Notes

(no notes provided by the developer)

## About this app (See raw [metadata.json](metadata.json))

**A CLAMS wrapper for NVIDIA NeMo Parakeet ASR models available on huggingface-hub with support for punctuation, capitalization, and word-level timestamping.**

- App ID: [http://apps.clams.ai/parakeet-wrapper/v1.0](http://apps.clams.ai/parakeet-wrapper/v1.0)
- App License: Apache-2.0
- Source Repository: [https://github.com/clamsproject/app-parakeet-wrapper](https://github.com/clamsproject/app-parakeet-wrapper) ([source tree of the submitted version](https://github.com/clamsproject/app-parakeet-wrapper/tree/v1.0))
- Analyzer Version: 20250714
- Analyzer License: cc-by-4.0


#### Inputs
(**Note**: "*" as a property value means that the property is required but can be any value.)

One of the following is required: [
- [http://mmif.clams.ai/vocabulary/AudioDocument/v1](http://mmif.clams.ai/vocabulary/AudioDocument/v1) (required)
(of any properties)

- [http://mmif.clams.ai/vocabulary/VideoDocument/v1](http://mmif.clams.ai/vocabulary/VideoDocument/v1) (required)
(of any properties)



]


#### Configurable Parameters
(**Note**: _Multivalued_ means the parameter can have one or more values.)

- `contextSize`: optional, defaults to `400`

- Type: integer
- Multivalued: False


> Local attention context size for the model. Can be any positive integer, or 0 to set global (full-context) attention. Larger context sizes may improve performance but require a lot more memory. For desktop CUDA device with 12GB VRAM, a context size of around 100 is recommended for full utilization of VRAM. Default is 400
- `modelSize`: optional, defaults to `0.6b`

- Type: string
- Multivalued: False
- Choices: `110m`, **_`0.6b`_**, `1.1b`


> Parakeet model size to use. Choices: 110m, 0.6b, 1.1b
- `pretty`: optional, defaults to `false`

- Type: boolean
- Multivalued: False
- Choices: **_`false`_**, `true`


> The JSON body of the HTTP response will be re-formatted with 2-space indentation
- `runningTime`: optional, defaults to `false`

- Type: boolean
- Multivalued: False
- Choices: **_`false`_**, `true`


> The running time of the app will be recorded in the view metadata
- `hwFetch`: optional, defaults to `false`

- Type: boolean
- Multivalued: False
- Choices: **_`false`_**, `true`


> The hardware information (architecture, GPU and vRAM) will be recorded in the view metadata


#### Outputs
(**Note**: "*" as a property value means that the property is required but can be any value.)

(**Note**: Not all output annotations are always generated.)

- [http://mmif.clams.ai/vocabulary/TextDocument/v1](http://mmif.clams.ai/vocabulary/TextDocument/v1)
(of any properties)

- [http://mmif.clams.ai/vocabulary/TimeFrame/v6](http://mmif.clams.ai/vocabulary/TimeFrame/v6)
(of any properties)

- [http://mmif.clams.ai/vocabulary/Alignment/v1](http://mmif.clams.ai/vocabulary/Alignment/v1)
(of any properties)

- [http://vocab.lappsgrid.org/Token](http://vocab.lappsgrid.org/Token)
(of any properties)

- [http://vocab.lappsgrid.org/Sentence](http://vocab.lappsgrid.org/Sentence)
(of any properties)

82 changes: 82 additions & 0 deletions docs/_apps/parakeet-wrapper/v1.0/metadata.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,82 @@
{
"name": "Parakeet Wrapper",
"description": "A CLAMS wrapper for NVIDIA NeMo Parakeet ASR models available on huggingface-hub with support for punctuation, capitalization, and word-level timestamping.",
"app_version": "v1.0",
"mmif_version": "1.1.0",
"analyzer_version": "20250714",
"app_license": "Apache-2.0",
"analyzer_license": "cc-by-4.0",
"identifier": "http://apps.clams.ai/parakeet-wrapper/v1.0",
"url": "https://github.com/clamsproject/app-parakeet-wrapper",
"input": [
[
{
"@type": "http://mmif.clams.ai/vocabulary/AudioDocument/v1",
"required": true
},
{
"@type": "http://mmif.clams.ai/vocabulary/VideoDocument/v1",
"required": true
}
]
],
"output": [
{
"@type": "http://mmif.clams.ai/vocabulary/TextDocument/v1"
},
{
"@type": "http://mmif.clams.ai/vocabulary/TimeFrame/v6"
},
{
"@type": "http://mmif.clams.ai/vocabulary/Alignment/v1"
},
{
"@type": "http://vocab.lappsgrid.org/Token"
},
{
"@type": "http://vocab.lappsgrid.org/Sentence"
}
],
"parameters": [
{
"name": "contextSize",
"description": "Local attention context size for the model. Can be any positive integer, or 0 to set global (full-context) attention. Larger context sizes may improve performance but require a lot more memory. For desktop CUDA device with 12GB VRAM, a context size of around 100 is recommended for full utilization of VRAM. Default is 400",
"type": "integer",
"default": 400,
"multivalued": false
},
{
"name": "modelSize",
"description": "Parakeet model size to use. Choices: 110m, 0.6b, 1.1b",
"type": "string",
"choices": [
"110m",
"0.6b",
"1.1b"
],
"default": "0.6b",
"multivalued": false
},
{
"name": "pretty",
"description": "The JSON body of the HTTP response will be re-formatted with 2-space indentation",
"type": "boolean",
"default": false,
"multivalued": false
},
{
"name": "runningTime",
"description": "The running time of the app will be recorded in the view metadata",
"type": "boolean",
"default": false,
"multivalued": false
},
{
"name": "hwFetch",
"description": "The hardware information (architecture, GPU and vRAM) will be recorded in the view metadata",
"type": "boolean",
"default": false,
"multivalued": false
}
]
}
5 changes: 5 additions & 0 deletions docs/_apps/parakeet-wrapper/v1.0/submission.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,5 @@
{
"time": "2025-07-29T15:07:13+00:00",
"submitter": "shel-ho",
"image": "ghcr.io/clamsproject/app-parakeet-wrapper:v1.0"
}
10 changes: 10 additions & 0 deletions docs/_data/app-index.json
Original file line number Diff line number Diff line change
@@ -1,4 +1,14 @@
{
"http://apps.clams.ai/parakeet-wrapper": {
"description": "A CLAMS wrapper for NVIDIA NeMo Parakeet ASR models available on huggingface-hub with support for punctuation, capitalization, and word-level timestamping.",
"latest_update": "2025-07-29T15:07:13+00:00",
"versions": [
[
"v1.0",
"shel-ho"
]
]
},
"http://apps.clams.ai/swt-detection": {
"description": "Detects scenes with text, like slates, chyrons and credits. This app can run in three modes, depending on `useClassifier`, `useStitcher` parameters. When `useClassifier=True`, it runs in the \"TimePoint mode\" and generates TimePoint annotations. When `useStitcher=True`, it runs in the \"TimeFrame mode\" and generates TimeFrame annotations based on existing TimePoint annotations -- if no TimePoint is found, it produces an error. By default, it runs in the 'both' mode and first generates TimePoint annotations and then TimeFrame annotations on them.",
"latest_update": "2025-07-28T15:43:20+00:00",
Expand Down
2 changes: 1 addition & 1 deletion docs/_data/apps.json

Large diffs are not rendered by default.

Loading