curl --request POST \
--url https://be.datagol.ai/noCo/api/v2/workspaces/{workspaceId}/extraction/advanced/test \
--header 'Content-Type: application/json' \
--data @- <<EOF
{
"content": "[00:00:02] Sarah Miller: Yeah, great to connect.\n\n[00:00:04] Mark Thompson: Likewise. So just for context — Alex and I started working together recently. We’re very early, mostly speaking with potential users and understanding their pain points.\n\n[00:00:18] Sarah Miller: Quick intro from my side — I’ve spent the last 10 years in startups. Started in fintech, then built a SaaS company focused on workflow tools. Exited last year and now exploring the next venture.\n\n[00:00:32] Alex Rivera: I’m Alex, engineering background. Around 15 years building developer tools and infrastructure products. This is my second startup.\n\n[00:00:44] John Doe: Nice to meet you both. I run a small fund investing in early-stage AI and infrastructure companies. Also previously built and bootstrapped a company for nearly a decade, so I understand the founder journey well.\n\n[00:01:00] Sarah Miller: Amazing. So our thesis is simple — AI isn’t just speeding up software development, it’s changing how software should be built entirely. Instead of static apps, we’re exploring the idea of “software you can summon” — lightweight, adaptive tools that exist only when needed.\n\n[00:01:20] John Doe: Interesting. Who’s the target user?\n\n[00:01:23] Alex Rivera: Product managers, designers, and technical founders who are building tools themselves but feel underserved by existing infrastructure. We want to rethink the layer beneath the coding agents.\n\n[00:01:38] John Doe: Makes sense. The key will be articulating why you, why now, and what gives you the right to win.\n\n[00:01:46] Sarah Miller: Exactly. That’s what we’re refining — and conversations like this really help.",
"advancedSchema": {
"type": "object",
"additionalProperties": false,
"properties": {
"title": {
"type": "string",
"description": "Optional title or identifier for the transcript (e.g., meeting name, interview subject)."
},
"duration_seconds": {
"type": "number",
"description": "Total duration of the transcribed audio/video in seconds."
},
"language_code": {
"type": "string",
"description": "BCP-47 language code of the transcript (e.g., 'en-US', 'es-ES')."
},
"speaker_count": {
"type": "number",
"description": "The total number of unique speakers identified in the transcript."
},
"segments": {
"type": "array",
"description": "An array of spoken segments or utterances, each representing a continuous speech turn by a single speaker.",
"items": {
"type": "object",
"properties": {
"speaker": {
"type": "string",
"description": "Identifier for the speaker of this segment (e.g., 'Speaker 1', 'John Doe')."
},
"start_time": {
"type": "number",
"description": "The start time of the segment in seconds from the beginning of the audio/video."
},
"end_time": {
"type": "number",
"description": "The end time of the segment in seconds from the beginning of the audio/video."
},
"text": {
"type": "string",
"description": "The full transcribed text of this segment."
},
"confidence": {
"type": "number",
"description": "Confidence score for the accuracy of this segment's transcription (0.0 to 1.0)."
},
"words": {
"type": "array",
"description": "An array of individual words within this segment, with their specific timings and confidence scores.",
"items": {
"type": "object",
"properties": {
"word": {
"type": "string",
"description": "The transcribed word."
},
"start_time": {
"type": "number",
"description": "The start time of the word in seconds relative to the beginning of the audio/video."
},
"end_time": {
"type": "number",
"description": "The end time of the word in seconds relative to the beginning of the audio/video."
},
"confidence": {
"type": "number",
"description": "Confidence score for the accuracy of this word's transcription (0.0 to 1.0)."
}
},
"required": [
"word",
"start_time",
"end_time"
]
}
}
},
"required": [
"speaker",
"start_time",
"end_time",
"text"
]
}
}
},
"required": [
"segments"
]
}
}
EOFRun an advanced test extraction.
Endpoint:
https://be.datagol.ai/noCo/api/v2/workspaces/{workspaceId}/extraction/advanced/test
Auth Required: Yes (x-auth-token)
| Field | Type | Required | Description |
|---|---|---|---|
| content | string | Yes | Inline content to test extraction |
| advancedSchema | object | Yes | JSON schema for advanced extraction |
| elementId | string | No | File ID for testing instead of content |
curl --request POST \
--url https://be.datagol.ai/noCo/api/v2/workspaces/{workspaceId}/extraction/advanced/test \
--header 'Content-Type: application/json' \
--data @- <<EOF
{
"content": "[00:00:02] Sarah Miller: Yeah, great to connect.\n\n[00:00:04] Mark Thompson: Likewise. So just for context — Alex and I started working together recently. We’re very early, mostly speaking with potential users and understanding their pain points.\n\n[00:00:18] Sarah Miller: Quick intro from my side — I’ve spent the last 10 years in startups. Started in fintech, then built a SaaS company focused on workflow tools. Exited last year and now exploring the next venture.\n\n[00:00:32] Alex Rivera: I’m Alex, engineering background. Around 15 years building developer tools and infrastructure products. This is my second startup.\n\n[00:00:44] John Doe: Nice to meet you both. I run a small fund investing in early-stage AI and infrastructure companies. Also previously built and bootstrapped a company for nearly a decade, so I understand the founder journey well.\n\n[00:01:00] Sarah Miller: Amazing. So our thesis is simple — AI isn’t just speeding up software development, it’s changing how software should be built entirely. Instead of static apps, we’re exploring the idea of “software you can summon” — lightweight, adaptive tools that exist only when needed.\n\n[00:01:20] John Doe: Interesting. Who’s the target user?\n\n[00:01:23] Alex Rivera: Product managers, designers, and technical founders who are building tools themselves but feel underserved by existing infrastructure. We want to rethink the layer beneath the coding agents.\n\n[00:01:38] John Doe: Makes sense. The key will be articulating why you, why now, and what gives you the right to win.\n\n[00:01:46] Sarah Miller: Exactly. That’s what we’re refining — and conversations like this really help.",
"advancedSchema": {
"type": "object",
"additionalProperties": false,
"properties": {
"title": {
"type": "string",
"description": "Optional title or identifier for the transcript (e.g., meeting name, interview subject)."
},
"duration_seconds": {
"type": "number",
"description": "Total duration of the transcribed audio/video in seconds."
},
"language_code": {
"type": "string",
"description": "BCP-47 language code of the transcript (e.g., 'en-US', 'es-ES')."
},
"speaker_count": {
"type": "number",
"description": "The total number of unique speakers identified in the transcript."
},
"segments": {
"type": "array",
"description": "An array of spoken segments or utterances, each representing a continuous speech turn by a single speaker.",
"items": {
"type": "object",
"properties": {
"speaker": {
"type": "string",
"description": "Identifier for the speaker of this segment (e.g., 'Speaker 1', 'John Doe')."
},
"start_time": {
"type": "number",
"description": "The start time of the segment in seconds from the beginning of the audio/video."
},
"end_time": {
"type": "number",
"description": "The end time of the segment in seconds from the beginning of the audio/video."
},
"text": {
"type": "string",
"description": "The full transcribed text of this segment."
},
"confidence": {
"type": "number",
"description": "Confidence score for the accuracy of this segment's transcription (0.0 to 1.0)."
},
"words": {
"type": "array",
"description": "An array of individual words within this segment, with their specific timings and confidence scores.",
"items": {
"type": "object",
"properties": {
"word": {
"type": "string",
"description": "The transcribed word."
},
"start_time": {
"type": "number",
"description": "The start time of the word in seconds relative to the beginning of the audio/video."
},
"end_time": {
"type": "number",
"description": "The end time of the word in seconds relative to the beginning of the audio/video."
},
"confidence": {
"type": "number",
"description": "Confidence score for the accuracy of this word's transcription (0.0 to 1.0)."
}
},
"required": [
"word",
"start_time",
"end_time"
]
}
}
},
"required": [
"speaker",
"start_time",
"end_time",
"text"
]
}
}
},
"required": [
"segments"
]
}
}
EOF