Dataset Preview
Duplicate
The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed
Error code:   DatasetGenerationError
Exception:    ArrowNotImplementedError
Message:      Cannot write struct type 'options' with no child field to Parquet. Consider adding a dummy child field.
Traceback:    Traceback (most recent call last):
                File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1887, in _prepare_split_single
                  writer.write_table(table)
                File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 673, in write_table
                  self._build_writer(inferred_schema=pa_table.schema)
                File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 714, in _build_writer
                  self.pa_writer = pq.ParquetWriter(
                                   ^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/pyarrow/parquet/core.py", line 1070, in __init__
                  self.writer = _parquet.ParquetWriter(
                                ^^^^^^^^^^^^^^^^^^^^^^^
                File "pyarrow/_parquet.pyx", line 2363, in pyarrow._parquet.ParquetWriter.__cinit__
                File "pyarrow/error.pxi", line 155, in pyarrow.lib.pyarrow_internal_check_status
                File "pyarrow/error.pxi", line 92, in pyarrow.lib.check_status
              pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'options' with no child field to Parquet. Consider adding a dummy child field.
              
              During handling of the above exception, another exception occurred:
              
              Traceback (most recent call last):
                File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1908, in _prepare_split_single
                  num_examples, num_bytes = writer.finalize()
                                            ^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 688, in finalize
                  self._build_writer(self.schema)
                File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 714, in _build_writer
                  self.pa_writer = pq.ParquetWriter(
                                   ^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/pyarrow/parquet/core.py", line 1070, in __init__
                  self.writer = _parquet.ParquetWriter(
                                ^^^^^^^^^^^^^^^^^^^^^^^
                File "pyarrow/_parquet.pyx", line 2363, in pyarrow._parquet.ParquetWriter.__cinit__
                File "pyarrow/error.pxi", line 155, in pyarrow.lib.pyarrow_internal_check_status
                File "pyarrow/error.pxi", line 92, in pyarrow.lib.check_status
              pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'options' with no child field to Parquet. Consider adding a dummy child field.
              
              The above exception was the direct cause of the following exception:
              
              Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1347, in compute_config_parquet_and_info_response
                  parquet_operations = convert_to_parquet(builder)
                                       ^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 980, in convert_to_parquet
                  builder.download_and_prepare(
                File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 884, in download_and_prepare
                  self._download_and_prepare(
                File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 947, in _download_and_prepare
                  self._prepare_split(split_generator, **prepare_split_kwargs)
                File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1736, in _prepare_split
                  for job_id, done, content in self._prepare_split_single(
                                               ^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1919, in _prepare_split_single
                  raise DatasetGenerationError("An error occurred while generating the dataset") from e
              datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

name
string
nodes
list
pinData
dict
connections
dict
active
bool
settings
dict
versionId
string
meta
dict
id
string
tags
list
My workflow
[ { "parameters": { "operation": "extractHtmlContent", "extractionValues": { "values": [ { "key": "title", "cssSelector": "h4.artTitle", "returnArray": true } ] }, "options": {}, "url": null, "assignments":...
{}
{ "HTTP Request Bursa dan Saham": { "main": [ [ { "node": "HTML", "type": "main", "index": 0 } ] ] }, "HTTP Request Emas": { "main": [ [ { "node": "HTML1", "type": "main", "index": 0 } ] ] }, "HTML": { "main": [ [ { "node": "Code Mapping", "type": "main", "index": 0 } ] ] }, "HTML1": { "main": [ [ { "node": "Code Mapping1", "type": "main", "index": 0 } ] ] }, "Edit Fields Emas": { "main": [ [ { "node": "Code1", "type": "main", "index": 0 } ] ] }, "Edit Fields Bursa Saham": { "main": [ [ { "node": "Code", "type": "main", "index": 0 } ] ] }, "Code": { "main": [ [ { "node": "Merge", "type": "main", "index": 0 } ] ] }, "Code1": { "main": [ [ { "node": "Merge", "type": "main", "index": 1 } ] ] }, "Google Gemini Chat Model": { "ai_languageModel": [ [ { "node": "Basic LLM Chain", "type": "ai_languageModel", "index": 0 } ] ] }, "Merge": { "main": [ [ { "node": "Data Cleaning", "type": "main", "index": 0 } ] ] }, "Basic LLM Chain": { "main": [ [ { "node": "Code4", "type": "main", "index": 0 }, { "node": "Basic LLM Chain1", "type": "main", "index": 0 } ] ] }, "Code4": { "main": [ [ { "node": "If", "type": "main", "index": 0 }, { "node": "If1", "type": "main", "index": 0 }, { "node": "If2", "type": "main", "index": 0 } ] ] }, "Code Mapping": { "main": [ [ { "node": "Edit Fields Bursa Saham", "type": "main", "index": 0 } ] ] }, "Code Mapping1": { "main": [ [ { "node": "Edit Fields Emas", "type": "main", "index": 0 } ] ] }, "Prompt Machine": { "main": [ [ { "node": "Basic LLM Chain", "type": "main", "index": 0 } ] ] }, "Data Cleaning": { "main": [ [ { "node": "Prompt Machine", "type": "main", "index": 0 } ] ] }, "Schedule Trigger": { "main": [ [ { "node": "HTTP Request Bursa dan Saham", "type": "main", "index": 0 }, { "node": "HTTP Request Emas", "type": "main", "index": 0 } ] ] }, "If": { "main": [ [ { "node": "Convert to File", "type": "main", "index": 0 } ] ] }, "If1": { "main": [ [ { "node": "Convert to File1", "type": "main", "index": 0 } ] ] }, "If2": { "main": [ [ { "node": "Convert to File2", "type": "main", "index": 0 } ] ] }, "Google Gemini Chat Model1": { "ai_languageModel": [ [ { "node": "Basic LLM Chain1", "type": "ai_languageModel", "index": 0 } ] ] }, "Basic LLM Chain1": { "main": [ [] ] }, "Convert to File": { "main": [ [] ] }, "Convert to File1": { "main": [ [] ] }, "Convert to File2": { "main": [ [] ] } }
false
{ "executionOrder": "v1" }
{ "instanceId": "" }
[]

No dataset card yet

Downloads last month
18