Easy Dataset
Github
English
English
  • Product Introduction
  • Installation and Use
  • Basic
    • Projects
      • Task Configuration
      • Model Configuration
      • Model Testing
    • Documents
      • Document Processing
      • Domain Tags
  • Questions
    • Question Generation
    • Question Management
  • Datasets
    • Dataset Generation
    • Dataset Management
    • Dataset Export
    • Dataset Marketplace
  • Evaluations
    • Fine-tuning Evaluation
  • Advanced
    • Text Spliting
    • Custom Prompts
    • Distilled Datasets
  • BLOG
    • Release Notes
    • Community Tutorials
  • Dataset Knowledge
  • MORE
    • FAQ
    • Privacy Policy
    • Contact Us
Powered by GitBook

Copyright © 2025 Easy Dataset

On this page
  • Q: How to generate an English dataset?
  • Q: Can't find the desired model provider and model in the model configuration?
  • Q: The model test is fine, but it reports an error when generating questions or datasets?
  • Q: The batch task processing speed is too slow?
  • Q: The batch task is interrupted suddenly, and it starts to complete quickly at a certain node?
  • Q: The questions or datasets are not output in the expected style?

Was this helpful?

Edit on GitHub
Export as PDF
  1. MORE

FAQ

PreviousDataset KnowledgeNextPrivacy Policy

Last updated 19 days ago

Was this helpful?

Q: How to generate an English dataset?

The system will decide the final language of the generated dataset based on the current user's language selection. Currently, it supports Chinese and English. The default language environment is Chinese. If you need to generate an English dataset, you need to manually switch to English.


Q: Can't find the desired model provider and model in the model configuration?

Currently, it supports OpenAI standard protocol model access, compatible with Ollama. The system only has some common model configurations built-in. If you can't find the desired model, you can customize the model provider, model name, API address, and key.


Q: The model test is fine, but it reports an error when generating questions or datasets?

In many cases, the system requires the model to output in a specified JSON format. If the model's understanding ability or context length is insufficient, the output may be unstable. It is recommended to replace it with a model with a larger parameter quantity and longer context length.


Q: The batch task processing speed is too slow?

The processing speed of the task is largely determined by the processing speed of the selected model. If it is a local model, please check the resource utilization rate. If it is a remote model, it is recommended to replace it with a faster and more stable platform.


Q: The batch task is interrupted suddenly, and it starts to complete quickly at a certain node?

It is likely that the model's rate limiting strategy has been triggered, which is common in unpaid Silicon Flow and free OpenRouter models. You can manually reduce the concurrent processing number in the task configuration, which is currently set to 5 by default.


Q: The questions or datasets are not output in the expected style?

You can add custom prompt words in the project configuration - prompt word configuration to actively intervene.