| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990 |
- from flask_restx import Resource, fields, reqparse
-
- from controllers.console import api, console_ns
- from controllers.console.datasets.error import WebsiteCrawlError
- from controllers.console.wraps import account_initialization_required, setup_required
- from libs.login import login_required
- from services.website_service import WebsiteCrawlApiRequest, WebsiteCrawlStatusApiRequest, WebsiteService
-
-
- @console_ns.route("/website/crawl")
- class WebsiteCrawlApi(Resource):
- @api.doc("crawl_website")
- @api.doc(description="Crawl website content")
- @api.expect(
- api.model(
- "WebsiteCrawlRequest",
- {
- "provider": fields.String(
- required=True,
- description="Crawl provider (firecrawl/watercrawl/jinareader)",
- enum=["firecrawl", "watercrawl", "jinareader"],
- ),
- "url": fields.String(required=True, description="URL to crawl"),
- "options": fields.Raw(required=True, description="Crawl options"),
- },
- )
- )
- @api.response(200, "Website crawl initiated successfully")
- @api.response(400, "Invalid crawl parameters")
- @setup_required
- @login_required
- @account_initialization_required
- def post(self):
- parser = reqparse.RequestParser()
- parser.add_argument(
- "provider",
- type=str,
- choices=["firecrawl", "watercrawl", "jinareader"],
- required=True,
- nullable=True,
- location="json",
- )
- parser.add_argument("url", type=str, required=True, nullable=True, location="json")
- parser.add_argument("options", type=dict, required=True, nullable=True, location="json")
- args = parser.parse_args()
-
- # Create typed request and validate
- try:
- api_request = WebsiteCrawlApiRequest.from_args(args)
- except ValueError as e:
- raise WebsiteCrawlError(str(e))
-
- # Crawl URL using typed request
- try:
- result = WebsiteService.crawl_url(api_request)
- except Exception as e:
- raise WebsiteCrawlError(str(e))
- return result, 200
-
-
- @console_ns.route("/website/crawl/status/<string:job_id>")
- class WebsiteCrawlStatusApi(Resource):
- @api.doc("get_crawl_status")
- @api.doc(description="Get website crawl status")
- @api.doc(params={"job_id": "Crawl job ID", "provider": "Crawl provider (firecrawl/watercrawl/jinareader)"})
- @api.response(200, "Crawl status retrieved successfully")
- @api.response(404, "Crawl job not found")
- @api.response(400, "Invalid provider")
- @setup_required
- @login_required
- @account_initialization_required
- def get(self, job_id: str):
- parser = reqparse.RequestParser()
- parser.add_argument(
- "provider", type=str, choices=["firecrawl", "watercrawl", "jinareader"], required=True, location="args"
- )
- args = parser.parse_args()
-
- # Create typed request and validate
- try:
- api_request = WebsiteCrawlStatusApiRequest.from_args(args, job_id)
- except ValueError as e:
- raise WebsiteCrawlError(str(e))
-
- # Get crawl status using typed request
- try:
- result = WebsiteService.get_crawl_status_typed(api_request)
- except Exception as e:
- raise WebsiteCrawlError(str(e))
- return result, 200
|