Ви не можете вибрати більше 25 тем Теми мають розпочинатися з літери або цифри, можуть містити дефіси (-) і не повинні перевищувати 35 символів.

3 місяці тому
123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869
  1. from flask_restful import Resource, reqparse
  2. from controllers.console import api
  3. from controllers.console.datasets.error import WebsiteCrawlError
  4. from controllers.console.wraps import account_initialization_required, setup_required
  5. from libs.login import login_required
  6. from services.website_service import WebsiteCrawlApiRequest, WebsiteCrawlStatusApiRequest, WebsiteService
  7. class WebsiteCrawlApi(Resource):
  8. @setup_required
  9. @login_required
  10. @account_initialization_required
  11. def post(self):
  12. parser = reqparse.RequestParser()
  13. parser.add_argument(
  14. "provider",
  15. type=str,
  16. choices=["firecrawl", "watercrawl", "jinareader"],
  17. required=True,
  18. nullable=True,
  19. location="json",
  20. )
  21. parser.add_argument("url", type=str, required=True, nullable=True, location="json")
  22. parser.add_argument("options", type=dict, required=True, nullable=True, location="json")
  23. args = parser.parse_args()
  24. # Create typed request and validate
  25. try:
  26. api_request = WebsiteCrawlApiRequest.from_args(args)
  27. except ValueError as e:
  28. raise WebsiteCrawlError(str(e))
  29. # Crawl URL using typed request
  30. try:
  31. result = WebsiteService.crawl_url(api_request)
  32. except Exception as e:
  33. raise WebsiteCrawlError(str(e))
  34. return result, 200
  35. class WebsiteCrawlStatusApi(Resource):
  36. @setup_required
  37. @login_required
  38. @account_initialization_required
  39. def get(self, job_id: str):
  40. parser = reqparse.RequestParser()
  41. parser.add_argument(
  42. "provider", type=str, choices=["firecrawl", "watercrawl", "jinareader"], required=True, location="args"
  43. )
  44. parser.add_argument("credential_id", type=str, required=True, nullable=True, location="args")
  45. args = parser.parse_args()
  46. # Create typed request and validate
  47. try:
  48. api_request = WebsiteCrawlStatusApiRequest.from_args(args, job_id)
  49. except ValueError as e:
  50. raise WebsiteCrawlError(str(e))
  51. # Get crawl status using typed request
  52. try:
  53. result = WebsiteService.get_crawl_status_typed(api_request)
  54. except Exception as e:
  55. raise WebsiteCrawlError(str(e))
  56. return result, 200
  57. api.add_resource(WebsiteCrawlApi, "/website/crawl")
  58. api.add_resource(WebsiteCrawlStatusApi, "/website/crawl/status/<string:job_id>")