Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
23 changes: 21 additions & 2 deletions crawl4ai/async_webcrawler.py
Original file line number Diff line number Diff line change
Expand Up @@ -1075,7 +1075,26 @@ async def result_transformer():
else:
try:
_results = await dispatcher.run_urls(crawler=self, urls=urls, config=config)
return [transform_result(res) for res in _results]
data = []
from .models import CrawlerTaskResult

for task_result in _results:
for res in task_result.result:
# all child urls within parent url will have same dispatch_results
tsk = CrawlerTaskResult(
task_id=task_result.task_id,
url=task_result.url,
result=res,
memory_usage=task_result.memory_usage,
peak_memory=task_result.peak_memory,
start_time=task_result.start_time,
end_time=task_result.end_time,
error_message=task_result.error_message,
retry_count=task_result.retry_count,
wait_time=task_result.wait_time,
)
data.append(tsk)
return [transform_result(res) for res in data]
finally:
# Auto-release session after batch completes
await maybe_release_session()
Expand Down Expand Up @@ -1169,4 +1188,4 @@ async def aseed_urls(
seeding_config
)
else:
raise ValueError("`domain_or_domains` must be a string or a list of strings.")
raise ValueError("`domain_or_domains` must be a string or a list of strings.")
6 changes: 6 additions & 0 deletions crawl4ai/deep_crawling/base_strategy.py
Original file line number Diff line number Diff line change
Expand Up @@ -36,6 +36,12 @@ async def result_wrapper():
return result_wrapper()
else:
try:
crwl_rslt = [
data
for crwl_cntr in result_obj
for data in crwl_cntr._results
]
result_obj = CrawlResultContainer(results=crwl_rslt)
return result_obj
finally:
self.deep_crawl_active.reset(token)
Expand Down
Loading