Skip to content
Open
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
37 changes: 27 additions & 10 deletions main.py
Original file line number Diff line number Diff line change
Expand Up @@ -1067,19 +1067,36 @@ def load_crawl_into_session(crawl_id):
crawler.base_url = crawl['base_url']
crawler.base_domain = crawl['base_domain']



# Initialize link manager if needed (it's None until a crawl starts)
if not crawler.link_manager:
from src.core.link_manager import LinkManager
crawler.link_manager = LinkManager(crawl['base_domain'])

# Load links into link manager
if crawler.link_manager:
crawler.link_manager.all_links = links
# Rebuild links_set
crawler.link_manager.links_set.clear()
for link in links:
link_key = f"{link['source_url']}|{link['target_url']}"
crawler.link_manager.links_set.add(link_key)
crawler.link_manager.all_links = links
# Rebuild links_set
crawler.link_manager.links_set.clear()
for link in links:
link_key = f"{link['source_url']}|{link['target_url']}"
crawler.link_manager.links_set.add(link_key)

# Populate discovered/visited URL sets so get_stats() returns correct counts
crawler.link_manager.all_discovered_urls = set(u['url'] for u in urls)
crawler.link_manager.visited_urls = set(u['url'] for u in urls)

# Initialize issue detector if needed (it's None until a crawl starts)
if not crawler.issue_detector:
from src.core.issue_detector import IssueDetector
crawler.issue_detector = IssueDetector()

# Load issues into issue detector
if crawler.issue_detector:
crawler.issue_detector.detected_issues = issues
crawler.issue_detector.detected_issues = issues




# Set Flask session flag for force full refresh
session['force_full_refresh'] = True

Expand Down Expand Up @@ -1437,4 +1454,4 @@ def open_browser():
serve(app, host='0.0.0.0', port=5000, threads=8)

if __name__ == '__main__':
main()
main()