Merge remote-tracking branch 'upstream/main' into fix_chrome

This commit is contained in:
yuanmengqi
2025-07-11 12:16:23 +00:00
31 changed files with 3853 additions and 206 deletions

View File

@@ -115,6 +115,11 @@ def is_expected_tabs(open_tabs: List[Dict[str, str]], rule: Dict[str, Any]) -> f
if match_type == "url":
expected_urls = rule['urls']
actual_urls = [tab['url'] for tab in open_tabs]
if not are_lists_equal(expected_urls, actual_urls, compare_urls):
logger.error("list not match")
logger.error(expected_urls)
logger.error(actual_urls)
return 0
return 1 if are_lists_equal(expected_urls, actual_urls, compare_urls) else 0
else:
logger.error(f"Unknown type: {match_type}")
@@ -343,7 +348,7 @@ def compare_archive(pred_path: str, gold_path: str, **kwargs) -> float:
return score / len(pred_files)
def compare_htmls(html_path1: str, html_path2: str) -> float:
def compare_htmls(html_path1: str, html_path2: str, **options) -> float:
"""
Compare two HTML files.
"""
@@ -351,20 +356,33 @@ def compare_htmls(html_path1: str, html_path2: str) -> float:
soup1 = BeautifulSoup(inf, 'lxml')
with open(html_path2, 'r', encoding='utf-8') as inf:
soup2 = BeautifulSoup(inf, 'lxml')
ignore_sdnum = options.get("ignore_sdnum", None)
def compare_elements(elem1, elem2):
if not (isinstance(elem1, Tag) and isinstance(elem2, Tag)):
if elem1 != elem2:
logger.info("not the same")
return elem1 == elem2
if elem1.name != elem2.name:
logger.info("html name not match")
return False
if elem1.text.strip() != elem2.text.strip():
logger.info("html text not match")
return False
if elem1.attrs != elem2.attrs:
if ignore_sdnum:
attrs1 = {k: v for k, v in elem1.attrs.items() if k != 'sdnum'}
attrs2 = {k: v for k, v in elem2.attrs.items() if k != 'sdnum'}
return attrs1 == attrs2
logger.info("html attrs not match")
logger.info(f"{elem1.attrs}")
logger.info(f"{elem2.attrs}")
return False
return True
for elem1, elem2 in zip(soup1.recursiveChildGenerator(), soup2.recursiveChildGenerator()):
if not compare_elements(elem1, elem2):
logger.info("html not match")
return .0
return 1.

View File

@@ -297,8 +297,24 @@ def compare_image_text(image_path, rule):
reader = easyocr.Reader(['en'])
result = reader.readtext(image_path)
extracted_text = ' '.join([entry[1] for entry in result])
# Log OCR results
logger.info(f"OCR extracted texts: {[entry[1] for entry in result]}")
logger.info(f"Combined extracted text: {extracted_text}")
if rule['type'] == 'text':
return 1 if rule['text'] in extracted_text else 0
target_text = rule['text']
match_found = target_text in extracted_text
# Log matching results
logger.info(f"Target text: '{target_text}'")
logger.info(f"Match found: {match_found}")
if match_found:
logger.info("✅ Text matching successful!")
else:
logger.info("❌ Text matching failed!")
return 1 if match_found else 0
else:
raise ValueError("Unsupported rule type")
@@ -986,3 +1002,10 @@ def compare_unique_train_records(processed_file, expected_files, **kwargs):
return 0
return 1
if __name__ == "__main__":
image_path = "/home/ubuntu/OSWorld/cache/02ce9a50-7af2-47ed-8596-af0c230501f8/ls.png"
print(compare_image_text(image_path, {
"type": "text",
"text": "ls"
}))

View File

@@ -213,7 +213,6 @@ _accessibility_ns_map = {
}
def check_accessibility_tree(result: str, rules: List[Dict[str, Any]], osname: str = "ubuntu") -> float:
"""
Args:
@@ -402,12 +401,17 @@ def check_direct_json_object(result, rules) -> float:
expected_value_list = expected_json.get(key)
logger.info(f"[DEBUG] Checking list key '{key}': expected_list={expected_value_list}, actual='{result.get(key)}'")
for each_expected_value in expected_value_list:
# Handle both list and string cases
if isinstance(result.get(key), list) and each_expected_value in result.get(key):
flag = 1
logger.info(f"[DEBUG] Found expected value '{each_expected_value}' in result list for key '{key}'")
break
elif isinstance(result.get(key), str) and each_expected_value == result.get(key):
flag = 1
logger.info(f"[DEBUG] Found expected value '{each_expected_value}' matches result string for key '{key}'")
break
if flag == 0:
logger.info(f"[DEBUG] No expected values found in result list for key '{key}', returning 0.0")
logger.info(f"[DEBUG] No expected values found in result for key '{key}', returning 0.0")
return 0.
elif isinstance(expected_json.get(key), str):
expected_str = expected_json.get(key)

View File

@@ -9,6 +9,31 @@ from pptx.enum.shapes import MSO_SHAPE_TYPE
logger = logging.getLogger("desktopenv.metric.slides")
# Add a new logger specifically for debugging PPTX comparisons
debug_logger = logging.getLogger("desktopenv.metric.slides.debug")
def enable_debug_logging():
"""Enable debug logging for PPTX comparison"""
debug_logger.setLevel(logging.DEBUG)
if not debug_logger.handlers:
handler = logging.StreamHandler()
handler.setLevel(logging.DEBUG)
formatter = logging.Formatter('[PPTX_DEBUG] %(message)s')
handler.setFormatter(formatter)
debug_logger.addHandler(handler)
# Add debug logger for detailed comparison output
debug_logger = logging.getLogger("desktopenv.metric.slides.debug")
def enable_debug_logging():
"""Enable detailed debug logging for PPTX comparison"""
debug_logger.setLevel(logging.DEBUG)
if not debug_logger.handlers:
handler = logging.StreamHandler()
formatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')
handler.setFormatter(formatter)
debug_logger.addHandler(handler)
def check_presenter_console_disable(config_file_path):
try:
@@ -135,11 +160,45 @@ def check_slide_numbers_color(pptx_file_path):
# return similarity_index
def get_all_text_shapes(slide):
"""递归获取slide中所有包含文本的shapes包括GROUP内部的"""
def extract_text_shapes(shape):
results = []
# 检查当前shape是否有文本
if hasattr(shape, "text") and hasattr(shape, "text_frame"):
results.append(shape)
# 如果是GROUP递归检查内部shapes
if hasattr(shape, 'shapes'):
for sub_shape in shape.shapes:
results.extend(extract_text_shapes(sub_shape))
return results
all_text_shapes = []
for shape in slide.shapes:
all_text_shapes.extend(extract_text_shapes(shape))
return all_text_shapes
def compare_pptx_files(file1_path, file2_path, **options):
# todo: not strictly match since not all information is compared because we cannot get the info through pptx
prs1 = Presentation(file1_path)
prs2 = Presentation(file2_path)
# Enable debug logging if requested
enable_debug = options.get("enable_debug", True)
if enable_debug:
enable_debug_logging()
debug_logger.debug(f"=== COMPARING PPTX FILES ===")
debug_logger.debug(f"File 1: {file1_path}")
debug_logger.debug(f"File 2: {file2_path}")
debug_logger.debug(f"File 1 slides: {len(prs1.slides)}")
debug_logger.debug(f"File 2 slides: {len(prs2.slides)}")
approximately_tolerance = options.get("approximately_tolerance", 0.005)
def is_approximately_equal(val1, val2, tolerance=approximately_tolerance):
"""Compare two values with a tolerance of 0.1% (0.005)"""
@@ -176,12 +235,17 @@ def compare_pptx_files(file1_path, file2_path, **options):
# compare the number of slides
if len(prs1.slides) != len(prs2.slides) and examine_number_of_slides:
if enable_debug:
debug_logger.debug(f"MISMATCH: Number of slides differ - File1: {len(prs1.slides)}, File2: {len(prs2.slides)}")
return 0
slide_idx = 0
# compare the content of each slide
for slide1, slide2 in zip(prs1.slides, prs2.slides):
slide_idx += 1
if enable_debug:
debug_logger.debug(f"--- Comparing Slide {slide_idx} ---")
debug_logger.debug(f"Slide {slide_idx} - Shapes count: File1={len(slide1.shapes)}, File2={len(slide2.shapes)}")
def get_slide_background_color(slide):
# background = slide.background
@@ -212,14 +276,35 @@ def compare_pptx_files(file1_path, file2_path, **options):
return None
if get_slide_notes(slide1).strip() != get_slide_notes(slide2).strip() and examine_note:
if enable_debug:
debug_logger.debug(f" MISMATCH: Slide {slide_idx} - Notes differ:")
debug_logger.debug(f" Notes1: '{get_slide_notes(slide1).strip()}'")
debug_logger.debug(f" Notes2: '{get_slide_notes(slide2).strip()}'")
return 0
# Get all text shapes including those inside GROUPs
text_shapes1 = get_all_text_shapes(slide1)
text_shapes2 = get_all_text_shapes(slide2)
if enable_debug:
debug_logger.debug(f"Slide {slide_idx} - Text shapes found: File1={len(text_shapes1)}, File2={len(text_shapes2)}")
# check if the number of slides is the same
if len(slide1.shapes) != len(slide2.shapes):
if enable_debug:
debug_logger.debug(f"MISMATCH: Slide {slide_idx} - Different number of shapes: File1={len(slide1.shapes)}, File2={len(slide2.shapes)}")
return 0
# check if the shapes are the same
shape_idx = 0
for shape1, shape2 in zip(slide1.shapes, slide2.shapes):
shape_idx += 1
if enable_debug:
debug_logger.debug(f" Shape {shape_idx} - Type: {shape1.shape_type} vs {shape2.shape_type}")
if hasattr(shape1, "text") and hasattr(shape2, "text"):
debug_logger.debug(f" Shape {shape_idx} - Text: '{shape1.text.strip()}' vs '{shape2.text.strip()}'")
debug_logger.debug(f" Shape {shape_idx} - Position: ({shape1.left}, {shape1.top}) vs ({shape2.left}, {shape2.top})")
debug_logger.debug(f" Shape {shape_idx} - Size: ({shape1.width}, {shape1.height}) vs ({shape2.width}, {shape2.height})")
if examine_title_bottom_position:
if hasattr(shape1, "text") and hasattr(shape2, "text") and shape1.text == shape2.text:
if shape1.text == "Product Comparison" and (shape1.top <= shape2.top or shape1.top < 3600000):
@@ -249,11 +334,7 @@ def compare_pptx_files(file1_path, file2_path, **options):
if slide_idx == 2 and shape1.shape_type == 13 and shape2.shape_type == 13:
if shape1.top >= shape2.top or shape1.top > 1980000:
return 0
elif (not is_approximately_equal(shape1.left, shape2.left) or
not is_approximately_equal(shape1.top, shape2.top) or
not is_approximately_equal(shape1.width, shape2.width) or
not is_approximately_equal(shape1.height, shape2.height)):
return 0
if examine_shape_for_shift_size:
if (not is_approximately_equal(shape1.left, shape2.left) or
@@ -264,11 +345,23 @@ def compare_pptx_files(file1_path, file2_path, **options):
"text") and shape1.text == shape2.text and shape1.text == "Elaborate on what you want to discuss."):
return 0
# CRITICAL: examine_shape check happens BEFORE examine_modify_height!
# If examine_shape=True (default), any shape dimension mismatch will cause immediate return 0,
# preventing examine_modify_height from ever being executed.
# For height modification tasks, you MUST set examine_shape=False to allow examine_modify_height to work.
if (
not is_approximately_equal(shape1.left, shape2.left) or
not is_approximately_equal(shape1.top, shape2.top) or
not is_approximately_equal(shape1.width, shape2.width) or
not is_approximately_equal(shape1.height, shape2.height)) and examine_shape:
if enable_debug:
debug_logger.debug(f" MISMATCH: Slide {slide_idx}, Shape {shape_idx} - Shape dimensions differ:")
debug_logger.debug(f" Left: {shape1.left} vs {shape2.left} (equal: {is_approximately_equal(shape1.left, shape2.left)})")
debug_logger.debug(f" Top: {shape1.top} vs {shape2.top} (equal: {is_approximately_equal(shape1.top, shape2.top)})")
debug_logger.debug(f" Width: {shape1.width} vs {shape2.width} (equal: {is_approximately_equal(shape1.width, shape2.width)})")
debug_logger.debug(f" Height: {shape1.height} vs {shape2.height} (equal: {is_approximately_equal(shape1.height, shape2.height)})")
if hasattr(shape1, "text") and hasattr(shape2, "text"):
debug_logger.debug(f" Shape text: '{shape1.text.strip()}' vs '{shape2.text.strip()}'")
return 0
if examine_image_size:
@@ -281,6 +374,11 @@ def compare_pptx_files(file1_path, file2_path, **options):
not is_approximately_equal(shape1.height, shape2.height)):
return 0
# examine_modify_height: Special logic for height modification tasks
# - For non-text shapes and FREEFORM shapes (type 5): Only check height differences
# - For other shapes: Check all dimensions (left, top, width, height)
# WARNING: This check only works if examine_shape=False, otherwise examine_shape will
# terminate the comparison before this code is reached!
if examine_modify_height:
if not hasattr(shape1, "text") and not hasattr(shape2,
"text") or shape1.shape_type == 5 and shape2.shape_type == 5:
@@ -295,24 +393,92 @@ def compare_pptx_files(file1_path, file2_path, **options):
if shape1.shape_type == MSO_SHAPE_TYPE.TABLE:
table1 = shape1.table
table2 = shape2.table
if enable_debug:
debug_logger.debug(f" Shape {shape_idx} - Comparing TABLE with {len(table1.rows)} rows and {len(table1.columns)} columns")
for row_idx in range(len(table1.rows)):
for col_idx in range(len(table1.columns)):
cell1 = table1.cell(row_idx, col_idx)
cell2 = table2.cell(row_idx, col_idx)
for para1, para2 in zip(cell1.text_frame.paragraphs, cell2.text_frame.paragraphs):
for run1, run2 in zip(para1.runs, para2.runs):
if run1.font.color.rgb != run2.font.color.rgb:
return 0
for para_idx, (para1, para2) in enumerate(zip(cell1.text_frame.paragraphs, cell2.text_frame.paragraphs)):
for run_idx, (run1, run2) in enumerate(zip(para1.runs, para2.runs)):
# Check font color
if hasattr(run1.font.color, "rgb") and hasattr(run2.font.color, "rgb"):
if run1.font.color.rgb != run2.font.color.rgb and examine_color_rgb:
if enable_debug:
debug_logger.debug(f" MISMATCH: Slide {slide_idx}, Shape {shape_idx} (TABLE) - Cell [{row_idx},{col_idx}], Para {para_idx}, Run {run_idx} - Font color differs:")
debug_logger.debug(f" Color1: {run1.font.color.rgb} vs Color2: {run2.font.color.rgb}")
debug_logger.debug(f" Cell text: '{cell1.text.strip()}' vs '{cell2.text.strip()}'")
debug_logger.debug(f" Run text: '{run1.text}' vs '{run2.text}'")
return 0
# Check font bold
if run1.font.bold != run2.font.bold:
if not ((run1.font.bold is None or run1.font.bold is False) and
(run2.font.bold is None or run2.font.bold is False)):
if enable_debug:
debug_logger.debug(f" MISMATCH: Slide {slide_idx}, Shape {shape_idx} (TABLE) - Cell [{row_idx},{col_idx}], Para {para_idx}, Run {run_idx} - Font bold differs:")
debug_logger.debug(f" Bold1: {run1.font.bold} vs Bold2: {run2.font.bold}")
debug_logger.debug(f" Run text: '{run1.text}' vs '{run2.text}'")
return 0
# Check font italic
if run1.font.italic != run2.font.italic:
if not ((run1.font.italic is None or run1.font.italic is False) and
(run2.font.italic is None or run2.font.italic is False)):
if enable_debug:
debug_logger.debug(f" MISMATCH: Slide {slide_idx}, Shape {shape_idx} (TABLE) - Cell [{row_idx},{col_idx}], Para {para_idx}, Run {run_idx} - Font italic differs:")
debug_logger.debug(f" Italic1: {run1.font.italic} vs Italic2: {run2.font.italic}")
debug_logger.debug(f" Run text: '{run1.text}' vs '{run2.text}'")
return 0
# Check font underline
if run1.font.underline != run2.font.underline:
if run1.font.underline is not None and run2.font.underline is not None:
if enable_debug:
debug_logger.debug(f" MISMATCH: Slide {slide_idx}, Shape {shape_idx} (TABLE) - Cell [{row_idx},{col_idx}], Para {para_idx}, Run {run_idx} - Font underline differs:")
debug_logger.debug(f" Underline1: {run1.font.underline} vs Underline2: {run2.font.underline}")
debug_logger.debug(f" Run text: '{run1.text}' vs '{run2.text}'")
return 0
if (run1.font.underline is None and run2.font.underline is True) or (run1.font.underline is True and run2.font.underline is None):
if enable_debug:
debug_logger.debug(f" MISMATCH: Slide {slide_idx}, Shape {shape_idx} (TABLE) - Cell [{row_idx},{col_idx}], Para {para_idx}, Run {run_idx} - Font underline differs (None vs True):")
debug_logger.debug(f" Underline1: {run1.font.underline} vs Underline2: {run2.font.underline}")
debug_logger.debug(f" Run text: '{run1.text}' vs '{run2.text}'")
return 0
if hasattr(shape1, "text") and hasattr(shape2, "text"):
if shape1.text.strip() != shape2.text.strip() and examine_text:
return 0
# check if the paragraphs are the same
para_idx = 0
for para1, para2 in zip(shape1.text_frame.paragraphs, shape2.text_frame.paragraphs):
if para1.alignment != para2.alignment and examine_alignment:
return 0
para_idx += 1
# Handle alignment comparison - treat None and LEFT (1) as equivalent
if examine_alignment:
from pptx.enum.text import PP_ALIGN
align1 = para1.alignment
align2 = para2.alignment
if enable_debug:
align1_name = "None" if align1 is None else getattr(align1, 'name', str(align1))
align2_name = "None" if align2 is None else getattr(align2, 'name', str(align2))
debug_logger.debug(f" Slide {slide_idx}, Shape {shape_idx}, Para {para_idx} - Alignment: '{align1_name}' vs '{align2_name}'")
debug_logger.debug(f" Slide {slide_idx}, Shape {shape_idx}, Para {para_idx} - Text: '{para1.text}' vs '{para2.text}'")
# Convert None to LEFT for comparison since None means default left alignment
if align1 is None:
align1 = PP_ALIGN.LEFT # LEFT alignment
if align2 is None:
align2 = PP_ALIGN.LEFT # LEFT alignment
if align1 != align2:
if enable_debug:
align1_final = getattr(align1, 'name', str(align1))
align2_final = getattr(align2, 'name', str(align2))
debug_logger.debug(f" MISMATCH: Slide {slide_idx}, Shape {shape_idx}, Para {para_idx} - Alignment differs: '{align1_final}' vs '{align2_final}'")
return 0
# check if the runs are the same
if para1.text != para2.text and examine_text:
@@ -325,26 +491,59 @@ def compare_pptx_files(file1_path, file2_path, **options):
# check if the font properties are the same
if run1.font.name != run2.font.name and examine_font_name:
if enable_debug:
debug_logger.debug(f" MISMATCH: Slide {slide_idx}, Shape {shape_idx}, Para {para_idx} - Font name differs:")
debug_logger.debug(f" Name1: '{run1.font.name}' vs Name2: '{run2.font.name}'")
debug_logger.debug(f" Text: '{run1.text}' vs '{run2.text}'")
return 0
if run1.font.size != run2.font.size and examine_font_size:
if enable_debug:
debug_logger.debug(f" MISMATCH: Slide {slide_idx}, Shape {shape_idx}, Para {para_idx} - Font size differs:")
debug_logger.debug(f" Size1: {run1.font.size} vs Size2: {run2.font.size}")
debug_logger.debug(f" Text: '{run1.text}' vs '{run2.text}'")
return 0
if run1.font.bold != run2.font.bold and examine_font_bold:
return 0
# Special handling for None vs False - both mean "not bold"
if not ((run1.font.bold is None or run1.font.bold is False) and
(run2.font.bold is None or run2.font.bold is False)):
if enable_debug:
debug_logger.debug(f" MISMATCH: Slide {slide_idx}, Shape {shape_idx}, Para {para_idx} - Font bold differs:")
debug_logger.debug(f" Bold1: {run1.font.bold} vs Bold2: {run2.font.bold}")
debug_logger.debug(f" Text: '{run1.text}' vs '{run2.text}'")
return 0
if run1.font.italic != run2.font.italic and examine_font_italic:
if run1.font.italic is not None and run2.font.italic is not None:
# Special handling for None vs False - both mean "not italic"
if not ((run1.font.italic is None or run1.font.italic is False) and
(run2.font.italic is None or run2.font.italic is False)):
if enable_debug:
debug_logger.debug(f" MISMATCH: Slide {slide_idx}, Shape {shape_idx}, Para {para_idx} - Font italic differs:")
debug_logger.debug(f" Italic1: {run1.font.italic} vs Italic2: {run2.font.italic}")
debug_logger.debug(f" Text: '{run1.text}' vs '{run2.text}'")
return 0
if hasattr(run1.font.color, "rgb") and hasattr(run2.font.color, "rgb"):
if run1.font.color.rgb != run2.font.color.rgb and examine_color_rgb:
if enable_debug:
debug_logger.debug(f" MISMATCH: Slide {slide_idx}, Shape {shape_idx}, Para {para_idx} - Font color differs:")
debug_logger.debug(f" Color1: {run1.font.color.rgb} vs Color2: {run2.font.color.rgb}")
debug_logger.debug(f" Text: '{run1.text}' vs '{run2.text}'")
return 0
if run1.font.underline != run2.font.underline and examine_font_underline:
if run1.font.underline is not None and run2.font.underline is not None:
if enable_debug:
debug_logger.debug(f" MISMATCH: Slide {slide_idx}, Shape {shape_idx}, Para {para_idx} - Font underline differs:")
debug_logger.debug(f" Underline1: {run1.font.underline} vs Underline2: {run2.font.underline}")
debug_logger.debug(f" Text: '{run1.text}' vs '{run2.text}'")
return 0
if (run1.font.underline is None and run2.font.underline is True) or (run1.font.underline is True and run2.font.underline is None):
if enable_debug:
debug_logger.debug(f" MISMATCH: Slide {slide_idx}, Shape {shape_idx}, Para {para_idx} - Font underline differs (None vs True):")
debug_logger.debug(f" Underline1: {run1.font.underline} vs Underline2: {run2.font.underline}")
debug_logger.debug(f" Text: '{run1.text}' vs '{run2.text}'")
return 0
if run1.font._element.attrib.get('strike', 'noStrike') != run2.font._element.attrib.get(
@@ -382,16 +581,89 @@ def compare_pptx_files(file1_path, file2_path, **options):
return bullets
if examine_bullets:
bullets1 = _extract_bullets(run1.part.blob.decode('utf-8'))
bullets2 = _extract_bullets(run2.part.blob.decode('utf-8'))
def _compare_bullets_with_tolerance(bullets1, bullets2):
"""Compare bullets with tolerance for minor differences"""
if len(bullets1) != len(bullets2):
return False
# Compare only non-empty bullets
if bullets1 != bullets2:
return 0
for (lvl1, char1, text1, color1), (lvl2, char2, text2, color2) in zip(bullets1, bullets2):
# Compare text (most important)
if text1 != text2:
return False
# Compare bullet character
if char1 != char2:
return False
# Compare level with tolerance (None and '0' are equivalent)
normalized_lvl1 = '0' if lvl1 is None else lvl1
normalized_lvl2 = '0' if lvl2 is None else lvl2
if normalized_lvl1 != normalized_lvl2:
return False
# Color comparison is more lenient - we don't fail on color differences
# since they might be due to theme or formatting differences
# if color1 != color2:
# return False
return True
if examine_bullets:
try:
bullets1 = _extract_bullets(run1.part.blob.decode('utf-8'))
bullets2 = _extract_bullets(run2.part.blob.decode('utf-8'))
# Compare bullets with tolerance for minor differences
if not _compare_bullets_with_tolerance(bullets1, bullets2):
return 0
except:
# If bullet extraction fails, skip bullet comparison
pass
# fixme: Actually there are more properties to be compared, we can add them later via parsing the xml data
# Additional check: compare all text shapes including those in GROUPs
if examine_alignment and len(text_shapes1) == len(text_shapes2):
for idx, (tshape1, tshape2) in enumerate(zip(text_shapes1, text_shapes2)):
if enable_debug:
debug_logger.debug(f" Additional text shape check {idx+1}: '{tshape1.text.strip()[:30]}' vs '{tshape2.text.strip()[:30]}'")
# Compare text content
if tshape1.text.strip() != tshape2.text.strip() and examine_text:
if enable_debug:
debug_logger.debug(f" MISMATCH: Text differs - '{tshape1.text.strip()}' vs '{tshape2.text.strip()}'")
return 0
# Compare alignment of each paragraph
for para_idx, (para1, para2) in enumerate(zip(tshape1.text_frame.paragraphs, tshape2.text_frame.paragraphs)):
from pptx.enum.text import PP_ALIGN
align1 = para1.alignment
align2 = para2.alignment
if enable_debug:
align1_name = "None" if align1 is None else getattr(align1, 'name', str(align1))
align2_name = "None" if align2 is None else getattr(align2, 'name', str(align2))
debug_logger.debug(f" Para {para_idx+1}: Alignment '{align1_name}' vs '{align2_name}'")
# Convert None to LEFT for comparison
if align1 is None:
align1 = PP_ALIGN.LEFT
if align2 is None:
align2 = PP_ALIGN.LEFT
if align1 != align2:
if enable_debug:
align1_final = getattr(align1, 'name', str(align1))
align2_final = getattr(align2, 'name', str(align2))
debug_logger.debug(f" MISMATCH: Alignment differs - '{align1_final}' vs '{align2_final}'")
return 0
elif len(text_shapes1) != len(text_shapes2):
if enable_debug:
debug_logger.debug(f"MISMATCH: Different number of text shapes - {len(text_shapes1)} vs {len(text_shapes2)}")
return 0
if enable_debug:
debug_logger.debug(f"=== COMPARISON SUCCESSFUL - Files match ===")
return 1