fix(s3): URL mismatch

This commit is contained in:
JamesFlare1212
2026-04-08 00:00:44 -04:00
parent 6c58eacc8f
commit 1e234624fb

View File

@@ -30,6 +30,37 @@ const STAFF_UPDATE_INTERVAL_MINS = parseInt(process.env.STAFF_UPDATE_INTERVAL_MI
const FIXED_STAFF_ACTIVITY_ID = process.env.FIXED_STAFF_ACTIVITY_ID;
const S3_IMAGE_PREFIX = (process.env.S3_PUBLIC_URL_PREFIX || 'files').replace(/\/$/, '');
/**
* Extracts S3 object key from a public URL.
* Handles both S3_PUBLIC_URL and S3_ENDPOINT formats.
*
* Examples:
* - https://cdn.example.com/bucket/files/abc123.avif → files/abc123.avif
* - https://s3.amazonaws.com/bucket/files/abc123.avif → files/abc123.avif
*
* @param url - The full S3 public URL
* @returns The object key (e.g., "files/abc123.avif") or null if parsing fails
*/
function extractObjectKeyFromUrl(url: string): string | null {
try {
const urlObj = new URL(url);
const path = urlObj.pathname;
const parts = path.replace(/^\//, '').split('/').filter(p => p);
if (parts.length >= 2) {
const key = parts.slice(1).join('/');
return key;
}
logger.warn(`Failed to extract object key from URL: ${url} - insufficient path parts`);
return null;
} catch (error) {
logger.warn(`Failed to parse URL: ${url}`, error);
return null;
}
}
// Crawler concurrency configuration
const CONCURRENT_API_CALLS = parseInt(process.env.CONCURRENT_API_CALLS || '8', 10);
const CRAWLER_REQUEST_TIMEOUT_MS = parseInt(process.env.CRAWLER_REQUEST_TIMEOUT_MS || '25000', 10);
@@ -283,23 +314,31 @@ export async function initializeOrUpdateStaffCache(forceUpdate: boolean = false)
/**
* Clean up orphaned S3 images
*
* FIX: Changed from URL-based matching to object key-based matching.
* Previous bug: URL mismatch between S3_PUBLIC_URL and S3_ENDPOINT caused
* valid images to be incorrectly marked as orphaned and deleted.
*
* New approach: Extract object keys from Redis-stored URLs and compare
* directly with S3 object keys. This is immune to URL configuration changes.
*/
export async function cleanupOrphanedS3Images(): Promise<void> {
logger.info('Starting S3 orphan image cleanup...');
const s3ObjectListPrefix = S3_IMAGE_PREFIX ? `${S3_IMAGE_PREFIX}/` : '';
try {
const referencedS3Urls = new Set<string>();
const referencedObjectKeys = new Set<string>();
const allActivityRedisKeys = await getAllActivityKeys();
const S3_ENDPOINT = process.env.S3_ENDPOINT;
let photoCount = 0;
let httpUrlCount = 0;
let base64Count = 0;
let nullCount = 0;
let addedToSet = 0;
let extractedKeyCount = 0;
let failedExtractionCount = 0;
let sampleHttpUrl = null;
let sampleBase64Url = null;
let sampleExtractedKey = null;
for (const redisKey of allActivityRedisKeys) {
const activityId = redisKey.substring(ACTIVITY_KEY_PREFIX.length);
@@ -311,8 +350,14 @@ export async function cleanupOrphanedS3Images(): Promise<void> {
if (typeof activityData.photo === 'string') {
if (activityData.photo.startsWith('http')) {
httpUrlCount++;
referencedS3Urls.add(activityData.photo);
addedToSet++;
const objectKey = extractObjectKeyFromUrl(activityData.photo);
if (objectKey) {
referencedObjectKeys.add(objectKey);
extractedKeyCount++;
if (!sampleExtractedKey) sampleExtractedKey = { url: activityData.photo, key: objectKey };
} else {
failedExtractionCount++;
}
if (!sampleHttpUrl) sampleHttpUrl = activityData.photo;
} else if (activityData.photo.startsWith('data:image')) {
base64Count++;
@@ -325,13 +370,18 @@ export async function cleanupOrphanedS3Images(): Promise<void> {
}
}
logger.info(`Found ${referencedS3Urls.size} unique S3 URLs referenced in Redis.`);
logger.info(`Found ${referencedObjectKeys.size} unique S3 object keys referenced in Redis.`);
logger.info(`Photo stats: ${httpUrlCount} HTTP URLs, ${base64Count} base64, ${nullCount} null/empty out of ${photoCount} activities with photo field.`);
logger.info(`Key extraction: ${extractedKeyCount} succeeded, ${failedExtractionCount} failed.`);
if (sampleHttpUrl) logger.debug(`Sample HTTP URL: ${sampleHttpUrl}`);
if (sampleExtractedKey) logger.debug(`Sample extracted key: URL="${sampleExtractedKey.url}" → key="${sampleExtractedKey.key}"`);
if (sampleBase64Url) logger.debug(`Sample base64 URL: ${sampleBase64Url}`);
if (httpUrlCount === 0 && base64Count > 0) {
logger.error('CRITICAL: All photos are base64 format! S3 upload may be failing or photo field not updated.');
}
if (failedExtractionCount > 0) {
logger.warn(`Failed to extract ${failedExtractionCount} object keys from URLs. These images won't be protected from deletion.`);
}
const s3ObjectKeys = await listS3Objects(s3ObjectListPrefix);
if (!s3ObjectKeys || s3ObjectKeys.length === 0) {
@@ -344,34 +394,32 @@ export async function cleanupOrphanedS3Images(): Promise<void> {
const orphanedObjectKeys: string[] = [];
let matchedCount = 0;
let notFoundCount = 0;
let firstMismatchExample = null;
let firstOrphanExample = null;
for (const objectKey of s3ObjectKeys) {
const s3Url = constructS3Url(objectKey);
if (s3Url) {
if (referencedS3Urls.has(s3Url)) {
if (referencedObjectKeys.has(objectKey)) {
matchedCount++;
} else {
notFoundCount++;
if (!firstMismatchExample) {
firstMismatchExample = { objectKey, s3Url };
}
if (!firstOrphanExample) {
firstOrphanExample = objectKey;
}
}
}
logger.info(`S3 URL matching: ${matchedCount} matched, ${notFoundCount} not found (orphaned).`);
if (notFoundCount > 0 && firstMismatchExample) {
logger.debug(`Example orphaned object: key="${firstMismatchExample.objectKey}", url="${firstMismatchExample.s3Url}"`);
logger.info(`S3 object key matching: ${matchedCount} matched, ${notFoundCount} not found (orphaned).`);
if (notFoundCount > 0 && firstOrphanExample) {
logger.debug(`Example orphaned object: key="${firstOrphanExample}"`);
}
if (matchedCount === 0 && s3ObjectKeys.length > 0 && referencedS3Urls.size > 0) {
logger.error('CRITICAL: No matches found! Sample from set: ' + [...referencedS3Urls].slice(0, 3).join(', '));
logger.error('CRITICAL: Sample from S3 keys: ' + s3ObjectKeys.slice(0, 3).map(k => constructS3Url(k)).join(', '));
if (matchedCount === 0 && s3ObjectKeys.length > 0 && referencedObjectKeys.size > 0) {
logger.error('CRITICAL: No matches found between S3 objects and Redis references!');
logger.error('CRITICAL: Sample from Redis keys: ' + [...referencedObjectKeys].slice(0, 3).join(', '));
logger.error('CRITICAL: Sample from S3 keys: ' + s3ObjectKeys.slice(0, 3).join(', '));
logger.error('CRITICAL: This indicates a serious bug in object key extraction or S3 listing.');
}
for (const objectKey of s3ObjectKeys) {
const s3Url = constructS3Url(objectKey);
if (s3Url && !referencedS3Urls.has(s3Url)) {
if (!referencedObjectKeys.has(objectKey)) {
orphanedObjectKeys.push(objectKey);
}
}