Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

perf: mongodb数据支持自动清理 #2783 #2960

Open
wants to merge 1 commit into
base: master
Choose a base branch
from
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions src/backend/job-backup/service-job-backup/build.gradle
Original file line number Diff line number Diff line change
Expand Up @@ -38,6 +38,7 @@ dependencies {
implementation("org.springframework.boot:spring-boot-starter-web")
implementation("org.springframework.boot:spring-boot-starter-jdbc")
implementation("org.springframework.boot:spring-boot-starter-jooq")
implementation("org.springframework.boot:spring-boot-starter-data-mongodb")
implementation "org.apache.commons:commons-collections4"
implementation "commons-io:commons-io"
implementation "ch.qos.logback:logback-core"
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,264 @@
/*
* Tencent is pleased to support the open source community by making BK-JOB蓝鲸智云作业平台 available.
*
* Copyright (C) 2021 THL A29 Limited, a Tencent company. All rights reserved.
*
* BK-JOB蓝鲸智云作业平台 is licensed under the MIT License.
*
* License for BK-JOB蓝鲸智云作业平台:
* --------------------------------------------------------------------
* Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated
* documentation files (the "Software"), to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and
* to permit persons to whom the Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in all copies or substantial portions of
* the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO
* THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
* AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF
* CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
* IN THE SOFTWARE.
*/

package com.tencent.bk.job.backup.archive;

import com.tencent.bk.job.backup.config.ArchiveDBProperties;
import com.tencent.bk.job.backup.constant.ArchiveModeEnum;
import com.tencent.bk.job.backup.constant.MongoDBLogTypeEnum;
import com.tencent.bk.job.backup.metrics.ArchiveErrorTaskCounter;
import com.tencent.bk.job.backup.model.dto.ArchiveSummary;
import lombok.extern.slf4j.Slf4j;
import org.slf4j.helpers.MessageFormatter;
import org.springframework.data.mongodb.core.MongoTemplate;

import java.time.LocalDate;
import java.time.format.DateTimeFormatter;
import java.util.List;
import java.util.Set;
import java.util.concurrent.CountDownLatch;
import java.util.concurrent.atomic.AtomicInteger;
import java.util.stream.Collectors;

/**
* mongodb归档基础实现
*/
@Slf4j
public abstract class AbstractMongoDBArchivist {
protected ArchiveDBProperties archiveDBProperties;
protected CountDownLatch countDownLatch;
protected ArchiveSummary archiveSummary;
protected boolean isAcquireLock;
protected ArchiveTaskLock archiveTaskLock;
protected MongoDBLogTypeEnum mongoDBLogTypeEnum;
protected AtomicInteger deleteCounter;
protected AtomicInteger backupCounter;
protected AtomicInteger failCounter;

protected final ArchiveErrorTaskCounter archiveErrorTaskCounter;
protected final MongoTemplate mongoTemplate;
protected final String SCRIPT_LOG_PREFIX = "job_log_script";
protected final String FILE_LOG_PREFIX = "job_log_file";
protected final String COLLECTION_NAME_SEPARATOR = "_";
protected final String COLLECTION_NAME_DATE_FORMATTER = "yyyy_MM_dd";

public AbstractMongoDBArchivist(MongoTemplate mongoTemplate,
ArchiveDBProperties archiveDBProperties,
ArchiveTaskLock archiveTaskLock,
CountDownLatch countDownLatch,
ArchiveErrorTaskCounter archiveErrorTaskCounter,
MongoDBLogTypeEnum mongoDBLogTypeEnum) {
this.mongoTemplate = mongoTemplate;
this.archiveDBProperties = archiveDBProperties;
this.countDownLatch = countDownLatch;
this.archiveSummary = new ArchiveSummary();
this.archiveTaskLock = archiveTaskLock;
this.archiveErrorTaskCounter = archiveErrorTaskCounter;
this.mongoDBLogTypeEnum = mongoDBLogTypeEnum;
deleteCounter = new AtomicInteger(0);
backupCounter = new AtomicInteger(0);
failCounter = new AtomicInteger(0);
}

public void archive() {
if (!archiveDBProperties.isEnabled()) {
archiveSummary.setSkip(true);
log.info("[{}] Archive is disabled, skip archive", getLogTypeStr());
return;
}
archiveSummary.setEnabled(true);
log.info("[{}] Start archive, keep days:{}", getLogTypeStr(), archiveDBProperties.getKeepDays());
List<String> targetCollectionNames = listArchiveCollectionName(archiveDBProperties.getKeepDays());
log.debug("[{}] archive, target collections:{}", getLogTypeStr(), targetCollectionNames);

long startTime = System.currentTimeMillis();
boolean success = true;
for (String collectionName : targetCollectionNames) {
if (!acquireLock(collectionName)) {
continue;
}
backupAndDelete(collectionName, isBackupEnable());
}

if (targetCollectionNames.size() > 0
&& targetCollectionNames.size() == failCounter.get()) {
archiveErrorTaskCounter.increment();
success = false;
}

long archiveCost = System.currentTimeMillis() - startTime;
log.info("[{}] Archive finished, totalCollectionSize:{},deleteCollectionSize:{}, backupCollectionSize:{}," +
"cost: {}ms",
getLogTypeStr(),
targetCollectionNames.size(),
deleteCounter.get(),
backupCounter.get(),
archiveCost
);
setArchiveSummary(getLogTypeStr(),
archiveDBProperties.getMode(),
archiveCost,
deleteCounter.get(),
backupCounter.get(),
targetCollectionNames.size(),
success
);
storeArchiveSummary();

countDownLatch.countDown();
}

private void backupAndDelete(String collectionName, boolean backupEnabled) {
long startTime = System.currentTimeMillis();
boolean success = true;
try {
if (backupEnabled) {
backupRecords(collectionName);
}
deleteRecord(collectionName);
} catch (Exception e) {
success = false;
failCounter.incrementAndGet();
String msg = MessageFormatter.format(
"[{}] archive error, collectionName:{}",
getLogTypeStr(),
collectionName
).getMessage();
log.error(msg, e);
} finally {
long archiveCost = System.currentTimeMillis() - startTime;
log.info(
"Archive {} finished, result: {}, cost: {}ms",
collectionName,
success ? "success" : "fail",
archiveCost
);
if (this.isAcquireLock) {
archiveTaskLock.unlock(collectionName);
}
}
}

protected void backupRecords(String collectionName) {
// 暂不支持备份
}

private void deleteRecord(String collectionName) {
mongoTemplate.dropCollection(collectionName);
deleteCounter.incrementAndGet();
}

/**
* 获取满足归档条件的集合名称
*/
private List<String> listArchiveCollectionName(int keepDays) {
Set<String> allCollectionNames = mongoTemplate.getCollectionNames();
List<String> targetCollectionNames = filterCollectionNames(allCollectionNames, keepDays);
return targetCollectionNames;
}

/**
* 从所有集合名称中过滤出要归档的集合名称,返回集合名称列表,并按日期排序
*/
private List<String> filterCollectionNames(Set<String> collectionNames, int keepDays) {
LocalDate endDate = LocalDate.now().minusDays(keepDays);
DateTimeFormatter formatter = DateTimeFormatter.ofPattern(COLLECTION_NAME_DATE_FORMATTER);
String endDateString = endDate.format(formatter);
log.info("[{}] archive,the deadline is {}", getLogTypeStr(), endDateString);
return collectionNames.stream()
.filter(collectionName -> {
if (collectionName.startsWith(getCollectionNamePrefix())) {
String dateString = collectionName.substring(getCollectionNamePrefix().length());
LocalDate collectionDate = LocalDate.parse(dateString, formatter);
return collectionDate.isBefore(endDate);
}
return false;
})
.sorted((collectionName1, collectionName2) -> {
LocalDate date1 = extractDateFromCollectionName(collectionName1);
LocalDate date2 = extractDateFromCollectionName(collectionName2);
return date1.compareTo(date2);
})
.collect(Collectors.toList());
}

private String getCollectionNamePrefix() {
Integer type = mongoDBLogTypeEnum.getValue();
String prefix = "";
if (MongoDBLogTypeEnum.SCRIPT.getValue().equals(type)) {
prefix = SCRIPT_LOG_PREFIX + COLLECTION_NAME_SEPARATOR;
} else if (MongoDBLogTypeEnum.FILE.getValue().equals(type)) {
prefix = FILE_LOG_PREFIX + COLLECTION_NAME_SEPARATOR;
}
return prefix;
}

private LocalDate extractDateFromCollectionName(String collectionName) {
String dateString = collectionName.substring(getCollectionNamePrefix().length());
return LocalDate.parse(dateString, DateTimeFormatter.ofPattern(COLLECTION_NAME_DATE_FORMATTER));
}

private String getLogTypeStr() {
Integer type = mongoDBLogTypeEnum.getValue();
String str = "script_log";
if (MongoDBLogTypeEnum.FILE.getValue().equals(type)) {
str = "file_log";
}
return str;
}

private boolean acquireLock(String collectionName) {
this.isAcquireLock = archiveTaskLock.lock(collectionName);
if (!isAcquireLock) {
log.info("[{}] Acquire lock fail", collectionName);
}
return isAcquireLock;
}

private boolean isBackupEnable() {
return archiveDBProperties.isEnabled()
&& ArchiveModeEnum.BACKUP_THEN_DELETE == ArchiveModeEnum.valOf(archiveDBProperties.getMode());
}

private void setArchiveSummary(String tableName,
String archiveMode,
long archiveCost,
long deleteCollectionSize,
long backupCollectionSize,
long totalCollectionSize,
boolean success) {
archiveSummary.setArchiveCost(archiveCost);
archiveSummary.setArchiveMode(archiveMode);
archiveSummary.setSuccess(success);
archiveSummary.setTableName(tableName);
archiveSummary.setDeleteCollectionSize(deleteCollectionSize);
archiveSummary.setBackupCollectionSize(backupCollectionSize);
archiveSummary.setTotalCollectionSize(totalCollectionSize);
}

private void storeArchiveSummary() {
ArchiveSummaryHolder.getInstance().addArchiveSummary(this.archiveSummary);
}

}
Original file line number Diff line number Diff line change
Expand Up @@ -30,6 +30,7 @@
import com.tencent.bk.job.backup.archive.impl.GseScriptAgentTaskArchivist;
import com.tencent.bk.job.backup.archive.impl.GseScriptExecuteObjTaskArchivist;
import com.tencent.bk.job.backup.archive.impl.GseTaskArchivist;
import com.tencent.bk.job.backup.archive.impl.MongoDBScriptLogArchivist;
import com.tencent.bk.job.backup.archive.impl.OperationLogArchivist;
import com.tencent.bk.job.backup.archive.impl.RollingConfigArchivist;
import com.tencent.bk.job.backup.archive.impl.StepInstanceArchivist;
Expand All @@ -42,6 +43,7 @@
import com.tencent.bk.job.backup.archive.impl.TaskInstanceHostArchivist;
import com.tencent.bk.job.backup.archive.impl.TaskInstanceVariableArchivist;
import com.tencent.bk.job.backup.config.ArchiveDBProperties;
import com.tencent.bk.job.backup.constant.MongoDBLogTypeEnum;
import com.tencent.bk.job.backup.dao.ExecuteArchiveDAO;
import com.tencent.bk.job.backup.dao.ExecuteRecordDAO;
import com.tencent.bk.job.backup.dao.impl.FileSourceTaskLogRecordDAO;
Expand All @@ -68,6 +70,7 @@
import org.joda.time.DateTime;
import org.slf4j.helpers.MessageFormatter;
import org.springframework.context.SmartLifecycle;
import org.springframework.data.mongodb.core.MongoTemplate;
import org.springframework.scheduling.annotation.Scheduled;

import java.util.concurrent.CountDownLatch;
Expand Down Expand Up @@ -99,6 +102,7 @@ public class JobExecuteArchiveManage implements SmartLifecycle {
private final ExecuteArchiveDAO executeArchiveDAO;
private final ArchiveTaskLock archiveTaskLock;
private final ArchiveErrorTaskCounter archiveErrorTaskCounter;
private final MongoTemplate mongoTemplate;


/**
Expand Down Expand Up @@ -128,7 +132,8 @@ public JobExecuteArchiveManage(TaskInstanceRecordDAO taskInstanceRecordDAO,
ArchiveDBProperties archiveDBProperties,
ExecutorService archiveExecutor,
ArchiveTaskLock archiveTaskLock,
ArchiveErrorTaskCounter archiveErrorTaskCounter) {
ArchiveErrorTaskCounter archiveErrorTaskCounter,
MongoTemplate mongoTemplate) {
log.info("Init JobExecuteArchiveManage! archiveConfig: {}", archiveDBProperties);
this.archiveDBProperties = archiveDBProperties;
this.archiveProgressService = archiveProgressService;
Expand All @@ -153,6 +158,7 @@ public JobExecuteArchiveManage(TaskInstanceRecordDAO taskInstanceRecordDAO,
this.executeArchiveDAO = executeArchiveDAO;
this.archiveTaskLock = archiveTaskLock;
this.archiveErrorTaskCounter = archiveErrorTaskCounter;
this.mongoTemplate = mongoTemplate;
}

@Scheduled(cron = "${job.backup.archive.execute.cron:0 0 4 * * *}")
Expand Down Expand Up @@ -271,7 +277,7 @@ private long getLastArchiveId(ExecuteRecordDAO<?> executeRecordDAO) {

private void archive(long maxNeedArchiveTaskInstanceId, long maxNeedArchiveStepInstanceId)
throws InterruptedException {
CountDownLatch countDownLatch = new CountDownLatch(17);
CountDownLatch countDownLatch = new CountDownLatch(19);
log.info("Submitting archive task...");

// task_instance
Expand Down Expand Up @@ -308,6 +314,10 @@ private void archive(long maxNeedArchiveTaskInstanceId, long maxNeedArchiveStepI
addGseScriptExecuteObjTaskArchiveTask(maxNeedArchiveTaskInstanceId, countDownLatch);
// gse_file_execute_obj_task
addGseFileExecuteObjTaskArchiveTask(maxNeedArchiveTaskInstanceId, countDownLatch);
// mongodb job_log_script_task
addMongoDBScriptLogArchiveTask(countDownLatch);
// mongodb job_log_file_task
addMongoDBFileLogArchiveTask(countDownLatch);

log.info("Archive task submitted. Waiting for complete...");
countDownLatch.await();
Expand Down Expand Up @@ -568,5 +578,29 @@ private void addTaskInstanceHostArchiveTask(Long maxNeedArchiveTaskInstanceId,
archiveErrorTaskCounter
).archive());
}

private void addMongoDBScriptLogArchiveTask(CountDownLatch countDownLatch) {
archiveExecutor.execute(() ->
new MongoDBScriptLogArchivist(
mongoTemplate,
archiveDBProperties,
archiveTaskLock,
countDownLatch,
archiveErrorTaskCounter,
MongoDBLogTypeEnum.SCRIPT
).archive());
}

private void addMongoDBFileLogArchiveTask(CountDownLatch countDownLatch) {
archiveExecutor.execute(() ->
new MongoDBScriptLogArchivist(
mongoTemplate,
archiveDBProperties,
archiveTaskLock,
countDownLatch,
archiveErrorTaskCounter,
MongoDBLogTypeEnum.FILE
).archive());
}
}
}
Loading
Loading