package de.micromata.borgbutler.cache;
|
|
import de.micromata.borgbutler.cache.memory.MemoryCache;
|
import de.micromata.borgbutler.cache.memory.MemoryCacheObject;
|
import de.micromata.borgbutler.config.BorgRepoConfig;
|
import de.micromata.borgbutler.data.Archive;
|
import de.micromata.borgbutler.data.FileSystemFilter;
|
import de.micromata.borgbutler.data.Repository;
|
import de.micromata.borgbutler.json.borg.BorgFilesystemItem;
|
import de.micromata.borgbutler.utils.ReplaceUtils;
|
import org.apache.commons.collections4.CollectionUtils;
|
import org.apache.commons.compress.compressors.gzip.GzipCompressorInputStream;
|
import org.apache.commons.compress.compressors.gzip.GzipCompressorOutputStream;
|
import org.apache.commons.io.FileUtils;
|
import org.apache.commons.io.output.ByteArrayOutputStream;
|
import org.apache.commons.lang3.StringUtils;
|
import org.nustaq.serialization.FSTConfiguration;
|
import org.nustaq.serialization.FSTObjectInput;
|
import org.nustaq.serialization.FSTObjectOutput;
|
import org.slf4j.Logger;
|
import org.slf4j.LoggerFactory;
|
|
import java.io.*;
|
import java.nio.file.Files;
|
import java.nio.file.attribute.BasicFileAttributes;
|
import java.nio.file.attribute.FileTime;
|
import java.util.*;
|
|
/**
|
* Cache for storing complete file lists of archives as gzipped files (using Java standard serialization for
|
* fastest access).
|
* <br>
|
* A file list (archive content) with over million file system items is over 100MB large (uncompressed).
|
* The compression is also useful for faster reading from the filesystem.
|
*/
|
class ArchiveFilelistCache {
|
private static Logger log = LoggerFactory.getLogger(ArchiveFilelistCache.class);
|
private static final String CACHE_ARCHIVE_LISTS_BASENAME = "archive-content-";
|
private static final String CACHE_FILE_GZIP_EXTENSION = ".gz";
|
private static final int MAX_SIZE_MEMORY_CACHE = 30 * 1024 * 1024; // 50 MB
|
private File cacheDir;
|
private int cacheArchiveContentMaxDiscSizeMB;
|
private long FILES_EXPIRE_TIME = 7 * 24 * 3660 * 1000; // Expires after 7 days.
|
// For avoiding concurrent writing of same files (e. g. after the user has pressed a button twice).
|
private Set<File> savingFiles = new HashSet<>();
|
private MemoryCache<Archive, RecentEntry> recents = new MemoryCache<>(MAX_SIZE_MEMORY_CACHE);
|
final FSTConfiguration conf = FSTConfiguration.createDefaultConfiguration();
|
|
ArchiveFilelistCache(File cacheDir, int cacheArchiveContentMaxDiscSizeMB) {
|
this.cacheDir = cacheDir;
|
this.cacheArchiveContentMaxDiscSizeMB = cacheArchiveContentMaxDiscSizeMB;
|
conf.registerClass(Integer.class, BorgFilesystemItem.class);
|
conf.setShareReferences(false);
|
}
|
|
public void save(BorgRepoConfig repoConfig, Archive archive, List<BorgFilesystemItem> filesystemItems) {
|
if (CollectionUtils.isEmpty(filesystemItems)) {
|
return;
|
}
|
File file = getFile(repoConfig, archive);
|
try {
|
synchronized (savingFiles) {
|
if (savingFiles.contains(file)) {
|
// File will already be written. This occurs if the user pressed a button twice.
|
log.info("Don't write the archive content twice.");
|
return;
|
}
|
savingFiles.add(file);
|
Collections.sort(filesystemItems); // Sort by path.
|
}
|
log.info("Saving archive content as file list: " + file.getAbsolutePath());
|
boolean ok = false;
|
|
int fileNumber = -1;
|
try (FSTObjectOutput outputStream
|
= new FSTObjectOutput(new BufferedOutputStream(new GzipCompressorOutputStream(new FileOutputStream(file))), conf)) {
|
outputStream.writeObject(filesystemItems.size(), Integer.class);
|
Iterator<BorgFilesystemItem> it = filesystemItems.iterator();
|
while (it.hasNext()) {
|
BorgFilesystemItem item = it.next();
|
item.setFileNumber(++fileNumber);
|
outputStream.writeObject(item, BorgFilesystemItem.class);
|
}
|
outputStream.writeObject("EOF");
|
ok = true;
|
} catch (IOException ex) {
|
log.error("Error while writing file list '" + file.getAbsolutePath() + "': " + ex.getMessage(), ex);
|
}
|
if (ok) {
|
// Storing current read gz archive in memory cache (recents):
|
ByteArrayOutputStream baos = new ByteArrayOutputStream();
|
try {
|
FileUtils.copyFile(file, baos);
|
recents.add(new RecentEntry(archive, baos.toByteArray()));
|
} catch (IOException ex) {
|
log.error("Error while writing gz archive to memory cache: " + ex.getMessage(), ex);
|
}
|
}
|
} finally {
|
synchronized (savingFiles) {
|
savingFiles.remove(file);
|
}
|
}
|
log.info("Saving done.");
|
}
|
|
/**
|
* @param repository
|
* @param archive
|
* @return true, if the content of the archive is already cached, otherwise false.
|
*/
|
public boolean contains(Repository repository, Archive archive) {
|
File file = getFile(repository, archive);
|
return file.exists();
|
}
|
|
/**
|
* Calls {@link #load(BorgRepoConfig, Archive, FileSystemFilter)} with filter null.
|
*
|
* @param repoConfig
|
* @param archive
|
* @return
|
*/
|
public List<BorgFilesystemItem> load(BorgRepoConfig repoConfig, Archive archive) {
|
return load(repoConfig, archive, null);
|
}
|
|
|
/**
|
* Will load and touch the archive file if exist. The file will be touched (last modified time will be set to now)
|
* for pruning oldest cache files. The last modified time will be the time of the last usage.
|
*
|
* @param repoConfig
|
* @param archive
|
* @param filter If given, only file items matching this filter are returned.
|
* @return
|
*/
|
public List<BorgFilesystemItem> load(BorgRepoConfig repoConfig, Archive archive, FileSystemFilter filter) {
|
File file = getFile(repoConfig, archive);
|
if (!file.exists()) {
|
return null;
|
}
|
return load(file, archive, filter);
|
}
|
|
/**
|
* @param file
|
* @param filter If given, only file items matching this filter are returned.
|
* @return
|
*/
|
public List<BorgFilesystemItem> load(File file, FileSystemFilter filter) {
|
return load(file, null, filter);
|
}
|
|
/**
|
* @param file
|
* @param archive Only for storing file system items as recent (may-be null)
|
* @param filter If given, only file items matching this filter are returned.
|
* @return
|
*/
|
public List<BorgFilesystemItem> load(File file, Archive archive, FileSystemFilter filter) {
|
if (!file.exists()) {
|
log.error("File '" + file.getAbsolutePath() + "' doesn't exist. Can't get archive content files.");
|
return null;
|
}
|
log.info("Loading archive content as file list from: " + file.getAbsolutePath());
|
try {
|
// Set last modified time of file:
|
Files.setAttribute(file.toPath(), "lastModifiedTime", FileTime.fromMillis(System.currentTimeMillis()));
|
} catch (IOException ex) {
|
log.error("Can't set lastModifiedTime on file '" + file.getAbsolutePath() + "'. Pruning old cache files may not work.");
|
}
|
RecentEntry recentEntry = recents.getRecent(archive);
|
byte[] bytes = null;
|
if (recentEntry != null) {
|
bytes = recentEntry.serializedGz;
|
} else {
|
ByteArrayOutputStream baos = new ByteArrayOutputStream();
|
try {
|
FileUtils.copyFile(file, baos);
|
bytes = baos.toByteArray();
|
recents.add(new RecentEntry(archive, bytes));
|
} catch (IOException ex) {
|
log.error("Error while restoring file: " + file.getAbsolutePath() + ": " + ex.getMessage(), ex);
|
return null;
|
}
|
}
|
List<BorgFilesystemItem> list = new ArrayList<>();
|
try (FSTObjectInput inputStream = new FSTObjectInput(new BufferedInputStream(new GzipCompressorInputStream(new ByteArrayInputStream(bytes))), conf)) {
|
int size = (Integer) inputStream.readObject(Integer.class);
|
for (int i = 0; i < size; i++) {
|
BorgFilesystemItem item = (BorgFilesystemItem) inputStream.readObject(BorgFilesystemItem.class);
|
if (filter == null || filter.matches(item)) {
|
list.add(item);
|
if (filter != null && filter.isFinished()) break;
|
}
|
}
|
} catch (Exception ex) {
|
log.error("Error while reading file list '" + file.getAbsolutePath() + "': " + ex.getMessage(), ex);
|
}
|
|
log.info("Loading done.");
|
return filter(list, filter);
|
}
|
|
private List<BorgFilesystemItem> filter(List<BorgFilesystemItem> filesystemItems, FileSystemFilter filter) {
|
if (filter != null) {
|
return filter.reduce(filesystemItems);
|
}
|
return filesystemItems;
|
}
|
|
/**
|
* Deletes archive contents older than 7 days and deletes the oldest archive contents if the max cache size is
|
* exceeded. The last modified time of a file is equals to the last usage by
|
* {@link #load(BorgRepoConfig, Archive, FileSystemFilter)}.
|
*/
|
public void cleanUp() {
|
File[] files = cacheDir.listFiles();
|
long currentMillis = System.currentTimeMillis();
|
for (File file : files) {
|
try {
|
if (!file.exists() || !isCacheFile(file)) continue;
|
// Get last modified time of file:
|
FileTime time = Files.readAttributes(file.toPath(), BasicFileAttributes.class).lastModifiedTime();
|
if (currentMillis - FILES_EXPIRE_TIME > time.toMillis()) {
|
log.info("Delete expired cache file (last usage " + time + " older than 7 days): " + file.getAbsolutePath());
|
file.delete();
|
}
|
} catch (IOException ex) {
|
log.error("Can't get last modified time from cache files (ignore file '" + file.getAbsolutePath() + "'): " + ex.getMessage(), ex);
|
}
|
}
|
int sizeInMB = getCacheDiskSizeInMB(files);
|
if (sizeInMB > cacheArchiveContentMaxDiscSizeMB) {
|
log.info("Maximum size of cache files exceeded (" + sizeInMB + "MB > " + cacheArchiveContentMaxDiscSizeMB
|
+ "MB). Deleting the old ones (with the oldest usage)...");
|
} else {
|
// Nothing to clean up anymore.
|
return;
|
}
|
SortedMap<FileTime, File> sortedFiles = new TreeMap<>();
|
for (File file : files) {
|
if (!file.exists() || !isCacheFile(file)) continue;
|
try {
|
// Get last modified time of file:
|
FileTime time = Files.readAttributes(file.toPath(), BasicFileAttributes.class).lastModifiedTime();
|
sortedFiles.put(time, file);
|
} catch (IOException ex) {
|
log.error("Can't get last modified time from cache files (ignore file '" + file.getAbsolutePath() + "'): " + ex.getMessage(), ex);
|
}
|
}
|
for (Map.Entry<FileTime, File> entry : sortedFiles.entrySet()) {
|
FileTime time = entry.getKey();
|
File file = entry.getValue();
|
if (!file.exists() || !isCacheFile(file)) continue;
|
log.info("Deleting cache file (last usage " + time + "): " + file.getAbsolutePath());
|
file.delete();
|
int newSizeInMB = getCacheDiskSizeInMB(files);
|
if (newSizeInMB < cacheArchiveContentMaxDiscSizeMB) {
|
log.info("New cache size is " + newSizeInMB + "MB. (" + (sizeInMB - newSizeInMB) + "MB deleted.)");
|
break;
|
}
|
}
|
}
|
|
public int getCacheDiskSizeInMB() {
|
return getCacheDiskSizeInMB(cacheDir.listFiles());
|
}
|
|
private int getCacheDiskSizeInMB(File[] files) {
|
int sizeInMB = 0;
|
for (File file : files) {
|
if (!file.exists()) continue;
|
if (!isCacheFile(file)) continue;
|
sizeInMB += (int) (file.length() / 1048576); // In MB
|
}
|
return sizeInMB;
|
}
|
|
public void removeAllCacheFiles() {
|
File[] files = cacheDir.listFiles();
|
for (File file : files) {
|
if (isCacheFile(file)) {
|
log.info("Deleting cache file: " + file.getAbsolutePath());
|
file.delete();
|
}
|
}
|
}
|
|
File getFile(Repository repository, Archive archive) {
|
return getFile(repository.getName(), archive);
|
}
|
|
File getFile(BorgRepoConfig repoConfig, Archive archive) {
|
return getFile(repoConfig.getRepo(), archive);
|
}
|
|
private File getFile(String repo, Archive archive) {
|
return new File(cacheDir, ReplaceUtils.encodeFilename(CACHE_ARCHIVE_LISTS_BASENAME + archive.getTime()
|
+ "-" + repo + "-" + archive.getName() + CACHE_FILE_GZIP_EXTENSION,
|
true));
|
}
|
|
private boolean isCacheFile(File file) {
|
return file.getName().startsWith(CACHE_ARCHIVE_LISTS_BASENAME);
|
}
|
|
private class RecentEntry extends MemoryCacheObject<Archive> {
|
private byte[] serializedGz;
|
|
@Override
|
protected boolean matches(Archive identifier) {
|
return StringUtils.equals(this.getIdentifier().getId(), identifier.getId());
|
}
|
|
@Override
|
protected int getSize() {
|
return serializedGz != null ? serializedGz.length : 0;
|
}
|
|
private RecentEntry(Archive archive, byte[] serializedGz) {
|
super(archive);
|
this.serializedGz = serializedGz;
|
}
|
}
|
}
|