mirror of
https://github.com/xivdev/Penumbra.git
synced 2025-12-13 12:14:17 +01:00
Why is this so much work?
This commit is contained in:
parent
651c7410ac
commit
b92a3161b5
53 changed files with 3054 additions and 2936 deletions
258
Penumbra/Mods/Editor/DuplicateManager.cs
Normal file
258
Penumbra/Mods/Editor/DuplicateManager.cs
Normal file
|
|
@ -0,0 +1,258 @@
|
|||
using System;
|
||||
using System.Collections;
|
||||
using System.Collections.Generic;
|
||||
using System.IO;
|
||||
using System.Linq;
|
||||
using System.Security.Cryptography;
|
||||
using System.Threading.Tasks;
|
||||
using Penumbra.String.Classes;
|
||||
|
||||
namespace Penumbra.Mods;
|
||||
|
||||
public class DuplicateManager
|
||||
{
|
||||
private readonly Mod.Manager _modManager;
|
||||
private readonly SHA256 _hasher = SHA256.Create();
|
||||
private readonly ModFileCollection _files;
|
||||
private readonly List<(FullPath[] Paths, long Size, byte[] Hash)> _duplicates = new();
|
||||
|
||||
public DuplicateManager(ModFileCollection files, Mod.Manager modManager)
|
||||
{
|
||||
_files = files;
|
||||
_modManager = modManager;
|
||||
}
|
||||
|
||||
public IReadOnlyList<(FullPath[] Paths, long Size, byte[] Hash)> Duplicates
|
||||
=> _duplicates;
|
||||
|
||||
public long SavedSpace { get; private set; } = 0;
|
||||
public bool Finished { get; private set; } = true;
|
||||
|
||||
public void StartDuplicateCheck(IEnumerable<FileRegistry> files)
|
||||
{
|
||||
if (!Finished)
|
||||
return;
|
||||
|
||||
Finished = false;
|
||||
var filesTmp = files.OrderByDescending(f => f.FileSize).ToArray();
|
||||
Task.Run(() => CheckDuplicates(filesTmp));
|
||||
}
|
||||
|
||||
public void DeleteDuplicates(Mod mod, ISubMod option, bool useModManager)
|
||||
{
|
||||
if (!Finished || _duplicates.Count == 0)
|
||||
return;
|
||||
|
||||
foreach (var (set, _, _) in _duplicates)
|
||||
{
|
||||
if (set.Length < 2)
|
||||
continue;
|
||||
|
||||
var remaining = set[0];
|
||||
foreach (var duplicate in set.Skip(1))
|
||||
HandleDuplicate(mod, duplicate, remaining, useModManager);
|
||||
}
|
||||
|
||||
_duplicates.Clear();
|
||||
DeleteEmptyDirectories(mod.ModPath);
|
||||
_files.UpdateAll(mod, option);
|
||||
}
|
||||
|
||||
public void Clear()
|
||||
{
|
||||
Finished = true;
|
||||
SavedSpace = 0;
|
||||
}
|
||||
|
||||
private void HandleDuplicate(Mod mod, FullPath duplicate, FullPath remaining, bool useModManager)
|
||||
{
|
||||
void HandleSubMod(ISubMod subMod, int groupIdx, int optionIdx)
|
||||
{
|
||||
var changes = false;
|
||||
var dict = subMod.Files.ToDictionary(kvp => kvp.Key,
|
||||
kvp => ChangeDuplicatePath(mod, kvp.Value, duplicate, remaining, kvp.Key, ref changes));
|
||||
if (!changes)
|
||||
return;
|
||||
|
||||
if (useModManager)
|
||||
{
|
||||
_modManager.OptionSetFiles(mod, groupIdx, optionIdx, dict);
|
||||
}
|
||||
else
|
||||
{
|
||||
var sub = (Mod.SubMod)subMod;
|
||||
sub.FileData = dict;
|
||||
if (groupIdx == -1)
|
||||
mod.SaveDefaultMod();
|
||||
else
|
||||
IModGroup.Save(mod.Groups[groupIdx], mod.ModPath, groupIdx);
|
||||
}
|
||||
}
|
||||
|
||||
ModEditor.ApplyToAllOptions(mod, HandleSubMod);
|
||||
|
||||
try
|
||||
{
|
||||
File.Delete(duplicate.FullName);
|
||||
}
|
||||
catch (Exception e)
|
||||
{
|
||||
Penumbra.Log.Error($"[DeleteDuplicates] Could not delete duplicate {duplicate.FullName} of {remaining.FullName}:\n{e}");
|
||||
}
|
||||
}
|
||||
|
||||
private static FullPath ChangeDuplicatePath(Mod mod, FullPath value, FullPath from, FullPath to, Utf8GamePath key, ref bool changes)
|
||||
{
|
||||
if (!value.Equals(from))
|
||||
return value;
|
||||
|
||||
changes = true;
|
||||
Penumbra.Log.Debug($"[DeleteDuplicates] Changing {key} for {mod.Name}\n : {from}\n -> {to}");
|
||||
return to;
|
||||
}
|
||||
|
||||
private void CheckDuplicates(IReadOnlyList<FileRegistry> files)
|
||||
{
|
||||
_duplicates.Clear();
|
||||
SavedSpace = 0;
|
||||
var list = new List<FullPath>();
|
||||
var lastSize = -1L;
|
||||
foreach (var file in files)
|
||||
{
|
||||
// Skip any UI Files because deduplication causes weird crashes for those.
|
||||
if (file.SubModUsage.Any(f => f.Item2.Path.StartsWith("ui/"u8)))
|
||||
continue;
|
||||
|
||||
if (Finished)
|
||||
return;
|
||||
|
||||
if (file.FileSize == lastSize)
|
||||
{
|
||||
list.Add(file.File);
|
||||
continue;
|
||||
}
|
||||
|
||||
if (list.Count >= 2)
|
||||
CheckMultiDuplicates(list, lastSize);
|
||||
|
||||
lastSize = file.FileSize;
|
||||
|
||||
list.Clear();
|
||||
list.Add(file.File);
|
||||
}
|
||||
|
||||
if (list.Count >= 2)
|
||||
CheckMultiDuplicates(list, lastSize);
|
||||
|
||||
_duplicates.Sort((a, b) => a.Size != b.Size ? b.Size.CompareTo(a.Size) : a.Paths[0].CompareTo(b.Paths[0]));
|
||||
Finished = true;
|
||||
}
|
||||
|
||||
private void CheckMultiDuplicates(IReadOnlyList<FullPath> list, long size)
|
||||
{
|
||||
var hashes = list.Select(f => (f, ComputeHash(f))).ToList();
|
||||
while (hashes.Count > 0)
|
||||
{
|
||||
if (Finished)
|
||||
return;
|
||||
|
||||
var set = new HashSet<FullPath> { hashes[0].Item1 };
|
||||
var hash = hashes[0];
|
||||
for (var j = 1; j < hashes.Count; ++j)
|
||||
{
|
||||
if (Finished)
|
||||
return;
|
||||
|
||||
if (CompareHashes(hash.Item2, hashes[j].Item2) && CompareFilesDirectly(hashes[0].Item1, hashes[j].Item1))
|
||||
set.Add(hashes[j].Item1);
|
||||
}
|
||||
|
||||
hashes.RemoveAll(p => set.Contains(p.Item1));
|
||||
if (set.Count > 1)
|
||||
{
|
||||
_duplicates.Add((set.OrderBy(f => f.FullName.Length).ToArray(), size, hash.Item2));
|
||||
SavedSpace += (set.Count - 1) * size;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
private static unsafe bool CompareFilesDirectly(FullPath f1, FullPath f2)
|
||||
{
|
||||
if (!f1.Exists || !f2.Exists)
|
||||
return false;
|
||||
|
||||
using var s1 = File.OpenRead(f1.FullName);
|
||||
using var s2 = File.OpenRead(f2.FullName);
|
||||
var buffer1 = stackalloc byte[256];
|
||||
var buffer2 = stackalloc byte[256];
|
||||
var span1 = new Span<byte>(buffer1, 256);
|
||||
var span2 = new Span<byte>(buffer2, 256);
|
||||
|
||||
while (true)
|
||||
{
|
||||
var bytes1 = s1.Read(span1);
|
||||
var bytes2 = s2.Read(span2);
|
||||
if (bytes1 != bytes2)
|
||||
return false;
|
||||
|
||||
if (!span1[..bytes1].SequenceEqual(span2[..bytes2]))
|
||||
return false;
|
||||
|
||||
if (bytes1 < 256)
|
||||
return true;
|
||||
}
|
||||
}
|
||||
|
||||
public static bool CompareHashes(byte[] f1, byte[] f2)
|
||||
=> StructuralComparisons.StructuralEqualityComparer.Equals(f1, f2);
|
||||
|
||||
public byte[] ComputeHash(FullPath f)
|
||||
{
|
||||
using var stream = File.OpenRead(f.FullName);
|
||||
return _hasher.ComputeHash(stream);
|
||||
}
|
||||
|
||||
/// <summary>
|
||||
/// Recursively delete all empty directories starting from the given directory.
|
||||
/// Deletes inner directories first, so that a tree of empty directories is actually deleted.
|
||||
/// </summary>
|
||||
private static void DeleteEmptyDirectories(DirectoryInfo baseDir)
|
||||
{
|
||||
try
|
||||
{
|
||||
if (!baseDir.Exists)
|
||||
return;
|
||||
|
||||
foreach (var dir in baseDir.EnumerateDirectories("*", SearchOption.TopDirectoryOnly))
|
||||
DeleteEmptyDirectories(dir);
|
||||
|
||||
baseDir.Refresh();
|
||||
if (!baseDir.EnumerateFileSystemInfos().Any())
|
||||
Directory.Delete(baseDir.FullName, false);
|
||||
}
|
||||
catch (Exception e)
|
||||
{
|
||||
Penumbra.Log.Error($"Could not delete empty directories in {baseDir.FullName}:\n{e}");
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
/// <summary> Deduplicate a mod simply by its directory without any confirmation or waiting time. </summary>
|
||||
internal void DeduplicateMod(DirectoryInfo modDirectory)
|
||||
{
|
||||
try
|
||||
{
|
||||
var mod = new Mod(modDirectory);
|
||||
mod.Reload(true, out _);
|
||||
|
||||
Finished = false;
|
||||
_files.UpdateAll(mod, mod.Default);
|
||||
CheckDuplicates(_files.Available.OrderByDescending(f => f.FileSize).ToArray());
|
||||
DeleteDuplicates(mod, mod.Default, false);
|
||||
}
|
||||
catch (Exception e)
|
||||
{
|
||||
Penumbra.Log.Warning($"Could not deduplicate mod {modDirectory.Name}:\n{e}");
|
||||
}
|
||||
}
|
||||
}
|
||||
Loading…
Add table
Add a link
Reference in a new issue