mirror of
https://github.com/redstrate/Physis.git
synced 2025-04-21 20:27:46 +00:00
852 lines
26 KiB
Rust
Executable file
852 lines
26 KiB
Rust
Executable file
// SPDX-FileCopyrightText: 2023 Joshua Goins <josh@redstrate.com>
|
|
// SPDX-License-Identifier: GPL-3.0-or-later
|
|
|
|
use core::cmp::min;
|
|
use std::fs;
|
|
use std::fs::{File, OpenOptions, read, read_dir};
|
|
use std::io::{BufWriter, Cursor, Seek, SeekFrom, Write};
|
|
use std::path::{Path, PathBuf};
|
|
|
|
use binrw::{binrw, BinWrite};
|
|
use binrw::BinRead;
|
|
use tracing::{debug, warn};
|
|
use crate::ByteBuffer;
|
|
|
|
use crate::common::{get_platform_string, Platform, Region};
|
|
use crate::common_file_operations::{get_string_len, read_bool_from, read_string, write_bool_as, write_string};
|
|
use crate::sqpack::{read_data_block_patch, write_data_block_patch};
|
|
|
|
#[binrw]
|
|
#[derive(Debug)]
|
|
#[brw(little)]
|
|
struct PatchHeader {
|
|
#[br(temp)]
|
|
#[bw(calc = *b"ZIPATCH")]
|
|
#[brw(pad_before = 1)]
|
|
#[brw(pad_after = 4)]
|
|
#[br(assert(magic == *b"ZIPATCH"))]
|
|
magic: [u8; 7],
|
|
}
|
|
|
|
#[binrw]
|
|
#[allow(dead_code)]
|
|
#[brw(little)]
|
|
struct PatchChunk {
|
|
#[brw(big)]
|
|
size: u32,
|
|
chunk_type: ChunkType,
|
|
#[br(if(chunk_type != ChunkType::EndOfFile))]
|
|
#[bw(if(*chunk_type != ChunkType::EndOfFile))]
|
|
crc32: u32,
|
|
}
|
|
|
|
#[binrw]
|
|
#[derive(PartialEq, Debug)]
|
|
enum ChunkType {
|
|
#[brw(magic = b"FHDR")]
|
|
FileHeader(
|
|
#[brw(pad_before = 2)]
|
|
#[brw(pad_after = 1)]
|
|
FileHeaderChunk,
|
|
),
|
|
#[brw(magic = b"APLY")]
|
|
ApplyOption(ApplyOptionChunk),
|
|
#[brw(magic = b"ADIR")]
|
|
AddDirectory(DirectoryChunk),
|
|
#[brw(magic = b"DELD")]
|
|
DeleteDirectory(DirectoryChunk),
|
|
#[brw(magic = b"SQPK")]
|
|
Sqpk(SqpkChunk),
|
|
#[brw(magic = b"EOF_")]
|
|
EndOfFile,
|
|
}
|
|
|
|
#[binrw]
|
|
#[derive(PartialEq, Debug)]
|
|
enum FileHeaderChunk {
|
|
#[brw(magic = 2u8)]
|
|
Version2(FileHeaderChunk2),
|
|
#[brw(magic = 3u8)]
|
|
Version3(FileHeaderChunk3),
|
|
}
|
|
|
|
#[binrw]
|
|
#[derive(PartialEq, Debug)]
|
|
#[brw(big)]
|
|
struct FileHeaderChunk2 {
|
|
#[br(count = 4)]
|
|
#[br(map = read_string)]
|
|
#[bw(map = write_string)]
|
|
name: String,
|
|
|
|
#[brw(pad_before = 8)]
|
|
depot_hash: u32,
|
|
}
|
|
|
|
#[binrw]
|
|
#[derive(PartialEq, Debug)]
|
|
#[brw(big)]
|
|
struct FileHeaderChunk3 {
|
|
#[br(count = 4)]
|
|
#[br(map = read_string)]
|
|
#[bw(map = write_string)]
|
|
name: String,
|
|
|
|
entry_files: u32,
|
|
|
|
add_directories: u32,
|
|
delete_directories: u32,
|
|
delete_data_size: u32,
|
|
delete_data_size_2: u32,
|
|
minor_version: u32,
|
|
repository_name: u32,
|
|
commands: u32,
|
|
sqpk_add_commands: u32,
|
|
sqpk_delete_commands: u32,
|
|
sqpk_expand_commands: u32,
|
|
sqpk_header_commands: u32,
|
|
#[brw(pad_after = 0xB8)]
|
|
sqpk_file_commands: u32,
|
|
}
|
|
|
|
#[binrw]
|
|
#[brw(repr = u32)]
|
|
#[brw(big)]
|
|
#[derive(PartialEq, Debug)]
|
|
enum ApplyOption {
|
|
IgnoreMissing = 1,
|
|
IgnoreOldMismatch = 2,
|
|
}
|
|
|
|
#[binrw]
|
|
#[derive(PartialEq, Debug)]
|
|
struct ApplyOptionChunk {
|
|
#[brw(pad_after = 4)]
|
|
option: ApplyOption,
|
|
#[brw(big)]
|
|
value: u32,
|
|
}
|
|
|
|
#[binrw]
|
|
#[derive(PartialEq, Debug)]
|
|
struct DirectoryChunk {
|
|
#[br(temp)]
|
|
#[bw(calc = get_string_len(name) as u32)]
|
|
name_length: u32,
|
|
|
|
#[br(count = name_length)]
|
|
#[br(map = read_string)]
|
|
#[bw(map = write_string)]
|
|
name: String,
|
|
}
|
|
|
|
#[binrw]
|
|
#[derive(PartialEq, Debug)]
|
|
enum SqpkOperation {
|
|
#[brw(magic = b'A')]
|
|
AddData(SqpkAddData),
|
|
#[brw(magic = b'D')]
|
|
DeleteData(SqpkDeleteData),
|
|
#[brw(magic = b'E')]
|
|
ExpandData(SqpkDeleteData),
|
|
#[brw(magic = b'F')]
|
|
FileOperation(SqpkFileOperationData),
|
|
#[brw(magic = b'H')]
|
|
HeaderUpdate(SqpkHeaderUpdateData),
|
|
#[brw(magic = b'X')]
|
|
PatchInfo(SqpkPatchInfo),
|
|
#[brw(magic = b'T')]
|
|
TargetInfo(SqpkTargetInfo),
|
|
#[brw(magic = b'I')]
|
|
Index(SqpkIndex),
|
|
}
|
|
|
|
#[binrw]
|
|
#[derive(PartialEq, Debug)]
|
|
struct SqpkPatchInfo {
|
|
status: u8,
|
|
#[brw(pad_after = 1)]
|
|
version: u8,
|
|
|
|
#[brw(big)]
|
|
install_size: u64,
|
|
}
|
|
|
|
#[binrw]
|
|
#[derive(PartialEq, Debug)]
|
|
enum SqpkFileOperation {
|
|
#[brw(magic = b'A')]
|
|
AddFile,
|
|
#[brw(magic = b'R')]
|
|
RemoveAll,
|
|
#[brw(magic = b'D')]
|
|
DeleteFile,
|
|
#[brw(magic = b'M')]
|
|
MakeDirTree,
|
|
}
|
|
|
|
#[binrw]
|
|
#[derive(PartialEq, Debug)]
|
|
#[brw(big)]
|
|
struct SqpkAddData {
|
|
#[brw(pad_before = 3)]
|
|
main_id: u16,
|
|
sub_id: u16,
|
|
file_id: u32,
|
|
|
|
#[br(map = | x : u32 | (x as u64) << 7 )]
|
|
block_offset: u64,
|
|
#[br(map = | x : u32 | (x as u64) << 7 )]
|
|
block_number: u64,
|
|
#[br(map = | x : u32 | (x as u64) << 7 )]
|
|
block_delete_number: u64,
|
|
|
|
#[br(count = block_number)]
|
|
block_data: Vec<u8>,
|
|
}
|
|
|
|
#[binrw]
|
|
#[derive(PartialEq, Debug)]
|
|
#[brw(big)]
|
|
struct SqpkDeleteData {
|
|
#[brw(pad_before = 3)]
|
|
main_id: u16,
|
|
sub_id: u16,
|
|
file_id: u32,
|
|
|
|
#[br(map = | x : u32 | (x as u64) << 7 )]
|
|
block_offset: u64,
|
|
#[brw(pad_after = 4)]
|
|
block_number: u32,
|
|
}
|
|
|
|
#[binrw]
|
|
#[derive(PartialEq, Debug)]
|
|
enum TargetFileKind {
|
|
#[brw(magic = b'D')]
|
|
Dat,
|
|
#[brw(magic = b'I')]
|
|
Index,
|
|
}
|
|
|
|
#[binrw]
|
|
#[derive(PartialEq, Debug)]
|
|
enum TargetHeaderKind {
|
|
#[brw(magic = b'V')]
|
|
Version,
|
|
#[brw(magic = b'I')]
|
|
Index,
|
|
#[brw(magic = b'D')]
|
|
Data,
|
|
}
|
|
|
|
#[binrw]
|
|
#[derive(PartialEq, Debug)]
|
|
#[brw(big)]
|
|
struct SqpkHeaderUpdateData {
|
|
file_kind: TargetFileKind,
|
|
header_kind: TargetHeaderKind,
|
|
|
|
#[brw(pad_before = 1)]
|
|
main_id: u16,
|
|
sub_id: u16,
|
|
file_id: u32,
|
|
|
|
#[br(count = 1024)]
|
|
header_data: Vec<u8>,
|
|
}
|
|
|
|
#[binrw]
|
|
#[derive(PartialEq, Debug)]
|
|
#[brw(big)]
|
|
struct SqpkFileOperationData {
|
|
#[brw(pad_after = 2)]
|
|
operation: SqpkFileOperation,
|
|
|
|
offset: u64,
|
|
file_size: u64,
|
|
|
|
// Note: counts the \0 at the end... for some reason
|
|
#[br(temp)]
|
|
#[bw(calc = get_string_len(path) as u32)]
|
|
path_length: u32,
|
|
|
|
#[brw(pad_after = 2)]
|
|
expansion_id: u16,
|
|
|
|
#[br(count = path_length)]
|
|
#[br(map = read_string)]
|
|
#[bw(map = write_string)]
|
|
path: String,
|
|
}
|
|
|
|
#[binrw]
|
|
#[derive(PartialEq, Debug)]
|
|
#[brw(big)]
|
|
struct SqpkTargetInfo {
|
|
#[brw(pad_before = 3)]
|
|
#[brw(pad_size_to = 2)]
|
|
platform: Platform, // Platform is read as a u16, but the enum is u8
|
|
region: Region,
|
|
#[br(map = read_bool_from::<u16>)]
|
|
#[bw(map = write_bool_as::<u16>)]
|
|
is_debug: bool,
|
|
version: u16,
|
|
#[brw(little)]
|
|
deleted_data_size: u64,
|
|
#[brw(little)]
|
|
#[brw(pad_after = 96)]
|
|
seek_count: u64,
|
|
}
|
|
|
|
#[binrw]
|
|
#[derive(PartialEq, Debug)]
|
|
enum SqpkIndexCommand {
|
|
#[brw(magic = b'A')]
|
|
Add,
|
|
#[brw(magic = b'D')]
|
|
Delete,
|
|
}
|
|
|
|
#[binrw]
|
|
#[derive(PartialEq, Debug)]
|
|
#[brw(big)]
|
|
struct SqpkIndex {
|
|
command: SqpkIndexCommand,
|
|
#[br(map = read_bool_from::<u8>)]
|
|
#[bw(map = write_bool_as::<u8>)]
|
|
is_synonym: bool,
|
|
|
|
#[brw(pad_before = 1)]
|
|
file_hash: u64,
|
|
|
|
block_offset: u32,
|
|
#[brw(pad_after = 8)] // data?
|
|
block_number: u32,
|
|
}
|
|
|
|
#[binrw]
|
|
#[derive(PartialEq, Debug)]
|
|
#[brw(big)]
|
|
struct SqpkChunk {
|
|
size: u32,
|
|
operation: SqpkOperation,
|
|
}
|
|
|
|
const WIPE_BUFFER: [u8; 1 << 16] = [0; 1 << 16];
|
|
|
|
fn wipe(mut file: &File, length: usize) -> Result<(), PatchError> {
|
|
let mut length: usize = length;
|
|
while length > 0 {
|
|
let num_bytes = min(WIPE_BUFFER.len(), length);
|
|
file.write_all(&WIPE_BUFFER[0..num_bytes])?;
|
|
length -= num_bytes;
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
fn wipe_from_offset(mut file: &File, length: usize, offset: u64) -> Result<(), PatchError> {
|
|
file.seek(SeekFrom::Start(offset))?;
|
|
wipe(file, length)
|
|
}
|
|
|
|
fn write_empty_file_block_at(
|
|
mut file: &File,
|
|
offset: u64,
|
|
block_number: u64,
|
|
) -> Result<(), PatchError> {
|
|
wipe_from_offset(file, (block_number << 7) as usize, offset)?;
|
|
|
|
file.seek(SeekFrom::Start(offset))?;
|
|
|
|
let block_size: i32 = 1 << 7;
|
|
file.write_all(block_size.to_le_bytes().as_slice())?;
|
|
|
|
let unknown: i32 = 0;
|
|
file.write_all(unknown.to_le_bytes().as_slice())?;
|
|
|
|
let file_size: i32 = 0;
|
|
file.write_all(file_size.to_le_bytes().as_slice())?;
|
|
|
|
let num_blocks: i32 = (block_number - 1).try_into().unwrap();
|
|
file.write_all(num_blocks.to_le_bytes().as_slice())?;
|
|
|
|
let used_blocks: i32 = 0;
|
|
file.write_all(used_blocks.to_le_bytes().as_slice())?;
|
|
|
|
Ok(())
|
|
}
|
|
|
|
fn get_expansion_folder_sub(sub_id: u16) -> String {
|
|
let expansion_id = sub_id >> 8;
|
|
|
|
get_expansion_folder(expansion_id)
|
|
}
|
|
|
|
fn get_expansion_folder(id: u16) -> String {
|
|
match id {
|
|
0 => "ffxiv".to_string(),
|
|
n => format!("ex{}", n),
|
|
}
|
|
}
|
|
|
|
#[derive(Debug)]
|
|
/// Errors emitted in the patching process
|
|
pub enum PatchError {
|
|
/// Failed to read parts of the file
|
|
InvalidPatchFile,
|
|
/// Failed to parse the patch format
|
|
ParseError,
|
|
}
|
|
|
|
impl From<std::io::Error> for PatchError {
|
|
// TODO: implement specific PatchErrors for stuff like out of storage space. invalidpatchfile is a bad name for this
|
|
fn from(_: std::io::Error) -> Self {
|
|
PatchError::InvalidPatchFile
|
|
}
|
|
}
|
|
|
|
impl From<binrw::Error> for PatchError {
|
|
fn from(_: binrw::Error) -> Self {
|
|
PatchError::ParseError
|
|
}
|
|
}
|
|
|
|
fn recurse(path: impl AsRef<Path>) -> Vec<PathBuf> {
|
|
let Ok(entries) = read_dir(path) else {
|
|
return vec![];
|
|
};
|
|
entries
|
|
.flatten()
|
|
.flat_map(|entry| {
|
|
let Ok(meta) = entry.metadata() else {
|
|
return vec![];
|
|
};
|
|
if meta.is_dir() {
|
|
return recurse(entry.path());
|
|
}
|
|
if meta.is_file() {
|
|
return vec![entry.path()];
|
|
}
|
|
vec![]
|
|
})
|
|
.collect()
|
|
}
|
|
|
|
pub struct ZiPatch;
|
|
|
|
impl ZiPatch {
|
|
/// Applies a boot or a game patch to the specified _data_dir_.
|
|
pub fn apply(data_dir: &str, patch_path: &str) -> Result<(), PatchError> {
|
|
let mut file = File::open(patch_path)?;
|
|
|
|
PatchHeader::read(&mut file)?;
|
|
|
|
let mut target_info: Option<SqpkTargetInfo> = None;
|
|
|
|
let get_dat_path =
|
|
|target_info: &SqpkTargetInfo, main_id: u16, sub_id: u16, file_id: u32| -> String {
|
|
let filename = format!(
|
|
"{:02x}{:04x}.{}.dat{}",
|
|
main_id,
|
|
sub_id,
|
|
get_platform_string(&target_info.platform),
|
|
file_id
|
|
);
|
|
let path: PathBuf = [
|
|
data_dir,
|
|
"sqpack",
|
|
&get_expansion_folder_sub(sub_id),
|
|
&filename,
|
|
]
|
|
.iter()
|
|
.collect();
|
|
|
|
path.to_str().unwrap().to_string()
|
|
};
|
|
|
|
let get_index_path =
|
|
|target_info: &SqpkTargetInfo, main_id: u16, sub_id: u16, file_id: u32| -> String {
|
|
let mut filename = format!(
|
|
"{:02x}{:04x}.{}.index",
|
|
main_id,
|
|
sub_id,
|
|
get_platform_string(&target_info.platform)
|
|
);
|
|
|
|
// index files have no special ending if it's file_id == 0
|
|
if file_id != 0 {
|
|
filename += &*format!("{}", file_id);
|
|
}
|
|
|
|
let path: PathBuf = [
|
|
data_dir,
|
|
"sqpack",
|
|
&get_expansion_folder_sub(sub_id),
|
|
&filename,
|
|
]
|
|
.iter()
|
|
.collect();
|
|
|
|
path.to_str().unwrap().to_string()
|
|
};
|
|
|
|
loop {
|
|
let chunk = PatchChunk::read(&mut file)?;
|
|
|
|
match chunk.chunk_type {
|
|
ChunkType::Sqpk(pchunk) => {
|
|
match pchunk.operation {
|
|
SqpkOperation::AddData(add) => {
|
|
let filename = get_dat_path(
|
|
target_info.as_ref().unwrap(),
|
|
add.main_id,
|
|
add.sub_id,
|
|
add.file_id,
|
|
);
|
|
|
|
let (left, _) = filename.rsplit_once('/').unwrap();
|
|
fs::create_dir_all(left)?;
|
|
|
|
let mut new_file = OpenOptions::new()
|
|
.write(true)
|
|
.create(true)
|
|
.truncate(false)
|
|
.open(filename)?;
|
|
|
|
new_file.seek(SeekFrom::Start(add.block_offset))?;
|
|
|
|
new_file.write_all(&add.block_data)?;
|
|
|
|
wipe(&new_file, add.block_delete_number as usize)?;
|
|
}
|
|
SqpkOperation::DeleteData(delete) => {
|
|
let filename = get_dat_path(
|
|
target_info.as_ref().unwrap(),
|
|
delete.main_id,
|
|
delete.sub_id,
|
|
delete.file_id,
|
|
);
|
|
|
|
let new_file = OpenOptions::new()
|
|
.write(true)
|
|
.create(true)
|
|
.truncate(false)
|
|
.open(filename)?;
|
|
|
|
write_empty_file_block_at(
|
|
&new_file,
|
|
delete.block_offset,
|
|
delete.block_number as u64,
|
|
)?;
|
|
}
|
|
SqpkOperation::ExpandData(expand) => {
|
|
let filename = get_dat_path(
|
|
target_info.as_ref().unwrap(),
|
|
expand.main_id,
|
|
expand.sub_id,
|
|
expand.file_id,
|
|
);
|
|
|
|
let (left, _) = filename.rsplit_once('/').unwrap();
|
|
fs::create_dir_all(left)?;
|
|
|
|
let new_file = OpenOptions::new()
|
|
.write(true)
|
|
.create(true)
|
|
.truncate(false)
|
|
.open(filename)?;
|
|
|
|
write_empty_file_block_at(
|
|
&new_file,
|
|
expand.block_offset,
|
|
expand.block_number as u64,
|
|
)?;
|
|
}
|
|
SqpkOperation::HeaderUpdate(header) => {
|
|
let file_path = match header.file_kind {
|
|
TargetFileKind::Dat => get_dat_path(
|
|
target_info.as_ref().unwrap(),
|
|
header.main_id,
|
|
header.sub_id,
|
|
header.file_id,
|
|
),
|
|
TargetFileKind::Index => get_index_path(
|
|
target_info.as_ref().unwrap(),
|
|
header.main_id,
|
|
header.sub_id,
|
|
header.file_id,
|
|
),
|
|
};
|
|
|
|
let (left, _) = file_path.rsplit_once('/').ok_or(PatchError::ParseError)?;
|
|
fs::create_dir_all(left)?;
|
|
|
|
let mut new_file = OpenOptions::new()
|
|
.write(true)
|
|
.create(true)
|
|
.truncate(false)
|
|
.open(file_path)?;
|
|
|
|
if header.header_kind != TargetHeaderKind::Version {
|
|
new_file.seek(SeekFrom::Start(1024))?;
|
|
}
|
|
|
|
new_file.write_all(&header.header_data)?;
|
|
}
|
|
SqpkOperation::FileOperation(fop) => {
|
|
let file_path = format!("{}/{}", data_dir, fop.path);
|
|
let (parent_directory, _) = file_path.rsplit_once('/').unwrap();
|
|
|
|
match fop.operation {
|
|
SqpkFileOperation::AddFile => {
|
|
fs::create_dir_all(parent_directory)?;
|
|
|
|
// reverse reading crc32
|
|
file.seek(SeekFrom::Current(-4))?;
|
|
|
|
let mut data: Vec<u8> = Vec::with_capacity(fop.file_size as usize);
|
|
|
|
while data.len() < fop.file_size as usize {
|
|
data.append(&mut read_data_block_patch(&mut file).unwrap());
|
|
}
|
|
|
|
// re-apply crc32
|
|
file.seek(SeekFrom::Current(4))?;
|
|
|
|
// now apply the file!
|
|
let new_file = OpenOptions::new()
|
|
.write(true)
|
|
.create(true)
|
|
.truncate(false)
|
|
.open(&file_path);
|
|
|
|
if let Ok(mut file) = new_file {
|
|
if fop.offset == 0 {
|
|
file.set_len(0)?;
|
|
}
|
|
|
|
file.seek(SeekFrom::Start(fop.offset))?;
|
|
file.write_all(&data)?;
|
|
} else {
|
|
warn!("{file_path} does not exist, skipping.");
|
|
}
|
|
}
|
|
SqpkFileOperation::DeleteFile => {
|
|
if fs::remove_file(file_path.as_str()).is_err() {
|
|
warn!("Failed to remove {file_path}");
|
|
}
|
|
}
|
|
SqpkFileOperation::RemoveAll => {
|
|
let path: PathBuf =
|
|
[data_dir, "sqpack", &get_expansion_folder(fop.expansion_id)]
|
|
.iter()
|
|
.collect();
|
|
|
|
if fs::read_dir(&path).is_ok() {
|
|
fs::remove_dir_all(&path)?;
|
|
}
|
|
}
|
|
SqpkFileOperation::MakeDirTree => {
|
|
fs::create_dir_all(parent_directory)?;
|
|
}
|
|
}
|
|
}
|
|
SqpkOperation::PatchInfo(_) => {
|
|
// Currently, there's nothing we need from PatchInfo. Intentional NOP.
|
|
debug!("PATCH: NOP PatchInfo");
|
|
}
|
|
SqpkOperation::TargetInfo(new_target_info) => {
|
|
target_info = Some(new_target_info);
|
|
}
|
|
SqpkOperation::Index(_) => {
|
|
// Currently, there's nothing we need from Index command. Intentional NOP.
|
|
debug!("PATCH: NOP Index");
|
|
}
|
|
}
|
|
}
|
|
ChunkType::FileHeader(_) => {
|
|
// Currently there's nothing very useful in the FileHeader, so it's an intentional NOP.
|
|
debug!("PATCH: NOP FileHeader");
|
|
}
|
|
ChunkType::ApplyOption(_) => {
|
|
// Currently, IgnoreMissing and IgnoreOldMismatch is not used in XIVQuickLauncher either. This stays as an intentional NOP.
|
|
debug!("PATCH: NOP ApplyOption");
|
|
}
|
|
ChunkType::AddDirectory(_) => {
|
|
debug!("PATCH: NOP AddDirectory");
|
|
}
|
|
ChunkType::DeleteDirectory(_) => {
|
|
debug!("PATCH: NOP DeleteDirectory");
|
|
}
|
|
ChunkType::EndOfFile => {
|
|
return Ok(());
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/// Creates a new ZiPatch describing the diff between `base_directory` and `new_directory`.
|
|
pub fn create(base_directory: &str, new_directory: &str) -> Option<ByteBuffer> {
|
|
let mut buffer = ByteBuffer::new();
|
|
|
|
{
|
|
let cursor = Cursor::new(&mut buffer);
|
|
let mut writer = BufWriter::new(cursor);
|
|
|
|
let header = PatchHeader {};
|
|
header.write(&mut writer).ok()?;
|
|
|
|
let base_files = crate::patch::recurse(base_directory);
|
|
let new_files = crate::patch::recurse(new_directory);
|
|
|
|
// A set of files not present in base, but in new (aka added files)
|
|
let added_files: Vec<&PathBuf> = new_files.iter().filter(|item| {
|
|
let metadata = fs::metadata(item).unwrap();
|
|
!base_files.contains(item) && metadata.len() > 0 // TODO: we filter out zero byte files here, but does SqEx do that?
|
|
}).collect();
|
|
|
|
// A set of files not present in the new directory, that used to be in base (aka removedf iles)
|
|
let removed_files: Vec<&PathBuf> = base_files.iter().filter(|item| !new_files.contains(item)).collect();
|
|
|
|
// Process added files
|
|
for file in added_files {
|
|
let file_data = read(file.to_str().unwrap()).unwrap();
|
|
let relative_path = file.strip_prefix(new_directory).unwrap().to_str().unwrap().to_string();
|
|
|
|
let add_file_chunk = PatchChunk {
|
|
size: 0,
|
|
chunk_type: ChunkType::Sqpk(SqpkChunk {
|
|
size: 0,
|
|
operation: SqpkOperation::FileOperation(SqpkFileOperationData {
|
|
operation: SqpkFileOperation::AddFile,
|
|
offset: 0,
|
|
file_size: file_data.len() as u64,
|
|
expansion_id: 0,
|
|
path: relative_path,
|
|
}),
|
|
}),
|
|
crc32: 0,
|
|
};
|
|
|
|
add_file_chunk.write(&mut writer).ok()?;
|
|
|
|
// reverse reading crc32
|
|
writer.seek(SeekFrom::Current(-4)).ok()?;
|
|
|
|
// add file data, dummy ver for now
|
|
write_data_block_patch(&mut writer, file_data);
|
|
|
|
// re-apply crc32
|
|
writer.seek(SeekFrom::Current(4)).ok()?;
|
|
}
|
|
|
|
// Process deleted files
|
|
for file in removed_files {
|
|
let relative_path = file.strip_prefix(base_directory).unwrap().to_str().unwrap().to_string();
|
|
|
|
let remove_file_chunk = PatchChunk {
|
|
size: 0,
|
|
chunk_type: ChunkType::Sqpk(SqpkChunk {
|
|
size: 0,
|
|
operation: SqpkOperation::FileOperation(SqpkFileOperationData {
|
|
operation: SqpkFileOperation::DeleteFile,
|
|
offset: 0,
|
|
file_size: 0,
|
|
expansion_id: 0,
|
|
path: relative_path,
|
|
}),
|
|
}),
|
|
crc32: 0,
|
|
};
|
|
|
|
remove_file_chunk.write(&mut writer).ok()?;
|
|
}
|
|
|
|
let eof_chunk = PatchChunk {
|
|
size: 0,
|
|
chunk_type: ChunkType::EndOfFile,
|
|
crc32: 0,
|
|
};
|
|
eof_chunk.write(&mut writer).ok()?;
|
|
}
|
|
|
|
Some(buffer)
|
|
}
|
|
}
|
|
|
|
// Note that these only deal with fake patch data. To test retail patches, see tests/patching_test.rs
|
|
#[cfg(test)]
|
|
mod tests {
|
|
use std::fs::{read, write};
|
|
use std::path::PathBuf;
|
|
use std::thread::sleep;
|
|
use std::time::Duration;
|
|
|
|
use super::*;
|
|
|
|
// Prepares a temporary data directory to use
|
|
fn prepare_data_dir() -> String {
|
|
let mut dir = std::env::temp_dir();
|
|
dir.push("physis-patch-tests");
|
|
if dir.exists() {
|
|
fs::remove_dir_all(&dir);
|
|
}
|
|
|
|
fs::create_dir_all(&dir);
|
|
|
|
dir.to_str().unwrap().to_string()
|
|
}
|
|
|
|
#[test]
|
|
fn test_invalid() {
|
|
let mut d = PathBuf::from(env!("CARGO_MANIFEST_DIR"));
|
|
d.push("resources/tests");
|
|
d.push("random");
|
|
|
|
let data_dir = prepare_data_dir();
|
|
|
|
write(data_dir.clone() + "/test.patch", &read(d).unwrap());
|
|
|
|
// Feeding it invalid data should not panic
|
|
ZiPatch::apply(&data_dir.clone(), &(data_dir + "/test.patch"));
|
|
}
|
|
|
|
#[test]
|
|
fn test_add_file_op() {
|
|
let mut d = PathBuf::from(env!("CARGO_MANIFEST_DIR"));
|
|
d.push("resources/tests");
|
|
d.push("random");
|
|
|
|
let data_dir = prepare_data_dir();
|
|
|
|
let mut resources_dir = PathBuf::from(env!("CARGO_MANIFEST_DIR"));
|
|
resources_dir.push("resources/tests");
|
|
|
|
// Let's create a patch that re-creates the resources dir into our data directory
|
|
let patch = ZiPatch::create(&*data_dir, resources_dir.to_str().unwrap()).unwrap();
|
|
|
|
write(data_dir.clone() + "/test.patch", &patch);
|
|
|
|
ZiPatch::apply(&data_dir.clone(), &(data_dir.clone() + "/test.patch"));
|
|
|
|
// FIXME: For some reason, running this test by itself is fine. However when running in the test suite, it trips over itself.
|
|
// So this is a really cruddy way to wait for the files to settle.
|
|
sleep(Duration::new(1, 0));
|
|
|
|
fs::remove_file(data_dir.clone() + "/test.patch");
|
|
|
|
let old_files = recurse(&resources_dir);
|
|
let new_files = recurse(&data_dir);
|
|
|
|
let old_relative_files: Vec<&Path> = old_files.iter().filter(|item| {
|
|
let metadata = fs::metadata(item).unwrap();
|
|
metadata.len() > 0 // fitler out zero byte files because ZiPatch::create does
|
|
}).map(|x| x.strip_prefix(&resources_dir).unwrap()).collect();
|
|
let new_relative_files: Vec<&Path> = new_files.iter().map(|x| x.strip_prefix(&data_dir).unwrap()).collect();
|
|
|
|
assert_eq!(old_relative_files, new_relative_files);
|
|
}
|
|
}
|
|
|