#[cfg(target_family = "unix")] use std::os::unix::prelude::*; use std::{ collections::HashMap, hash::Hash, path::{Path, PathBuf}, time, vec, }; use anyhow::{anyhow, bail, Context, Result}; use bstr::ByteSlice; use diffy::{apply as diffy_apply, Line, Patch}; use git2_hooks::HookResult; use regex::Regex; use serde::Serialize; use super::{ branch::{ self, Branch, BranchCreateRequest, BranchId, BranchOwnershipClaims, Hunk, OwnershipClaim, }, branch_to_remote_branch, context, errors, target, Iterator, RemoteBranch, VirtualBranchesHandle, }; use crate::{ askpass::AskpassBroker, dedup::{dedup, dedup_fmt}, gb_repository, git::{ self, diff::{self, diff_files_to_hunks}, show, Commit, Refname, RemoteRefname, }, keys, project_repository::{self, conflicts, LogUntil}, projects, reader, sessions, users, }; use crate::{error::Error, git::diff::GitHunk}; type AppliedStatuses = Vec<(branch::Branch, HashMap>)>; // this struct is a mapping to the view `Branch` type in Typescript // found in src-tauri/src/routes/repo/[project_id]/types.ts // it holds a materialized view for presentation purposes of the Branch struct in Rust // which is our persisted data structure for virtual branches // // it is not persisted, it is only used for presentation purposes through the ipc // #[derive(Debug, PartialEq, Clone, Serialize)] #[serde(rename_all = "camelCase")] #[allow(clippy::struct_excessive_bools)] pub struct VirtualBranch { pub id: BranchId, pub name: String, pub notes: String, pub active: bool, pub files: Vec, pub commits: Vec, pub requires_force: bool, // does this branch require a force push to the upstream? pub conflicted: bool, // is this branch currently in a conflicted state (only for the workspace) pub order: usize, // the order in which this branch should be displayed in the UI pub upstream: Option, // the upstream branch where this branch pushes to, if any pub upstream_name: Option, // the upstream branch where this branch will push to on next push pub base_current: bool, // is this vbranch based on the current base branch? if false, this needs to be manually merged with conflicts pub ownership: BranchOwnershipClaims, pub updated_at: u128, pub selected_for_changes: bool, pub head: git::Oid, } #[derive(Debug, PartialEq, Clone, Serialize)] #[serde(rename_all = "camelCase")] pub struct VirtualBranches { pub branches: Vec, pub skipped_files: Vec, } // this is the struct that maps to the view `Commit` type in Typescript // it is derived from walking the git commits between the `Branch.head` commit // and the `Target.sha` commit, or, everything that is uniquely committed to // the virtual branch we assign it to. an array of them are returned as part of // the `VirtualBranch` struct // // it is not persisted, it is only used for presentation purposes through the ipc // #[derive(Debug, PartialEq, Clone, Serialize)] #[serde(rename_all = "camelCase")] pub struct VirtualBranchCommit { pub id: git::Oid, pub description: String, pub created_at: u128, pub author: Author, pub is_remote: bool, pub files: Vec, pub is_integrated: bool, pub parent_ids: Vec, pub branch_id: BranchId, } // this struct is a mapping to the view `File` type in Typescript // found in src-tauri/src/routes/repo/[project_id]/types.ts // it holds a materialized view for presentation purposes of one entry of the // `Branch.ownership` vector in Rust. an array of them are returned as part of // the `VirtualBranch` struct, which map to each entry of the `Branch.ownership` vector // // it is not persisted, it is only used for presentation purposes through the ipc // #[derive(Debug, PartialEq, Clone, Serialize)] #[serde(rename_all = "camelCase")] pub struct VirtualBranchFile { pub id: String, pub path: PathBuf, pub hunks: Vec, pub modified_at: u128, pub conflicted: bool, pub binary: bool, pub large: bool, } // this struct is a mapping to the view `Hunk` type in Typescript // found in src-tauri/src/routes/repo/[project_id]/types.ts // it holds a materialized view for presentation purposes of one entry of the // each hunk in one `Branch.ownership` vector entry in Rust. // an array of them are returned as part of the `VirtualBranchFile` struct // // it is not persisted, it is only used for presentation purposes through the ipc // #[derive(Debug, PartialEq, Clone, Serialize)] #[serde(rename_all = "camelCase")] pub struct VirtualBranchHunk { pub id: String, pub diff: String, pub modified_at: u128, pub file_path: PathBuf, pub hash: String, pub old_start: u32, pub start: u32, pub end: u32, pub binary: bool, pub locked: bool, pub locked_to: Option, pub change_type: diff::ChangeType, } #[derive(Debug, Serialize, Hash, Clone, PartialEq, Eq)] #[serde(rename_all = "camelCase")] pub struct Author { pub name: String, pub email: String, pub gravatar_url: url::Url, } impl From> for Author { fn from(value: git::Signature) -> Self { let name = value.name().unwrap_or_default().to_string(); let email = value.email().unwrap_or_default().to_string(); let gravatar_url = url::Url::parse(&format!( "https://www.gravatar.com/avatar/{:x}?s=100&r=g&d=retro", md5::compute(email.to_lowercase()) )) .unwrap(); Author { name, email, gravatar_url, } } } pub fn normalize_branch_name(name: &str) -> String { let pattern = Regex::new("[^A-Za-z0-9_/.#]+").unwrap(); pattern.replace_all(name, "-").to_string() } fn get_default_target( session_reader: &sessions::Reader, project: &projects::Project, ) -> Result, reader::Error> { let target_reader = target::Reader::new( session_reader, VirtualBranchesHandle::new(&project.gb_dir()), project.use_toml_vbranches_state(), ); match target_reader.read_default() { Ok(target) => Ok(Some(target)), Err(reader::Error::NotFound) => Ok(None), Err(error) => Err(error), } } pub fn apply_branch( gb_repository: &gb_repository::Repository, project_repository: &project_repository::Repository, branch_id: &BranchId, signing_key: Option<&keys::PrivateKey>, user: Option<&users::User>, ) -> Result<(), errors::ApplyBranchError> { if project_repository.is_resolving() { return Err(errors::ApplyBranchError::Conflict( errors::ProjectConflict { project_id: project_repository.project().id, }, )); } let current_session = gb_repository .get_or_create_current_session() .context("failed to get or create current session")?; let current_session_reader = sessions::Reader::open(gb_repository, ¤t_session) .context("failed to open current session")?; let repo = &project_repository.git_repository; let default_target = get_default_target(¤t_session_reader, project_repository.project()) .context("failed to get default target")? .ok_or_else(|| { errors::ApplyBranchError::DefaultTargetNotSet(errors::DefaultTargetNotSet { project_id: project_repository.project().id, }) })?; let writer = branch::Writer::new( gb_repository, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), ) .context("failed to create branch writer")?; let mut branch = match branch::Reader::new( ¤t_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ) .read(branch_id) { Ok(branch) => Ok(branch), Err(reader::Error::NotFound) => Err(errors::ApplyBranchError::BranchNotFound( errors::BranchNotFound { project_id: project_repository.project().id, branch_id: *branch_id, }, )), Err(error) => Err(errors::ApplyBranchError::Other(error.into())), }?; if branch.applied { return Ok(()); } let target_commit = repo .find_commit(default_target.sha) .context("failed to find target commit")?; let target_tree = target_commit.tree().context("failed to get target tree")?; // calculate the merge base and make sure it's the same as the target commit // if not, we need to merge or rebase the branch to get it up to date let merge_base = repo .merge_base(default_target.sha, branch.head) .context(format!( "failed to find merge base between {} and {}", default_target.sha, branch.head ))?; if merge_base != default_target.sha { // Branch is out of date, merge or rebase it let merge_base_tree = repo .find_commit(merge_base) .context(format!("failed to find merge base commit {}", merge_base))? .tree() .context("failed to find merge base tree")?; let branch_tree = repo .find_tree(branch.tree) .context("failed to find branch tree")?; let mut merge_index = repo .merge_trees(&merge_base_tree, &branch_tree, &target_tree) .context("failed to merge trees")?; if merge_index.has_conflicts() { // currently we can only deal with the merge problem branch for mut branch in super::get_status_by_branch( gb_repository, project_repository, Some(&target_commit.id()), )? .0 .into_iter() .map(|(branch, _)| branch) .filter(|branch| branch.applied) { branch.applied = false; writer.write(&mut branch)?; } // apply the branch branch.applied = true; writer.write(&mut branch)?; // checkout the conflicts repo.checkout_index(&mut merge_index) .allow_conflicts() .conflict_style_merge() .force() .checkout() .context("failed to checkout index")?; // mark conflicts let conflicts = merge_index .conflicts() .context("failed to get merge index conflicts")?; let mut merge_conflicts = Vec::new(); for path in conflicts.flatten() { if let Some(ours) = path.our { let path = std::str::from_utf8(&ours.path) .context("failed to convert path to utf8")? .to_string(); merge_conflicts.push(path); } } conflicts::mark( project_repository, &merge_conflicts, Some(default_target.sha), )?; return Ok(()); } let head_commit = repo .find_commit(branch.head) .context("failed to find head commit")?; let merged_branch_tree_oid = merge_index .write_tree_to(repo) .context("failed to write tree")?; let merged_branch_tree = repo .find_tree(merged_branch_tree_oid) .context("failed to find tree")?; let ok_with_force_push = project_repository.project().ok_with_force_push; if branch.upstream.is_some() && !ok_with_force_push { // branch was pushed to upstream, and user doesn't like force pushing. // create a merge commit to avoid the need of force pushing then. let new_branch_head = project_repository.commit( user, format!( "Merged {}/{} into {}", default_target.branch.remote(), default_target.branch.branch(), branch.name ) .as_str(), &merged_branch_tree, &[&head_commit, &target_commit], signing_key, )?; // ok, update the virtual branch branch.head = new_branch_head; branch.tree = merged_branch_tree_oid; writer.write(&mut branch)?; } else { // branch was not pushed to upstream yet. attempt a rebase, let (_, committer) = project_repository.git_signatures(user)?; let mut rebase_options = git2::RebaseOptions::new(); rebase_options.quiet(true); rebase_options.inmemory(true); let mut rebase = repo .rebase( Some(branch.head), Some(target_commit.id()), None, Some(&mut rebase_options), ) .context("failed to rebase")?; let mut rebase_success = true; // check to see if these commits have already been pushed let mut last_rebase_head = branch.head; while rebase.next().is_some() { let index = rebase .inmemory_index() .context("failed to get inmemory index")?; if index.has_conflicts() { rebase_success = false; break; } if let Ok(commit_id) = rebase.commit(None, &committer.clone().into(), None) { last_rebase_head = commit_id.into(); } else { rebase_success = false; break; } } if rebase_success { // rebase worked out, rewrite the branch head rebase.finish(None).context("failed to finish rebase")?; branch.head = last_rebase_head; branch.tree = merged_branch_tree_oid; } else { // rebase failed, do a merge commit rebase.abort().context("failed to abort rebase")?; // get tree from merge_tree_oid let merge_tree = repo .find_tree(merged_branch_tree_oid) .context("failed to find tree")?; // commit the merge tree oid let new_branch_head = project_repository .commit( user, format!( "Merged {}/{} into {}", default_target.branch.remote(), default_target.branch.branch(), branch.name ) .as_str(), &merge_tree, &[&head_commit, &target_commit], signing_key, ) .context("failed to commit merge")?; branch.head = new_branch_head; branch.tree = merged_branch_tree_oid; } } } let wd_tree = project_repository.get_wd_tree()?; let branch_tree = repo .find_tree(branch.tree) .context("failed to find branch tree")?; // check index for conflicts let mut merge_index = repo .merge_trees(&target_tree, &wd_tree, &branch_tree) .context("failed to merge trees")?; if merge_index.has_conflicts() { return Err(errors::ApplyBranchError::BranchConflicts(*branch_id)); } // apply the branch branch.applied = true; writer.write(&mut branch)?; ensure_selected_for_changes( ¤t_session_reader, &writer, project_repository.project(), ) .context("failed to ensure selected for changes")?; // checkout the merge index repo.checkout_index(&mut merge_index) .force() .checkout() .context("failed to checkout index")?; super::integration::update_gitbutler_integration(gb_repository, project_repository)?; Ok(()) } pub fn unapply_ownership( gb_repository: &gb_repository::Repository, project_repository: &project_repository::Repository, ownership: &BranchOwnershipClaims, ) -> Result<(), errors::UnapplyOwnershipError> { if conflicts::is_resolving(project_repository) { return Err(errors::UnapplyOwnershipError::Conflict( errors::ProjectConflict { project_id: project_repository.project().id, }, )); } let latest_session = gb_repository .get_latest_session() .context("failed to get or create current session")? .ok_or_else(|| { errors::UnapplyOwnershipError::DefaultTargetNotSet(errors::DefaultTargetNotSet { project_id: project_repository.project().id, }) })?; let latest_session_reader = sessions::Reader::open(gb_repository, &latest_session) .context("failed to open current session")?; let default_target = get_default_target(&latest_session_reader, project_repository.project()) .context("failed to get default target")? .ok_or_else(|| { errors::UnapplyOwnershipError::DefaultTargetNotSet(errors::DefaultTargetNotSet { project_id: project_repository.project().id, }) })?; let applied_branches = Iterator::new( &latest_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ) .context("failed to create branch iterator")? .collect::, reader::Error>>() .context("failed to read virtual branches")? .into_iter() .filter(|b| b.applied) .collect::>(); let integration_commit = super::integration::update_gitbutler_integration(gb_repository, project_repository)?; let (applied_statuses, _) = get_applied_status( gb_repository, project_repository, &integration_commit, &default_target.sha, applied_branches, ) .context("failed to get status by branch")?; let hunks_to_unapply = applied_statuses .iter() .map( |(_branch, branch_files)| -> Result> { let mut hunks_to_unapply = Vec::new(); for (path, hunks) in branch_files { let ownership_hunks: Vec<&Hunk> = ownership .claims .iter() .filter(|o| o.file_path == *path) .flat_map(|f| &f.hunks) .collect(); for hunk in hunks { if ownership_hunks.contains(&&Hunk::from(hunk)) { hunks_to_unapply.push((path.clone(), hunk)); } } } hunks_to_unapply.sort_by(|a, b| a.1.old_start.cmp(&b.1.old_start)); Ok(hunks_to_unapply) }, ) .collect::>>()? .into_iter() .flatten() .collect::>(); let mut diff = HashMap::new(); for h in hunks_to_unapply { if let Some(reversed_hunk) = diff::reverse_hunk(h.1) { diff.entry(h.0).or_insert_with(Vec::new).push(reversed_hunk); } else { return Err(errors::UnapplyOwnershipError::Other(anyhow::anyhow!( "failed to reverse hunk" ))); } } let repo = &project_repository.git_repository; let target_commit = repo .find_commit(integration_commit) .context("failed to find target commit")?; let base_tree = target_commit.tree().context("failed to get target tree")?; let final_tree = applied_statuses.into_iter().fold( target_commit.tree().context("failed to get target tree"), |final_tree, status| { let final_tree = final_tree?; let tree_oid = write_tree(project_repository, &integration_commit, &status.1)?; let branch_tree = repo.find_tree(tree_oid)?; let mut result = repo.merge_trees(&base_tree, &final_tree, &branch_tree)?; let final_tree_oid = result.write_tree_to(repo)?; repo.find_tree(final_tree_oid) .context("failed to find tree") }, )?; let final_tree_oid = write_tree_onto_tree(project_repository, &final_tree, &diff)?; let final_tree = repo .find_tree(final_tree_oid) .context("failed to find tree")?; repo.checkout_tree(&final_tree) .force() .remove_untracked() .checkout() .context("failed to checkout tree")?; super::integration::update_gitbutler_integration(gb_repository, project_repository)?; Ok(()) } // reset a file in the project to the index state pub fn reset_files( project_repository: &project_repository::Repository, files: &Vec, ) -> Result<(), errors::UnapplyOwnershipError> { if conflicts::is_resolving(project_repository) { return Err(errors::UnapplyOwnershipError::Conflict( errors::ProjectConflict { project_id: project_repository.project().id, }, )); } // for each tree, we need to checkout the entry from the index at that path // or if it doesn't exist, remove the file from the working directory let repo = &project_repository.git_repository; let index = repo.index().context("failed to get index")?; for file in files { let entry = index.get_path(Path::new(file), 0); if entry.is_some() { repo.checkout_index_path(Path::new(file)) .context("failed to checkout index")?; } else { // find the project root let project_root = &project_repository.project().path; let path = Path::new(file); //combine the project root with the file path let path = &project_root.join(path); std::fs::remove_file(path).context("failed to remove file")?; } } Ok(()) } // to unapply a branch, we need to write the current tree out, then remove those file changes from the wd pub fn unapply_branch( gb_repository: &gb_repository::Repository, project_repository: &project_repository::Repository, branch_id: &BranchId, ) -> Result, errors::UnapplyBranchError> { let session = &gb_repository .get_or_create_current_session() .context("failed to get or create currnt session")?; let current_session_reader = sessions::Reader::open(gb_repository, session).context("failed to open current session")?; let branch_reader = branch::Reader::new( ¤t_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ); let mut target_branch = branch_reader.read(branch_id).map_err(|error| match error { reader::Error::NotFound => { errors::UnapplyBranchError::BranchNotFound(errors::BranchNotFound { project_id: project_repository.project().id, branch_id: *branch_id, }) } error => errors::UnapplyBranchError::Other(error.into()), })?; if !target_branch.applied { return Ok(Some(target_branch)); } let default_target = get_default_target(¤t_session_reader, project_repository.project()) .context("failed to get default target")? .ok_or_else(|| { errors::UnapplyBranchError::DefaultTargetNotSet(errors::DefaultTargetNotSet { project_id: project_repository.project().id, }) })?; let repo = &project_repository.git_repository; let target_commit = repo .find_commit(default_target.sha) .context("failed to find target commit")?; let branch_writer = branch::Writer::new( gb_repository, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), ) .context("failed to create writer")?; let final_tree = if conflicts::is_resolving(project_repository) { { target_branch.applied = false; target_branch.selected_for_changes = None; branch_writer.write(&mut target_branch)?; } conflicts::clear(project_repository).context("failed to clear conflicts")?; target_commit.tree().context("failed to get target tree")? } else { // if we are not resolving, we need to merge the rest of the applied branches let applied_branches = Iterator::new( ¤t_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ) .context("failed to create branch iterator")? .collect::, reader::Error>>() .context("failed to read virtual branches")? .into_iter() .filter(|b| b.applied) .collect::>(); let integration_commit = super::integration::update_gitbutler_integration(gb_repository, project_repository)?; let (applied_statuses, _) = get_applied_status( gb_repository, project_repository, &integration_commit, &default_target.sha, applied_branches, ) .context("failed to get status by branch")?; let status = applied_statuses .iter() .find(|(s, _)| s.id == target_branch.id) .context("failed to find status for branch"); if let Ok((branch, files)) = status { update_conflict_markers(project_repository, files)?; if files.is_empty() && branch.head == default_target.sha { // if there is nothing to unapply, remove the branch straight away branch_writer .delete(&target_branch) .context("Failed to remove branch")?; ensure_selected_for_changes( ¤t_session_reader, &branch_writer, project_repository.project(), ) .context("failed to ensure selected for changes")?; project_repository.delete_branch_reference(&target_branch)?; return Ok(None); } target_branch.tree = write_tree(project_repository, &target_branch.head, files)?; target_branch.applied = false; target_branch.selected_for_changes = None; branch_writer.write(&mut target_branch)?; } let target_commit = repo .find_commit(default_target.sha) .context("failed to find target commit")?; // ok, update the wd with the union of the rest of the branches let base_tree = target_commit.tree().context("failed to get target tree")?; // go through the other applied branches and merge them into the final tree // then check that out into the working directory let final_tree = applied_statuses .into_iter() .filter(|(branch, _)| &branch.id != branch_id) .fold( target_commit.tree().context("failed to get target tree"), |final_tree, status| { let final_tree = final_tree?; let branch = status.0; let tree_oid = write_tree(project_repository, &branch.head, &status.1)?; let branch_tree = repo.find_tree(tree_oid)?; let mut result = repo.merge_trees(&base_tree, &final_tree, &branch_tree)?; let final_tree_oid = result.write_tree_to(repo)?; repo.find_tree(final_tree_oid) .context("failed to find tree") }, )?; ensure_selected_for_changes( ¤t_session_reader, &branch_writer, project_repository.project(), ) .context("failed to ensure selected for changes")?; final_tree }; // checkout final_tree into the working directory repo.checkout_tree(&final_tree) .force() .remove_untracked() .checkout() .context("failed to checkout tree")?; super::integration::update_gitbutler_integration(gb_repository, project_repository)?; Ok(Some(target_branch)) } fn find_base_tree<'a>( repo: &'a git::Repository, branch_commit: &'a git::Commit<'a>, target_commit: &'a git::Commit<'a>, ) -> Result> { // find merge base between target_commit and branch_commit let merge_base = repo .merge_base(target_commit.id(), branch_commit.id()) .context("failed to find merge base")?; // turn oid into a commit let merge_base_commit = repo .find_commit(merge_base) .context("failed to find merge base commit")?; let base_tree = merge_base_commit .tree() .context("failed to get base tree object")?; Ok(base_tree) } pub fn list_virtual_branches( gb_repository: &gb_repository::Repository, project_repository: &project_repository::Repository, ) -> Result<(Vec, bool, Vec), errors::ListVirtualBranchesError> { let mut branches: Vec = Vec::new(); let default_target = gb_repository .default_target() .context("failed to get default target")? .ok_or_else(|| { errors::ListVirtualBranchesError::DefaultTargetNotSet(errors::DefaultTargetNotSet { project_id: project_repository.project().id, }) })?; let integration_commit = super::integration::update_gitbutler_integration(gb_repository, project_repository)?; let (statuses, skipped_files) = get_status_by_branch(gb_repository, project_repository, Some(&integration_commit))?; let max_selected_for_changes = statuses .iter() .filter_map(|(branch, _)| branch.selected_for_changes) .max() .unwrap_or(-1); for (branch, files) in &statuses { let repo = &project_repository.git_repository; update_conflict_markers(project_repository, files)?; let upstream_branch = match branch .upstream .as_ref() .map(|name| repo.find_branch(&git::Refname::from(name))) .transpose() { Err(git::Error::NotFound(_)) => Ok(None), Err(error) => Err(error), Ok(branch) => Ok(branch), } .context(format!( "failed to find upstream branch for {}", branch.name ))?; let upstram_branch_commit = upstream_branch .as_ref() .map(git::Branch::peel_to_commit) .transpose() .context(format!( "failed to find upstream branch commit for {}", branch.name ))?; // find upstream commits if we found an upstream reference let mut pushed_commits = HashMap::new(); if let Some(upstream) = &upstram_branch_commit { let merge_base = repo.merge_base(upstream.id(), default_target.sha) .context(format!( "failed to find merge base between {} and {}", upstream.id(), default_target.sha ))?; for oid in project_repository.l(upstream.id(), LogUntil::Commit(merge_base))? { pushed_commits.insert(oid, true); } } let mut is_integrated = false; let mut is_remote = false; // find all commits on head that are not on target.sha let commits = project_repository .log(branch.head, LogUntil::Commit(default_target.sha)) .context(format!("failed to get log for branch {}", branch.name))? .iter() .map(|commit| { is_remote = if is_remote { is_remote } else { pushed_commits.contains_key(&commit.id()) }; // only check for integration if we haven't already found an integration is_integrated = if is_integrated { is_integrated } else { is_commit_integrated(project_repository, &default_target, commit)? }; commit_to_vbranch_commit( project_repository, branch, commit, is_integrated, is_remote, ) }) .collect::>>()?; // if the branch is not applied, check to see if it's mergeable and up to date let mut base_current = true; if !branch.applied { // determine if this branch is up to date with the target/base let merge_base = repo .merge_base(default_target.sha, branch.head) .context("failed to find merge base")?; if merge_base != default_target.sha { base_current = false; } } let upstream = upstream_branch .map(|upstream_branch| branch_to_remote_branch(&upstream_branch)) .transpose()? .flatten(); let mut files = diffs_to_virtual_files(project_repository, files); files.sort_by(|a, b| { branch .ownership .claims .iter() .position(|o| o.file_path.eq(&a.path)) .unwrap_or(999) .cmp( &branch .ownership .claims .iter() .position(|id| id.file_path.eq(&b.path)) .unwrap_or(999), ) }); let requires_force = is_requires_force(project_repository, branch)?; let branch = VirtualBranch { id: branch.id, name: branch.name.clone(), notes: branch.notes.clone(), active: branch.applied, files, order: branch.order, commits, requires_force, upstream, upstream_name: branch .upstream .clone() .and_then(|r| Refname::from(r).branch().map(Into::into)), conflicted: conflicts::is_resolving(project_repository), base_current, ownership: branch.ownership.clone(), updated_at: branch.updated_timestamp_ms, selected_for_changes: branch.selected_for_changes == Some(max_selected_for_changes), head: branch.head, }; branches.push(branch); } let branches = branches_with_large_files_abridged(branches); let mut branches = branches_with_hunk_locks(branches, project_repository)?; branches.sort_by(|a, b| a.order.cmp(&b.order)); let uses_diff_context = project_repository .project() .use_diff_context .unwrap_or(false); Ok((branches, uses_diff_context, skipped_files)) } fn branches_with_large_files_abridged(mut branches: Vec) -> Vec { for branch in &mut branches { for file in &mut branch.files { // Diffs larger than 500kb are considered large if file.hunks.iter().any(|hunk| hunk.diff.len() > 500_000) { file.large = true; file.hunks .iter_mut() .for_each(|hunk| hunk.diff = String::new()); } } } branches } fn branches_with_hunk_locks( mut branches: Vec, project_repository: &project_repository::Repository, ) -> Result> { let all_commits: Vec = branches .clone() .iter() .filter(|branch| branch.active) .flat_map(|vbranch| vbranch.commits.clone()) .collect(); for commit in all_commits { let commit = project_repository.git_repository.find_commit(commit.id)?; let parent = commit.parent(0).context("failed to get parent commit")?; let commit_tree = commit.tree().context("failed to get commit tree")?; let parent_tree = parent.tree().context("failed to get parent tree")?; let commited_file_diffs = diff::trees( &project_repository.git_repository, &parent_tree, &commit_tree, 3, )?; let commited_file_diffs = diff::diff_files_to_hunks(&commited_file_diffs); for branch in &mut branches { for file in &mut branch.files { for hunk in &mut file.hunks { let locked = commited_file_diffs .get(&file.path) .map_or(false, |committed_hunks| { committed_hunks.iter().any(|committed_hunk| { joined( committed_hunk.new_start, committed_hunk.new_start + committed_hunk.new_lines, hunk.start, hunk.end, ) }) }); if locked { hunk.locked = true; hunk.locked_to = Some(commit.id()); } } } } } Ok(branches) } fn joined(start_a: u32, end_a: u32, start_b: u32, end_b: u32) -> bool { ((start_a >= start_b && start_a <= end_b) || (end_a >= start_b && end_a <= end_b)) || ((start_b >= start_a && start_b <= end_a) || (end_b >= start_a && end_b <= end_a)) } fn files_with_hunk_context( repository: &git::Repository, mut files: Vec, context_lines: usize, branch_head: git::Oid, ) -> Result> { for file in &mut files { if file.binary { continue; } // Get file content as it looked before the diffs let branch_head_commit = repository.find_commit(branch_head)?; let head_tree = branch_head_commit.tree()?; let file_content_before = show::show_file_at_tree(repository, file.path.clone(), &head_tree) .context("failed to get file contents at base")?; let file_lines_before = file_content_before.split('\n').collect::>(); // Update each hunk with contex lines before & after file.hunks = file .hunks .iter() .map(|hunk| { if hunk.diff.is_empty() { // noop on empty diff hunk.clone() } else { let hunk_with_ctx = context::hunk_with_context( &hunk.diff, hunk.old_start as usize, hunk.start as usize, hunk.binary, context_lines, &file_lines_before, hunk.change_type, ); to_virtual_branch_hunk(hunk.clone(), hunk_with_ctx) } }) .collect::>(); } Ok(files) } fn to_virtual_branch_hunk( mut hunk: VirtualBranchHunk, diff_with_context: diff::GitHunk, ) -> VirtualBranchHunk { hunk.diff = diff_with_context.diff; hunk.start = diff_with_context.new_start; hunk.end = diff_with_context.new_start + diff_with_context.new_lines; hunk } fn is_requires_force( project_repository: &project_repository::Repository, branch: &branch::Branch, ) -> Result { let upstream = if let Some(upstream) = &branch.upstream { upstream } else { return Ok(false); }; let reference = match project_repository .git_repository .refname_to_id(&upstream.to_string()) { Ok(reference) => reference, Err(git::Error::NotFound(_)) => return Ok(false), Err(other) => return Err(other).context("failed to find upstream reference"), }; let upstream_commit = project_repository .git_repository .find_commit(reference) .context("failed to find upstream commit")?; let merge_base = project_repository .git_repository .merge_base(upstream_commit.id(), branch.head)?; Ok(merge_base != upstream_commit.id()) } fn list_virtual_commit_files( project_repository: &project_repository::Repository, commit: &git::Commit, ) -> Result> { if commit.parent_count() == 0 { return Ok(vec![]); } let parent = commit.parent(0).context("failed to get parent commit")?; let commit_tree = commit.tree().context("failed to get commit tree")?; let parent_tree = parent.tree().context("failed to get parent tree")?; let diff = diff::trees( &project_repository.git_repository, &parent_tree, &commit_tree, 3, )?; let diff = diff::diff_files_to_hunks(&diff); let hunks_by_filepath = virtual_hunks_by_filepath(&project_repository.project().path, &diff); Ok(virtual_hunks_to_virtual_files( project_repository, &hunks_by_filepath .values() .flatten() .cloned() .collect::>(), )) } fn commit_to_vbranch_commit( repository: &project_repository::Repository, branch: &branch::Branch, commit: &git::Commit, is_integrated: bool, is_remote: bool, ) -> Result { let timestamp = u128::try_from(commit.time().seconds())?; let signature = commit.author(); let message = commit.message().unwrap().to_string(); let files = list_virtual_commit_files(repository, commit).context("failed to list commit files")?; let parent_ids = commit.parents()?.iter().map(Commit::id).collect::>(); let commit = VirtualBranchCommit { id: commit.id(), created_at: timestamp * 1000, author: Author::from(signature), description: message, is_remote, files, is_integrated, parent_ids, branch_id: branch.id, }; Ok(commit) } pub fn create_virtual_branch( gb_repository: &gb_repository::Repository, project_repository: &project_repository::Repository, create: &BranchCreateRequest, ) -> Result { let current_session = gb_repository .get_or_create_current_session() .context("failed to get or create currnt session")?; let current_session_reader = sessions::Reader::open(gb_repository, ¤t_session) .context("failed to open current session")?; let default_target = get_default_target(¤t_session_reader, project_repository.project()) .context("failed to get default target")? .ok_or_else(|| { errors::CreateVirtualBranchError::DefaultTargetNotSet(errors::DefaultTargetNotSet { project_id: project_repository.project().id, }) })?; let commit = project_repository .git_repository .find_commit(default_target.sha) .context("failed to find default target commit")?; let tree = commit .tree() .context("failed to find defaut target commit tree")?; let mut all_virtual_branches = Iterator::new( ¤t_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ) .context("failed to create branch iterator")? .collect::, reader::Error>>() .context("failed to read virtual branches")?; all_virtual_branches.sort_by_key(|branch| branch.order); let order = create .order .unwrap_or(all_virtual_branches.len()) .clamp(0, all_virtual_branches.len()); let branch_writer = branch::Writer::new( gb_repository, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), ) .context("failed to create writer")?; let selected_for_changes = if let Some(selected_for_changes) = create.selected_for_changes { if selected_for_changes { for mut other_branch in Iterator::new( ¤t_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ) .context("failed to create branch iterator")? .collect::, reader::Error>>() .context("failed to read virtual branches")? { other_branch.selected_for_changes = None; branch_writer.write(&mut other_branch)?; } Some(chrono::Utc::now().timestamp_millis()) } else { None } } else { (!all_virtual_branches .iter() .any(|b| b.selected_for_changes.is_some())) .then_some(chrono::Utc::now().timestamp_millis()) }; // make space for the new branch for (i, branch) in all_virtual_branches.iter().enumerate() { let mut branch = branch.clone(); let new_order = if i < order { i } else { i + 1 }; if branch.order != new_order { branch.order = new_order; branch_writer .write(&mut branch) .context("failed to write branch")?; } } let now = time::UNIX_EPOCH .elapsed() .context("failed to get elapsed time")? .as_millis(); let name = dedup( &all_virtual_branches .iter() .map(|b| b.name.as_str()) .collect::>(), create .name .as_ref() .unwrap_or(&"Virtual branch".to_string()), ); let mut branch = Branch { id: BranchId::generate(), name, notes: String::new(), applied: true, upstream: None, upstream_head: None, tree: tree.id(), head: default_target.sha, created_timestamp_ms: now, updated_timestamp_ms: now, ownership: BranchOwnershipClaims::default(), order, selected_for_changes, }; if let Some(ownership) = &create.ownership { set_ownership( ¤t_session_reader, &branch_writer, &mut branch, ownership, project_repository.project(), ) .context("failed to set ownership")?; } branch_writer .write(&mut branch) .context("failed to write branch")?; project_repository.add_branch_reference(&branch)?; Ok(branch) } pub fn merge_virtual_branch_upstream( gb_repository: &gb_repository::Repository, project_repository: &project_repository::Repository, branch_id: &BranchId, signing_key: Option<&keys::PrivateKey>, user: Option<&users::User>, ) -> Result<(), errors::MergeVirtualBranchUpstreamError> { if conflicts::is_conflicting::<&Path>(project_repository, None)? { return Err(errors::MergeVirtualBranchUpstreamError::Conflict( errors::ProjectConflict { project_id: project_repository.project().id, }, )); } let current_session = gb_repository .get_or_create_current_session() .context("failed to get current session")?; let current_session_reader = sessions::Reader::open(gb_repository, ¤t_session) .context("failed to open current session")?; // get the branch let branch_reader = branch::Reader::new( ¤t_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ); let mut branch = match branch_reader.read(branch_id) { Ok(branch) => Ok(branch), Err(reader::Error::NotFound) => Err( errors::MergeVirtualBranchUpstreamError::BranchNotFound(errors::BranchNotFound { project_id: project_repository.project().id, branch_id: *branch_id, }), ), Err(error) => Err(errors::MergeVirtualBranchUpstreamError::Other(error.into())), }?; // check if the branch upstream can be merged into the wd cleanly let repo = &project_repository.git_repository; // get upstream from the branch and find the remote branch let mut upstream_commit = None; let upstream_branch = branch .upstream .as_ref() .context("no upstream branch found")?; if let Ok(upstream_oid) = repo.refname_to_id(&upstream_branch.to_string()) { if let Ok(upstream_commit_obj) = repo.find_commit(upstream_oid) { upstream_commit = Some(upstream_commit_obj); } } // if there is no upstream commit, then there is nothing to do if upstream_commit.is_none() { // no upstream commit, no merge to be done return Ok(()); } // there is an upstream commit, so lets check it out let upstream_commit = upstream_commit.unwrap(); let remote_tree = upstream_commit.tree().context("failed to get tree")?; if upstream_commit.id() == branch.head { // upstream is already merged, nothing to do return Ok(()); } // if any other branches are applied, unapply them let applied_branches = Iterator::new( ¤t_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ) .context("failed to create branch iterator")? .collect::, reader::Error>>() .context("failed to read virtual branches")? .into_iter() .filter(|b| b.applied) .filter(|b| b.id != *branch_id) .collect::>(); // unapply all other branches for other_branch in applied_branches { unapply_branch(gb_repository, project_repository, &other_branch.id) .context("failed to unapply branch")?; } // get merge base from remote branch commit and target commit let merge_base = repo .merge_base(upstream_commit.id(), branch.head) .context("failed to find merge base")?; let merge_tree = repo .find_commit(merge_base) .and_then(|c| c.tree()) .context(format!( "failed to find merge base commit {} tree", merge_base ))?; // get wd tree let wd_tree = project_repository.get_wd_tree()?; // try to merge our wd tree with the upstream tree let mut merge_index = repo .merge_trees(&merge_tree, &wd_tree, &remote_tree) .context("failed to merge trees")?; if merge_index.has_conflicts() { // checkout the conflicts repo.checkout_index(&mut merge_index) .allow_conflicts() .conflict_style_merge() .force() .checkout() .context("failed to checkout index")?; // mark conflicts let conflicts = merge_index.conflicts().context("failed to get conflicts")?; let mut merge_conflicts = Vec::new(); for path in conflicts.flatten() { if let Some(ours) = path.our { let path = std::str::from_utf8(&ours.path) .context("failed to convert path to utf8")? .to_string(); merge_conflicts.push(path); } } conflicts::mark( project_repository, &merge_conflicts, Some(upstream_commit.id()), )?; } else { let merge_tree_oid = merge_index .write_tree_to(repo) .context("failed to write tree")?; let merge_tree = repo .find_tree(merge_tree_oid) .context("failed to find merge tree")?; let branch_writer = branch::Writer::new( gb_repository, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), ) .context("failed to create writer")?; if *project_repository.project().ok_with_force_push { // attempt a rebase let (_, committer) = project_repository.git_signatures(user)?; let mut rebase_options = git2::RebaseOptions::new(); rebase_options.quiet(true); rebase_options.inmemory(true); let mut rebase = repo .rebase( Some(branch.head), Some(upstream_commit.id()), None, Some(&mut rebase_options), ) .context("failed to rebase")?; let mut rebase_success = true; // check to see if these commits have already been pushed let mut last_rebase_head = upstream_commit.id(); while rebase.next().is_some() { let index = rebase .inmemory_index() .context("failed to get inmemory index")?; if index.has_conflicts() { rebase_success = false; break; } if let Ok(commit_id) = rebase.commit(None, &committer.clone().into(), None) { last_rebase_head = commit_id.into(); } else { rebase_success = false; break; } } if rebase_success { // rebase worked out, rewrite the branch head rebase.finish(None).context("failed to finish rebase")?; project_repository .git_repository .checkout_tree(&merge_tree) .force() .checkout() .context("failed to checkout tree")?; branch.head = last_rebase_head; branch.tree = merge_tree_oid; branch_writer.write(&mut branch)?; super::integration::update_gitbutler_integration( gb_repository, project_repository, )?; return Ok(()); } rebase.abort().context("failed to abort rebase")?; } let head_commit = repo .find_commit(branch.head) .context("failed to find head commit")?; let new_branch_head = project_repository.commit( user, format!( "Merged {}/{} into {}", upstream_branch.remote(), upstream_branch.branch(), branch.name ) .as_str(), &merge_tree, &[&head_commit, &upstream_commit], signing_key, )?; // checkout the merge tree repo.checkout_tree(&merge_tree) .force() .checkout() .context("failed to checkout tree")?; // write the branch data branch.head = new_branch_head; branch.tree = merge_tree_oid; branch_writer.write(&mut branch)?; } super::integration::update_gitbutler_integration(gb_repository, project_repository)?; Ok(()) } pub fn update_branch( gb_repository: &gb_repository::Repository, project_repository: &project_repository::Repository, branch_update: branch::BranchUpdateRequest, ) -> Result { let current_session = gb_repository .get_or_create_current_session() .context("failed to get or create currnt session")?; let current_session_reader = sessions::Reader::open(gb_repository, ¤t_session) .context("failed to open current session")?; let branch_reader = branch::Reader::new( ¤t_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ); let branch_writer = branch::Writer::new( gb_repository, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), ) .context("failed to create writer")?; let mut branch = branch_reader .read(&branch_update.id) .map_err(|error| match error { reader::Error::NotFound => { errors::UpdateBranchError::BranchNotFound(errors::BranchNotFound { project_id: project_repository.project().id, branch_id: branch_update.id, }) } _ => errors::UpdateBranchError::Other(error.into()), })?; if let Some(ownership) = branch_update.ownership { set_ownership( ¤t_session_reader, &branch_writer, &mut branch, &ownership, project_repository.project(), ) .context("failed to set ownership")?; } if let Some(name) = branch_update.name { let all_virtual_branches = Iterator::new( ¤t_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ) .context("failed to create branch iterator")? .collect::, reader::Error>>() .context("failed to read virtual branches")?; project_repository.delete_branch_reference(&branch)?; branch.name = dedup( &all_virtual_branches .iter() .map(|b| b.name.as_str()) .collect::>(), &name, ); project_repository.add_branch_reference(&branch)?; }; if let Some(updated_upstream) = branch_update.upstream { let default_target = get_default_target(¤t_session_reader, project_repository.project()) .context("failed to get default target")? .ok_or_else(|| { errors::UpdateBranchError::DefaultTargetNotSet(errors::DefaultTargetNotSet { project_id: project_repository.project().id, }) })?; let remote_branch = format!( "refs/remotes/{}/{}", default_target.branch.remote(), normalize_branch_name(&updated_upstream) ) .parse::() .unwrap(); branch.upstream = Some(remote_branch); }; if let Some(notes) = branch_update.notes { branch.notes = notes; }; if let Some(order) = branch_update.order { branch.order = order; }; if let Some(selected_for_changes) = branch_update.selected_for_changes { branch.selected_for_changes = if selected_for_changes { for mut other_branch in Iterator::new( ¤t_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ) .context("failed to create branch iterator")? .collect::, reader::Error>>() .context("failed to read virtual branches")? .into_iter() .filter(|b| b.id != branch.id) { other_branch.selected_for_changes = None; branch_writer.write(&mut other_branch)?; } Some(chrono::Utc::now().timestamp_millis()) } else { None }; }; branch_writer .write(&mut branch) .context("failed to write target branch")?; Ok(branch) } pub fn delete_branch( gb_repository: &gb_repository::Repository, project_repository: &project_repository::Repository, branch_id: &BranchId, ) -> Result<(), Error> { let current_session = gb_repository .get_or_create_current_session() .context("failed to get or create currnt session")?; let current_session_reader = sessions::Reader::open(gb_repository, ¤t_session) .context("failed to open current session")?; let branch_reader = branch::Reader::new( ¤t_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ); let branch_writer = branch::Writer::new( gb_repository, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), ) .context("failed to create writer")?; let branch = match branch_reader.read(branch_id) { Ok(branch) => Ok(branch), Err(reader::Error::NotFound) => return Ok(()), Err(error) => Err(error), } .context("failed to read branch")?; if branch.applied && unapply_branch(gb_repository, project_repository, branch_id)?.is_none() { return Ok(()); } branch_writer .delete(&branch) .context("Failed to remove branch")?; project_repository.delete_branch_reference(&branch)?; ensure_selected_for_changes( ¤t_session_reader, &branch_writer, project_repository.project(), ) .context("failed to ensure selected for changes")?; Ok(()) } fn ensure_selected_for_changes( current_session_reader: &sessions::Reader, branch_writer: &branch::Writer, project: &projects::Project, ) -> Result<()> { let mut applied_branches = Iterator::new( current_session_reader, VirtualBranchesHandle::new(&project.gb_dir()), project.use_toml_vbranches_state(), ) .context("failed to create branch iterator")? .collect::, reader::Error>>() .context("failed to read virtual branches")? .into_iter() .filter(|b| b.applied) .collect::>(); if applied_branches.is_empty() { println!("no applied branches"); return Ok(()); } if applied_branches .iter() .any(|b| b.selected_for_changes.is_some()) { println!("some branches already selected for changes"); return Ok(()); } applied_branches.sort_by_key(|branch| branch.order); applied_branches[0].selected_for_changes = Some(chrono::Utc::now().timestamp_millis()); branch_writer.write(&mut applied_branches[0])?; Ok(()) } fn set_ownership( session_reader: &sessions::Reader, branch_writer: &branch::Writer, target_branch: &mut branch::Branch, ownership: &branch::BranchOwnershipClaims, project: &projects::Project, ) -> Result<()> { if target_branch.ownership.eq(ownership) { // nothing to update return Ok(()); } let virtual_branches = Iterator::new( session_reader, VirtualBranchesHandle::new(&project.gb_dir()), project.use_toml_vbranches_state(), ) .context("failed to create branch iterator")? .collect::, reader::Error>>() .context("failed to read virtual branches")?; let mut claim_outcomes = branch::reconcile_claims(virtual_branches, target_branch, &ownership.claims)?; for claim_outcome in &mut claim_outcomes { if !claim_outcome.removed_claims.is_empty() { branch_writer .write(&mut claim_outcome.updated_branch) .context("failed to write ownership for branch".to_string())?; } } // Updates the claiming branch that was passed as mutable state with the new ownership claims // TODO: remove mutable reference to target_branch target_branch.ownership = ownership.clone(); Ok(()) } fn get_mtime(cache: &mut HashMap, file_path: &PathBuf) -> u128 { if let Some(mtime) = cache.get(file_path) { *mtime } else { let mtime = file_path .metadata() .map_or_else( |_| time::SystemTime::now(), |metadata| { metadata .modified() .or(metadata.created()) .unwrap_or_else(|_| time::SystemTime::now()) }, ) .duration_since(time::UNIX_EPOCH) .unwrap() .as_millis(); cache.insert(file_path.clone(), mtime); mtime } } pub fn virtual_hunks_by_filepath( project_path: &Path, diff: &HashMap>, ) -> HashMap> { let mut mtimes: HashMap = HashMap::new(); diff.iter() .map(|(file_path, hunks)| { let hunks = hunks .iter() .map(|hunk| VirtualBranchHunk { id: format!("{}-{}", hunk.new_start, hunk.new_start + hunk.new_lines), modified_at: get_mtime(&mut mtimes, &project_path.join(file_path)), file_path: file_path.clone(), diff: hunk.diff.clone(), old_start: hunk.old_start, start: hunk.new_start, end: hunk.new_start + hunk.new_lines, binary: hunk.binary, hash: Hunk::hash(&hunk.diff), locked: false, locked_to: None, change_type: hunk.change_type, }) .collect::>(); (file_path.clone(), hunks) }) .collect::>() } pub type BranchStatus = HashMap>; // list the virtual branches and their file statuses (statusi?) #[allow(clippy::type_complexity)] pub fn get_status_by_branch( gb_repository: &gb_repository::Repository, project_repository: &project_repository::Repository, integration_commit: Option<&git::Oid>, ) -> Result<(Vec<(branch::Branch, BranchStatus)>, Vec)> { let latest_session = gb_repository .get_latest_session() .context("failed to get latest session")? .context("latest session not found")?; let session_reader = sessions::Reader::open(gb_repository, &latest_session) .context("failed to open current session")?; let default_target = match get_default_target(&session_reader, project_repository.project()) .context("failed to read default target")? { Some(target) => target, None => { return Ok((vec![], vec![])); } }; let virtual_branches = Iterator::new( &session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ) .context("failed to create branch iterator")? .collect::, reader::Error>>() .context("failed to read virtual branches")?; let applied_virtual_branches = virtual_branches .iter() .filter(|branch| branch.applied) .cloned() .collect::>(); let (applied_status, skipped_files) = get_applied_status( gb_repository, project_repository, // TODO: Keep this optional or update lots of tests? integration_commit.unwrap_or(&default_target.sha), &default_target.sha, applied_virtual_branches, )?; let non_applied_virtual_branches = virtual_branches .into_iter() .filter(|branch| !branch.applied) .collect::>(); let non_applied_status = get_non_applied_status(project_repository, non_applied_virtual_branches)?; Ok(( applied_status .into_iter() .chain(non_applied_status) .collect(), skipped_files, )) } // given a list of non applied virtual branches, return the status of each file, comparing the default target with // virtual branch latest tree // // ownerships are not taken into account here, as they are not relevant for non applied branches fn get_non_applied_status( project_repository: &project_repository::Repository, virtual_branches: Vec, ) -> Result> { virtual_branches .into_iter() .map( |branch| -> Result<(branch::Branch, HashMap>)> { if branch.applied { bail!("branch {} is applied", branch.name); } let branch_tree = project_repository .git_repository .find_tree(branch.tree) .context(format!("failed to find tree {}", branch.tree))?; let head_tree = project_repository .git_repository .find_commit(branch.head) .context("failed to find target commit")? .tree() .context("failed to find target tree")?; let diff = diff::trees( &project_repository.git_repository, &head_tree, &branch_tree, 3, )?; Ok((branch, diff::diff_files_to_hunks(&diff))) }, ) .collect::>>() } // given a list of applied virtual branches, return the status of each file, comparing the default target with // the working directory // // ownerships are updated if nessessary fn get_applied_status( gb_repository: &gb_repository::Repository, project_repository: &project_repository::Repository, integration_commit: &git::Oid, target_sha: &git::Oid, mut virtual_branches: Vec, ) -> Result<(AppliedStatuses, Vec)> { let base_file_diffs = diff::workdir(&project_repository.git_repository, integration_commit, 3) .context("failed to diff workdir")?; let mut base_diffs: HashMap> = diff_files_to_hunks(&base_file_diffs); let mut skipped_files: Vec = Vec::new(); for (_, file_diff) in base_file_diffs { if file_diff.skipped { skipped_files.push(file_diff); } } // sort by order, so that the default branch is first (left in the ui) virtual_branches.sort_by(|a, b| a.order.cmp(&b.order)); if virtual_branches.is_empty() && !base_diffs.is_empty() { // no virtual branches, but hunks: create default branch virtual_branches = vec![create_virtual_branch( gb_repository, project_repository, &BranchCreateRequest::default(), ) .context("failed to create default branch")?]; } // align branch ownership to the real hunks: // - update shifted hunks // - remove non existent hunks let mut diffs_by_branch: HashMap>> = virtual_branches .iter() .map(|branch| (branch.id, HashMap::new())) .collect(); let mut mtimes = HashMap::new(); let mut git_hunk_map = HashMap::new(); for branch in &virtual_branches { if !branch.applied { bail!("branch {} is not applied", branch.name); } let commits = project_repository.log(branch.head, LogUntil::Commit(*target_sha)); if let Ok(commits) = commits { for commit in commits { let commit = project_repository .git_repository .find_commit(commit.id()) .unwrap(); let parent = commit .parent(0) .context("failed to get parent commit") .unwrap(); let commit_tree = commit.tree().context("failed to get commit tree").unwrap(); let parent_tree = parent.tree().context("failed to get parent tree").unwrap(); let commited_file_diffs = diff::trees( &project_repository.git_repository, &parent_tree, &commit_tree, 3, ); let commited_file_diffs = diff::diff_files_to_hunks(&commited_file_diffs.unwrap()); for (path, committed_git_hunks) in commited_file_diffs.iter() { if let Some(uncommitted_git_hunks) = base_diffs.get_mut(path) { for uncommitted_git_hunk in uncommitted_git_hunks { for committed_git_hunk in committed_git_hunks { if joined( uncommitted_git_hunk.new_start, uncommitted_git_hunk.new_start + uncommitted_git_hunk.new_lines, committed_git_hunk.new_start, committed_git_hunk.new_start + committed_git_hunk.new_lines, ) { let hash = Hunk::hash(&uncommitted_git_hunk.diff); git_hunk_map.insert(hash, branch.id); } } } } } } } } for branch in &mut virtual_branches { if !branch.applied { bail!("branch {} is not applied", branch.name); } let old_claims = branch.ownership.claims.clone(); let new_claims = old_claims .iter() .filter_map(|claim| { let git_diff_hunks = match base_diffs.get_mut(&claim.file_path) { None => return None, Some(hunks) => hunks, }; let mtime = get_mtime(&mut mtimes, &claim.file_path); let claimed_hunks: Vec = claim .hunks .iter() .filter_map(|claimed_hunk| { // if any of the current hunks intersects with the owned hunk, we want to keep it for (i, git_diff_hunk) in git_diff_hunks.iter().enumerate() { let hash = Hunk::hash(&git_diff_hunk.diff); if let Some(locked_to) = git_hunk_map.get(&hash) { if locked_to != &branch.id { return None; } } if claimed_hunk.eq(&Hunk::from(git_diff_hunk)) { // try to re-use old timestamp let timestamp = claimed_hunk.timestam_ms().unwrap_or(mtime); diffs_by_branch .entry(branch.id) .or_default() .entry(claim.file_path.clone()) .or_default() .push(git_diff_hunk.clone()); git_diff_hunks.remove(i); return Some( claimed_hunk .with_timestamp(timestamp) .with_hash(hash.as_str()), ); } else if claimed_hunk.intersects(git_diff_hunk) { diffs_by_branch .entry(branch.id) .or_default() .entry(claim.file_path.clone()) .or_default() .insert(0, git_diff_hunk.clone()); let updated_hunk = Hunk { start: git_diff_hunk.new_start, end: git_diff_hunk.new_start + git_diff_hunk.new_lines, timestamp_ms: Some(mtime), hash: Some(hash.clone()), }; git_diff_hunks.remove(i); return Some(updated_hunk); } } None }) .collect(); if claimed_hunks.is_empty() { // No need for an empty claim None } else { Some(OwnershipClaim { file_path: claim.file_path.clone(), hunks: claimed_hunks, }) } }) .collect(); branch.ownership = BranchOwnershipClaims { claims: new_claims }; } let max_selected_for_changes = virtual_branches .iter() .filter_map(|b| b.selected_for_changes) .max() .unwrap_or(-1); let default_vbranch_pos = virtual_branches .iter() .position(|b| b.selected_for_changes == Some(max_selected_for_changes)) .unwrap_or(0); for (filepath, hunks) in base_diffs { for hunk in hunks { let hash = Hunk::hash(&hunk.diff); let vbranch_pos = if let Some(locked_to) = git_hunk_map.get(&hash) { let p = virtual_branches.iter().position(|vb| vb.id == *locked_to); match p { Some(p) => p, _ => default_vbranch_pos, } } else { default_vbranch_pos }; virtual_branches[vbranch_pos] .ownership .put(&OwnershipClaim { file_path: filepath.clone(), hunks: vec![Hunk::from(&hunk) .with_timestamp(get_mtime(&mut mtimes, &filepath)) .with_hash(Hunk::hash(hunk.diff.as_str()).as_str())], }); diffs_by_branch .entry(virtual_branches[vbranch_pos].id) .or_default() .entry(filepath.clone()) .or_default() .push(hunk.clone()); } } let mut hunks_by_branch = diffs_by_branch .into_iter() .map(|(branch_id, hunks)| { ( virtual_branches .iter() .find(|b| b.id.eq(&branch_id)) .unwrap() .clone(), hunks, ) }) .collect::>(); // write updated state if not resolving if !project_repository.is_resolving() { let branch_writer = branch::Writer::new( gb_repository, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), ) .context("failed to create writer")?; for (vbranch, files) in &mut hunks_by_branch { vbranch.tree = write_tree(project_repository, integration_commit, files)?; branch_writer .write(vbranch) .context(format!("failed to write virtual branch {}", vbranch.name))?; } } Ok((hunks_by_branch, skipped_files)) } fn virtual_hunks_to_virtual_files( project_repository: &project_repository::Repository, hunks: &[VirtualBranchHunk], ) -> Vec { hunks .iter() .fold(HashMap::>::new(), |mut acc, hunk| { acc.entry(hunk.file_path.clone()) .or_default() .push(hunk.clone()); acc }) .into_iter() .map(|(file_path, hunks)| VirtualBranchFile { id: file_path.display().to_string(), path: file_path.clone(), hunks: hunks.clone(), binary: hunks.iter().any(|h| h.binary), large: false, modified_at: hunks.iter().map(|h| h.modified_at).max().unwrap_or(0), conflicted: conflicts::is_conflicting( project_repository, Some(&file_path.display().to_string()), ) .unwrap_or(false), }) .collect::>() } // reset virtual branch to a specific commit pub fn reset_branch( gb_repository: &gb_repository::Repository, project_repository: &project_repository::Repository, branch_id: &BranchId, target_commit_oid: git::Oid, ) -> Result<(), errors::ResetBranchError> { let current_session = gb_repository.get_or_create_current_session()?; let current_session_reader = sessions::Reader::open(gb_repository, ¤t_session)?; let default_target = get_default_target(¤t_session_reader, project_repository.project()) .context("failed to read default target")? .ok_or_else(|| { errors::ResetBranchError::DefaultTargetNotSet(errors::DefaultTargetNotSet { project_id: project_repository.project().id, }) })?; let branch_reader = branch::Reader::new( ¤t_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ); let mut branch = match branch_reader.read(branch_id) { Ok(branch) => Ok(branch), Err(reader::Error::NotFound) => Err(errors::ResetBranchError::BranchNotFound( errors::BranchNotFound { branch_id: *branch_id, project_id: project_repository.project().id, }, )), Err(error) => Err(errors::ResetBranchError::Other(error.into())), }?; if branch.head == target_commit_oid { // nothing to do return Ok(()); } if default_target.sha != target_commit_oid && !project_repository .l(branch.head, LogUntil::Commit(default_target.sha))? .contains(&target_commit_oid) { return Err(errors::ResetBranchError::CommitNotFoundInBranch( target_commit_oid, )); } let branch_writer = branch::Writer::new( gb_repository, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), ) .context("failed to create writer")?; branch.head = target_commit_oid; branch_writer .write(&mut branch) .context("failed to write branch")?; super::integration::update_gitbutler_integration(gb_repository, project_repository) .context("failed to update gitbutler integration")?; Ok(()) } fn diffs_to_virtual_files( project_repository: &project_repository::Repository, diffs: &HashMap>, ) -> Vec { let hunks_by_filepath = virtual_hunks_by_filepath(&project_repository.project().path, diffs); virtual_hunks_to_virtual_files( project_repository, &hunks_by_filepath .values() .flatten() .cloned() .collect::>(), ) } // this function takes a list of file ownership, // constructs a tree from those changes on top of the target // and writes it as a new tree for storage pub fn write_tree( project_repository: &project_repository::Repository, target: &git::Oid, files: &HashMap>, ) -> Result { write_tree_onto_commit(project_repository, *target, files) } pub fn write_tree_onto_commit( project_repository: &project_repository::Repository, commit_oid: git::Oid, files: &HashMap>, ) -> Result { // read the base sha into an index let git_repository = &project_repository.git_repository; let head_commit = git_repository.find_commit(commit_oid)?; let base_tree = head_commit.tree()?; write_tree_onto_tree(project_repository, &base_tree, files) } pub fn write_tree_onto_tree( project_repository: &project_repository::Repository, base_tree: &git::Tree, files: &HashMap>, ) -> Result { let git_repository = &project_repository.git_repository; let mut builder = git_repository.treebuilder(Some(base_tree)); // now update the index with content in the working directory for each file for (filepath, hunks) in files { // convert this string to a Path let rel_path = Path::new(&filepath); let full_path = project_repository.path().join(rel_path); let is_submodule = full_path.is_dir() && hunks.len() == 1 && hunks[0].diff.contains("Subproject commit"); // if file exists if full_path.exists() { // if file is executable, use 755, otherwise 644 let mut filemode = git::FileMode::Blob; // check if full_path file is executable if let Ok(metadata) = std::fs::symlink_metadata(&full_path) { #[cfg(target_family = "unix")] { if metadata.permissions().mode() & 0o111 != 0 { filemode = git::FileMode::BlobExecutable; } } #[cfg(target_os = "windows")] { // TODO(qix-): Pull from `core.filemode` config option to determine // TODO(qix-): the behavior on windows. For now, we set this to true. // TODO(qix-): It's not ideal, but it gets us to a windows build faster. filemode = git::FileMode::BlobExecutable; } if metadata.file_type().is_symlink() { filemode = git::FileMode::Link; } } // get the blob if filemode == git::FileMode::Link { // it's a symlink, make the content the path of the link let link_target = std::fs::read_link(&full_path)?; // if the link target is inside the project repository, make it relative let link_target = link_target .strip_prefix(project_repository.path()) .unwrap_or(&link_target); let blob_oid = git_repository.blob( link_target .to_str() .ok_or_else(|| { anyhow!("path contains invalid utf-8 characters: {link_target:?}") })? .as_bytes(), )?; builder.upsert(rel_path, blob_oid, filemode); } else if let Ok(tree_entry) = base_tree.get_path(rel_path) { if hunks.len() == 1 && hunks[0].binary { let new_blob_oid = &hunks[0].diff; // convert string to Oid let new_blob_oid = new_blob_oid.parse().context("failed to diff as oid")?; builder.upsert(rel_path, new_blob_oid, filemode); } else { // blob from tree_entry let blob = tree_entry .to_object(git_repository) .unwrap() .peel_to_blob() .context("failed to get blob")?; let mut blob_contents = blob.content().to_str()?.to_string(); let mut hunks = hunks.clone(); hunks.sort_by_key(|hunk| hunk.new_start); let mut all_diffs = String::new(); for hunk in hunks { all_diffs.push_str(&hunk.diff); } let patch = Patch::from_str(&all_diffs)?; blob_contents = apply(&blob_contents, &patch).context(format!( "failed to apply\n{}\nonto:\n{}", &all_diffs, &blob_contents ))?; // create a blob let new_blob_oid = git_repository.blob(blob_contents.as_bytes())?; // upsert into the builder builder.upsert(rel_path, new_blob_oid, filemode); } } else if is_submodule { let mut blob_contents = String::new(); let mut hunks = hunks.clone(); hunks.sort_by_key(|hunk| hunk.new_start); for hunk in hunks { let patch = Patch::from_str(&hunk.diff)?; blob_contents = apply(&blob_contents, &patch) .context(format!("failed to apply {}", &hunk.diff))?; } // create a blob let new_blob_oid = git_repository.blob(blob_contents.as_bytes())?; // upsert into the builder builder.upsert(rel_path, new_blob_oid, filemode); } else { // create a git blob from a file on disk let blob_oid = git_repository .blob_path(&full_path) .context(format!("failed to create blob from path {:?}", &full_path))?; builder.upsert(rel_path, blob_oid, filemode); } } else if base_tree.get_path(rel_path).is_ok() { // remove file from index if it exists in the base tree builder.remove(rel_path); } else { // file not in index or base tree, do nothing // this is the } } // now write out the tree let tree_oid = builder.write().context("failed to write updated tree")?; Ok(tree_oid) } fn _print_tree(repo: &git2::Repository, tree: &git2::Tree) -> Result<()> { println!("tree id: {}", tree.id()); for entry in tree { println!( " entry: {} {}", entry.name().unwrap_or_default(), entry.id() ); // get entry contents let object = entry.to_object(repo).context("failed to get object")?; let blob = object.as_blob().context("failed to get blob")?; // convert content to string if let Ok(content) = std::str::from_utf8(blob.content()) { println!(" blob: {}", content); } else { println!(" blob: BINARY"); } } Ok(()) } #[allow(clippy::too_many_arguments)] pub fn commit( gb_repository: &gb_repository::Repository, project_repository: &project_repository::Repository, branch_id: &BranchId, message: &str, ownership: Option<&branch::BranchOwnershipClaims>, signing_key: Option<&keys::PrivateKey>, user: Option<&users::User>, run_hooks: bool, ) -> Result { let mut message_buffer = message.to_owned(); if run_hooks { let hook_result = project_repository .git_repository .run_hook_commit_msg(&mut message_buffer) .context("failed to run hook")?; if let HookResult::RunNotSuccessful { stdout, .. } = hook_result { return Err(errors::CommitError::CommitMsgHookRejected(stdout)); } let hook_result = project_repository .git_repository .run_hook_pre_commit() .context("failed to run hook")?; if let HookResult::RunNotSuccessful { stdout, .. } = hook_result { return Err(errors::CommitError::CommitHookRejected(stdout)); } } let message = &message_buffer; let integration_commit = super::integration::update_gitbutler_integration(gb_repository, project_repository)?; // get the files to commit let (mut statuses, _) = get_status_by_branch(gb_repository, project_repository, Some(&integration_commit)) .context("failed to get status by branch")?; let (ref mut branch, files) = statuses .iter_mut() .find(|(branch, _)| branch.id == *branch_id) .ok_or_else(|| { errors::CommitError::BranchNotFound(errors::BranchNotFound { project_id: project_repository.project().id, branch_id: *branch_id, }) })?; update_conflict_markers(project_repository, files)?; if conflicts::is_conflicting::<&Path>(project_repository, None)? { return Err(errors::CommitError::Conflicted(errors::ProjectConflict { project_id: project_repository.project().id, })); } let tree_oid = if let Some(ownership) = ownership { let files = files .iter() .filter_map(|(filepath, hunks)| { let hunks = hunks .iter() .filter(|hunk| { ownership .claims .iter() .find(|f| f.file_path.eq(filepath)) .map_or(false, |f| { f.hunks.iter().any(|h| { h.start == hunk.new_start && h.end == hunk.new_start + hunk.new_lines }) }) }) .cloned() .collect::>(); if hunks.is_empty() { None } else { Some((filepath.clone(), hunks)) } }) .collect::>(); write_tree_onto_commit(project_repository, branch.head, &files)? } else { write_tree_onto_commit(project_repository, branch.head, files)? }; let git_repository = &project_repository.git_repository; let parent_commit = git_repository .find_commit(branch.head) .context(format!("failed to find commit {:?}", branch.head))?; let tree = git_repository .find_tree(tree_oid) .context(format!("failed to find tree {:?}", tree_oid))?; // now write a commit, using a merge parent if it exists let extra_merge_parent = conflicts::merge_parent(project_repository).context("failed to get merge parent")?; let commit_oid = match extra_merge_parent { Some(merge_parent) => { let merge_parent = git_repository .find_commit(merge_parent) .context(format!("failed to find merge parent {:?}", merge_parent))?; let commit_oid = project_repository.commit( user, message, &tree, &[&parent_commit, &merge_parent], signing_key, )?; conflicts::clear(project_repository).context("failed to clear conflicts")?; commit_oid } None => project_repository.commit(user, message, &tree, &[&parent_commit], signing_key)?, }; if run_hooks { project_repository .git_repository .run_hook_post_commit() .context("failed to run hook")?; } // update the virtual branch head let writer = branch::Writer::new( gb_repository, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), ) .context("failed to create writer")?; branch.tree = tree_oid; branch.head = commit_oid; writer.write(branch).context("failed to write branch")?; super::integration::update_gitbutler_integration(gb_repository, project_repository) .context("failed to update gitbutler integration")?; Ok(commit_oid) } pub fn push( project_repository: &project_repository::Repository, gb_repository: &gb_repository::Repository, branch_id: &BranchId, with_force: bool, credentials: &git::credentials::Helper, askpass: Option<(AskpassBroker, Option)>, ) -> Result<(), errors::PushError> { let current_session = gb_repository .get_or_create_current_session() .context("failed to get or create currnt session") .map_err(errors::PushError::Other)?; let current_session_reader = sessions::Reader::open(gb_repository, ¤t_session) .context("failed to open current session") .map_err(errors::PushError::Other)?; let branch_reader = branch::Reader::new( ¤t_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ); let branch_writer = branch::Writer::new( gb_repository, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), ) .context("failed to create writer")?; let mut vbranch = branch_reader.read(branch_id).map_err(|error| match error { reader::Error::NotFound => errors::PushError::BranchNotFound(errors::BranchNotFound { project_id: project_repository.project().id, branch_id: *branch_id, }), error => errors::PushError::Other(error.into()), })?; let remote_branch = if let Some(upstream_branch) = vbranch.upstream.as_ref() { upstream_branch.clone() } else { let default_target = get_default_target(¤t_session_reader, project_repository.project()) .context("failed to get default target")? .ok_or_else(|| { errors::PushError::DefaultTargetNotSet(errors::DefaultTargetNotSet { project_id: project_repository.project().id, }) })?; let remote_branch = format!( "refs/remotes/{}/{}", default_target.branch.remote(), normalize_branch_name(&vbranch.name) ) .parse::() .context("failed to parse remote branch name")?; let remote_branches = project_repository.git_remote_branches()?; let existing_branches = remote_branches .iter() .map(RemoteRefname::branch) .map(str::to_lowercase) // git is weird about case sensitivity here, assume not case sensitive .collect::>(); remote_branch.with_branch(&dedup_fmt( &existing_branches .iter() .map(String::as_str) .collect::>(), remote_branch.branch(), "-", )) }; project_repository.push( &vbranch.head, &remote_branch, with_force, credentials, None, askpass.clone(), )?; vbranch.upstream = Some(remote_branch.clone()); vbranch.upstream_head = Some(vbranch.head); branch_writer .write(&mut vbranch) .context("failed to write target branch after push")?; project_repository.fetch( remote_branch.remote(), credentials, askpass.map(|(broker, _)| (broker, "modal".to_string())), )?; Ok(()) } fn is_commit_integrated( project_repository: &project_repository::Repository, target: &target::Target, commit: &git::Commit, ) -> Result { let remote_branch = project_repository .git_repository .find_branch(&target.branch.clone().into())?; let remote_head = remote_branch.peel_to_commit()?; let upstream_commits = project_repository.l( remote_head.id(), project_repository::LogUntil::Commit(target.sha), )?; if target.sha.eq(&commit.id()) { // could not be integrated if heads are the same. return Ok(false); } if upstream_commits.is_empty() { // could not be integrated - there is nothing new upstream. return Ok(false); } if upstream_commits.contains(&commit.id()) { return Ok(true); } let merge_base_id = project_repository .git_repository .merge_base(target.sha, commit.id())?; if merge_base_id.eq(&commit.id()) { // if merge branch is the same as branch head and there are upstream commits // then it's integrated return Ok(true); } let merge_base = project_repository .git_repository .find_commit(merge_base_id)?; let merge_base_tree = merge_base.tree()?; let upstream = project_repository .git_repository .find_commit(remote_head.id())?; let upstream_tree = upstream.tree()?; if merge_base_tree.id() == upstream_tree.id() { // if merge base is the same as upstream tree, then it's integrated return Ok(true); } // try to merge our tree into the upstream tree let mut merge_index = project_repository .git_repository .merge_trees(&merge_base_tree, &commit.tree()?, &upstream_tree) .context("failed to merge trees")?; if merge_index.has_conflicts() { return Ok(false); } let merge_tree_oid = merge_index .write_tree_to(&project_repository.git_repository) .context("failed to write tree")?; // if the merge_tree is the same as the new_target_tree and there are no files (uncommitted changes) // then the vbranch is fully merged Ok(merge_tree_oid == upstream_tree.id()) } pub fn is_remote_branch_mergeable( gb_repository: &gb_repository::Repository, project_repository: &project_repository::Repository, branch_name: &git::RemoteRefname, ) -> Result { // get the current target let latest_session = gb_repository.get_latest_session()?.ok_or_else(|| { errors::IsRemoteBranchMergableError::DefaultTargetNotSet(errors::DefaultTargetNotSet { project_id: project_repository.project().id, }) })?; let session_reader = sessions::Reader::open(gb_repository, &latest_session) .context("failed to open current session")?; let default_target = get_default_target(&session_reader, project_repository.project()) .context("failed to get default target")? .ok_or_else(|| { errors::IsRemoteBranchMergableError::DefaultTargetNotSet(errors::DefaultTargetNotSet { project_id: project_repository.project().id, }) })?; let target_commit = project_repository .git_repository .find_commit(default_target.sha) .context("failed to find target commit")?; let branch = match project_repository .git_repository .find_branch(&branch_name.into()) { Ok(branch) => Ok(branch), Err(git::Error::NotFound(_)) => Err(errors::IsRemoteBranchMergableError::BranchNotFound( branch_name.clone(), )), Err(error) => Err(errors::IsRemoteBranchMergableError::Other(error.into())), }?; let branch_oid = branch.target().context("detatched head")?; let branch_commit = project_repository .git_repository .find_commit(branch_oid) .context("failed to find branch commit")?; let base_tree = find_base_tree( &project_repository.git_repository, &branch_commit, &target_commit, )?; let wd_tree = project_repository.get_wd_tree()?; let branch_tree = branch_commit.tree().context("failed to find branch tree")?; let mergeable = !project_repository .git_repository .merge_trees(&base_tree, &branch_tree, &wd_tree) .context("failed to merge trees")? .has_conflicts(); Ok(mergeable) } pub fn is_virtual_branch_mergeable( gb_repository: &gb_repository::Repository, project_repository: &project_repository::Repository, branch_id: &BranchId, ) -> Result { let latest_session = gb_repository.get_latest_session()?.ok_or_else(|| { errors::IsVirtualBranchMergeable::DefaultTargetNotSet(errors::DefaultTargetNotSet { project_id: project_repository.project().id, }) })?; let session_reader = sessions::Reader::open(gb_repository, &latest_session) .context("failed to open current session reader")?; let branch_reader = branch::Reader::new( &session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ); let branch = match branch_reader.read(branch_id) { Ok(branch) => Ok(branch), Err(reader::Error::NotFound) => Err(errors::IsVirtualBranchMergeable::BranchNotFound( errors::BranchNotFound { project_id: project_repository.project().id, branch_id: *branch_id, }, )), Err(error) => Err(errors::IsVirtualBranchMergeable::Other(error.into())), }?; if branch.applied { return Ok(true); } let default_target = get_default_target(&session_reader, project_repository.project()) .context("failed to read default target")? .ok_or_else(|| { errors::IsVirtualBranchMergeable::DefaultTargetNotSet(errors::DefaultTargetNotSet { project_id: project_repository.project().id, }) })?; // determine if this branch is up to date with the target/base let merge_base = project_repository .git_repository .merge_base(default_target.sha, branch.head) .context("failed to find merge base")?; if merge_base != default_target.sha { return Ok(false); } let branch_commit = project_repository .git_repository .find_commit(branch.head) .context("failed to find branch commit")?; let target_commit = project_repository .git_repository .find_commit(default_target.sha) .context("failed to find target commit")?; let base_tree = find_base_tree( &project_repository.git_repository, &branch_commit, &target_commit, )?; let wd_tree = project_repository.get_wd_tree()?; // determine if this tree is mergeable let branch_tree = project_repository .git_repository .find_tree(branch.tree) .context("failed to find branch tree")?; let is_mergeable = !project_repository .git_repository .merge_trees(&base_tree, &branch_tree, &wd_tree) .context("failed to merge trees")? .has_conflicts(); Ok(is_mergeable) } pub fn amend( gb_repository: &gb_repository::Repository, project_repository: &project_repository::Repository, branch_id: &BranchId, target_ownership: &BranchOwnershipClaims, ) -> Result { if conflicts::is_conflicting::<&Path>(project_repository, None)? { return Err(errors::AmendError::Conflict(errors::ProjectConflict { project_id: project_repository.project().id, })); } let current_session = gb_repository .get_or_create_current_session() .context("failed to get or create current session")?; let current_session_reader = sessions::Reader::open(gb_repository, ¤t_session) .context("failed to open current session")?; let all_branches = Iterator::new( ¤t_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ) .context("failed to create branch iterator")? .collect::, reader::Error>>() .context("failed to read virtual branches")? .into_iter() .collect::>(); if !all_branches.iter().any(|b| b.id == *branch_id) { return Err(errors::AmendError::BranchNotFound(errors::BranchNotFound { project_id: project_repository.project().id, branch_id: *branch_id, })); } let applied_branches = all_branches .into_iter() .filter(|b| b.applied) .collect::>(); if !applied_branches.iter().any(|b| b.id == *branch_id) { return Err(errors::AmendError::BranchNotFound(errors::BranchNotFound { project_id: project_repository.project().id, branch_id: *branch_id, })); } let default_target = get_default_target(¤t_session_reader, project_repository.project()) .context("failed to read default target")? .ok_or_else(|| { errors::AmendError::DefaultTargetNotSet(errors::DefaultTargetNotSet { project_id: project_repository.project().id, }) })?; let integration_commit = super::integration::update_gitbutler_integration(gb_repository, project_repository)?; let (mut applied_statuses, _) = get_applied_status( gb_repository, project_repository, &integration_commit, &default_target.sha, applied_branches, )?; let (ref mut target_branch, target_status) = applied_statuses .iter_mut() .find(|(b, _)| b.id == *branch_id) .ok_or_else(|| { errors::AmendError::BranchNotFound(errors::BranchNotFound { project_id: project_repository.project().id, branch_id: *branch_id, }) })?; if target_branch.upstream.is_some() && !project_repository.project().ok_with_force_push { // amending to a pushed head commit will cause a force push that is not allowed return Err(errors::AmendError::ForcePushNotAllowed( errors::ForcePushNotAllowed { project_id: project_repository.project().id, }, )); } if project_repository .l( target_branch.head, project_repository::LogUntil::Commit(default_target.sha), )? .is_empty() { return Err(errors::AmendError::BranchHasNoCommits); } let head_commit = project_repository .git_repository .find_commit(target_branch.head) .context("failed to find head commit")?; let diffs_to_amend = target_ownership .claims .iter() .filter_map(|file_ownership| { let hunks = target_status .get(&file_ownership.file_path) .map(|hunks| { hunks .iter() .filter(|hunk| { file_ownership.hunks.iter().any(|owned_hunk| { owned_hunk.start == hunk.new_start && owned_hunk.end == hunk.new_start + hunk.new_lines }) }) .cloned() .collect::>() }) .unwrap_or_default(); if hunks.is_empty() { None } else { Some((file_ownership.file_path.clone(), hunks)) } }) .collect::>(); if diffs_to_amend.is_empty() { return Err(errors::AmendError::TargetOwnerhshipNotFound( target_ownership.clone(), )); } let new_tree_oid = write_tree_onto_commit(project_repository, target_branch.head, &diffs_to_amend)?; let new_tree = project_repository .git_repository .find_tree(new_tree_oid) .context("failed to find new tree")?; let parents = head_commit .parents() .context("failed to find head commit parents")?; let commit_oid = project_repository .git_repository .commit( None, &head_commit.author(), &head_commit.committer(), head_commit.message().unwrap_or_default(), &new_tree, &parents.iter().collect::>(), ) .context("failed to create commit")?; let branch_writer = branch::Writer::new( gb_repository, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), ) .context("failed to create writer")?; target_branch.head = commit_oid; branch_writer.write(target_branch)?; super::integration::update_gitbutler_integration(gb_repository, project_repository)?; Ok(commit_oid) } pub fn cherry_pick( gb_repository: &gb_repository::Repository, project_repository: &project_repository::Repository, branch_id: &BranchId, target_commit_oid: git::Oid, ) -> Result, errors::CherryPickError> { if conflicts::is_conflicting::<&Path>(project_repository, None)? { return Err(errors::CherryPickError::Conflict(errors::ProjectConflict { project_id: project_repository.project().id, })); } let current_session = gb_repository .get_or_create_current_session() .context("failed to get or create current session")?; let current_session_reader = sessions::Reader::open(gb_repository, ¤t_session) .context("failed to open current session")?; let branch_reader = branch::Reader::new( ¤t_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ); let mut branch = branch_reader .read(branch_id) .context("failed to read branch")?; if !branch.applied { // todo? return Err(errors::CherryPickError::NotApplied); } let target_commit = project_repository .git_repository .find_commit(target_commit_oid) .map_err(|error| match error { git::Error::NotFound(_) => errors::CherryPickError::CommitNotFound(target_commit_oid), error => errors::CherryPickError::Other(error.into()), })?; let branch_head_commit = project_repository .git_repository .find_commit(branch.head) .context("failed to find branch tree")?; let default_target = get_default_target(¤t_session_reader, project_repository.project()) .context("failed to read default target")? .context("no default target set")?; // if any other branches are applied, unapply them let applied_branches = Iterator::new( ¤t_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ) .context("failed to create branch iterator")? .collect::, reader::Error>>() .context("failed to read virtual branches")? .into_iter() .filter(|b| b.applied) .collect::>(); let integration_commit = super::integration::update_gitbutler_integration(gb_repository, project_repository)?; let (applied_statuses, _) = get_applied_status( gb_repository, project_repository, &integration_commit, &default_target.sha, applied_branches, )?; let branch_files = applied_statuses .iter() .find(|(b, _)| b.id == *branch_id) .map(|(_, f)| f) .context("branch status not found")?; // create a wip commit. we'll use it to offload cherrypick conflicts calculation to libgit. let wip_commit = { let wip_tree_oid = write_tree(project_repository, &branch.head, branch_files)?; let wip_tree = project_repository .git_repository .find_tree(wip_tree_oid) .context("failed to find tree")?; let signature = git::Signature::now("GitButler", "gitbutler@gitbutler.com") .context("failed to make gb signature")?; let oid = project_repository .git_repository .commit( None, &signature, &signature, "wip cherry picking commit", &wip_tree, &[&branch_head_commit], ) .context("failed to commit wip work")?; project_repository .git_repository .find_commit(oid) .context("failed to find wip commit")? }; let mut cherrypick_index = project_repository .git_repository .cherry_pick(&wip_commit, &target_commit) .context("failed to cherry pick")?; // unapply other branches for other_branch in applied_statuses .iter() .filter(|(b, _)| b.id != branch.id) .map(|(b, _)| b) { unapply_branch(gb_repository, project_repository, &other_branch.id) .context("failed to unapply branch")?; } let commit_oid = if cherrypick_index.has_conflicts() { // checkout the conflicts project_repository .git_repository .checkout_index(&mut cherrypick_index) .allow_conflicts() .conflict_style_merge() .force() .checkout() .context("failed to checkout conflicts")?; // mark conflicts let conflicts = cherrypick_index .conflicts() .context("failed to get conflicts")?; let mut merge_conflicts = Vec::new(); for path in conflicts.flatten() { if let Some(ours) = path.our { let path = std::str::from_utf8(&ours.path) .context("failed to convert path")? .to_string(); merge_conflicts.push(path); } } conflicts::mark(project_repository, &merge_conflicts, Some(branch.head))?; None } else { let merge_tree_oid = cherrypick_index .write_tree_to(&project_repository.git_repository) .context("failed to write merge tree")?; let merge_tree = project_repository .git_repository .find_tree(merge_tree_oid) .context("failed to find merge tree")?; let branch_head_commit = project_repository .git_repository .find_commit(branch.head) .context("failed to find branch head commit")?; let commit_oid = project_repository .git_repository .commit( None, &target_commit.author(), &target_commit.committer(), target_commit.message().unwrap_or_default(), &merge_tree, &[&branch_head_commit], ) .context("failed to create commit")?; // checkout final_tree into the working directory project_repository .git_repository .checkout_tree(&merge_tree) .force() .remove_untracked() .checkout() .context("failed to checkout final tree")?; // update branch status let writer = branch::Writer::new( gb_repository, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), ) .context("failed to create writer")?; branch.head = commit_oid; writer .write(&mut branch) .context("failed to write branch")?; Some(commit_oid) }; super::integration::update_gitbutler_integration(gb_repository, project_repository) .context("failed to update gitbutler integration")?; Ok(commit_oid) } /// squashes a commit from a virtual branch into it's parent. pub fn squash( gb_repository: &gb_repository::Repository, project_repository: &project_repository::Repository, branch_id: &BranchId, commit_oid: git::Oid, ) -> Result<(), errors::SquashError> { if conflicts::is_conflicting::<&Path>(project_repository, None)? { return Err(errors::SquashError::Conflict(errors::ProjectConflict { project_id: project_repository.project().id, })); } let current_session = gb_repository .get_or_create_current_session() .context("failed to get or create current session")?; let current_session_reader = sessions::Reader::open(gb_repository, ¤t_session) .context("failed to open current session")?; let branch_reader = branch::Reader::new( ¤t_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ); let default_target = get_default_target(¤t_session_reader, project_repository.project()) .context("failed to read default target")? .ok_or_else(|| { errors::SquashError::DefaultTargetNotSet(errors::DefaultTargetNotSet { project_id: project_repository.project().id, }) })?; let mut branch = branch_reader.read(branch_id).map_err(|error| match error { reader::Error::NotFound => errors::SquashError::BranchNotFound(errors::BranchNotFound { project_id: project_repository.project().id, branch_id: *branch_id, }), error => errors::SquashError::Other(error.into()), })?; let branch_commit_oids = project_repository.l( branch.head, project_repository::LogUntil::Commit(default_target.sha), )?; if !branch_commit_oids.contains(&commit_oid) { return Err(errors::SquashError::CommitNotFound(commit_oid)); } let commit_to_squash = project_repository .git_repository .find_commit(commit_oid) .context("failed to find commit")?; let parent_commit = commit_to_squash .parent(0) .context("failed to find parent commit")?; let pushed_commit_oids = branch.upstream_head.map_or_else( || Ok(vec![]), |upstream_head| { project_repository.l( upstream_head, project_repository::LogUntil::Commit(default_target.sha), ) }, )?; if pushed_commit_oids.contains(&parent_commit.id()) && !project_repository.project().ok_with_force_push { // squashing into a pushed commit will cause a force push that is not allowed return Err(errors::SquashError::ForcePushNotAllowed( errors::ForcePushNotAllowed { project_id: project_repository.project().id, }, )); } if !branch_commit_oids.contains(&parent_commit.id()) { return Err(errors::SquashError::CantSquashRootCommit); } let ids_to_rebase = { let ids = branch_commit_oids .split(|oid| oid.eq(&commit_oid)) .collect::>(); ids.first().copied() }; // create a commit that: // * has the tree of the target commit // * has the message combined of the target commit and parent commit // * has parents of the parents commit. let parents = parent_commit .parents() .context("failed to find head commit parents")?; let new_commit_oid = project_repository .git_repository .commit( None, &commit_to_squash.author(), &commit_to_squash.committer(), &format!( "{}\n{}", parent_commit.message().unwrap_or_default(), commit_to_squash.message().unwrap_or_default(), ), &commit_to_squash.tree().context("failed to find tree")?, &parents.iter().collect::>(), ) .context("failed to commit")?; let new_head_id = if let Some(ids_to_rebase) = ids_to_rebase { let mut ids_to_rebase = ids_to_rebase.to_vec(); ids_to_rebase.reverse(); // now, rebase unchanged commits onto the new commit let commits_to_rebase = ids_to_rebase .iter() .map(|oid| project_repository.git_repository.find_commit(*oid)) .collect::, _>>() .context("failed to read commits to rebase")?; commits_to_rebase .into_iter() .fold( project_repository .git_repository .find_commit(new_commit_oid) .context("failed to find new commit"), |head, to_rebase| { let head = head?; let mut cherrypick_index = project_repository .git_repository .cherry_pick(&head, &to_rebase) .context("failed to cherry pick")?; if cherrypick_index.has_conflicts() { bail!("failed to rebase"); } let merge_tree_oid = cherrypick_index .write_tree_to(&project_repository.git_repository) .context("failed to write merge tree")?; let merge_tree = project_repository .git_repository .find_tree(merge_tree_oid) .context("failed to find merge tree")?; let commit_oid = project_repository .git_repository .commit( None, &to_rebase.author(), &to_rebase.committer(), to_rebase.message().unwrap_or_default(), &merge_tree, &[&head], ) .context("failed to create commit")?; project_repository .git_repository .find_commit(commit_oid) .context("failed to find commit") }, )? .id() } else { new_commit_oid }; // save new branch head let writer = branch::Writer::new( gb_repository, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), ) .context("failed to create writer")?; branch.head = new_head_id; writer .write(&mut branch) .context("failed to write branch")?; super::integration::update_gitbutler_integration(gb_repository, project_repository)?; Ok(()) } pub fn update_commit_message( gb_repository: &gb_repository::Repository, project_repository: &project_repository::Repository, branch_id: &BranchId, commit_oid: git::Oid, message: &str, ) -> Result<(), errors::UpdateCommitMessageError> { if message.is_empty() { return Err(errors::UpdateCommitMessageError::EmptyMessage); } if conflicts::is_conflicting::<&Path>(project_repository, None)? { return Err(errors::UpdateCommitMessageError::Conflict( errors::ProjectConflict { project_id: project_repository.project().id, }, )); } let current_session = gb_repository .get_or_create_current_session() .context("failed to get or create current session")?; let current_session_reader = sessions::Reader::open(gb_repository, ¤t_session) .context("failed to open current session")?; let branch_reader = branch::Reader::new( ¤t_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ); let default_target = get_default_target(¤t_session_reader, project_repository.project()) .context("failed to read default target")? .ok_or_else(|| { errors::UpdateCommitMessageError::DefaultTargetNotSet(errors::DefaultTargetNotSet { project_id: project_repository.project().id, }) })?; let mut branch = branch_reader.read(branch_id).map_err(|error| match error { reader::Error::NotFound => { errors::UpdateCommitMessageError::BranchNotFound(errors::BranchNotFound { project_id: project_repository.project().id, branch_id: *branch_id, }) } error => errors::UpdateCommitMessageError::Other(error.into()), })?; let branch_commit_oids = project_repository.l( branch.head, project_repository::LogUntil::Commit(default_target.sha), )?; if !branch_commit_oids.contains(&commit_oid) { return Err(errors::UpdateCommitMessageError::CommitNotFound(commit_oid)); } let pushed_commit_oids = branch.upstream_head.map_or_else( || Ok(vec![]), |upstream_head| { project_repository.l( upstream_head, project_repository::LogUntil::Commit(default_target.sha), ) }, )?; if pushed_commit_oids.contains(&commit_oid) && !project_repository.project().ok_with_force_push { // updating the message of a pushed commit will cause a force push that is not allowed return Err(errors::UpdateCommitMessageError::ForcePushNotAllowed( errors::ForcePushNotAllowed { project_id: project_repository.project().id, }, )); } let target_commit = project_repository .git_repository .find_commit(commit_oid) .context("failed to find commit")?; let ids_to_rebase = { let ids = branch_commit_oids .split(|oid| oid.eq(&commit_oid)) .collect::>(); ids.first().copied() }; let parents = target_commit .parents() .context("failed to find head commit parents")?; let new_commit_oid = project_repository .git_repository .commit( None, &target_commit.author(), &target_commit.committer(), message, &target_commit.tree().context("failed to find tree")?, &parents.iter().collect::>(), ) .context("failed to commit")?; let new_head_id = if let Some(ids_to_rebase) = ids_to_rebase { let mut ids_to_rebase = ids_to_rebase.to_vec(); ids_to_rebase.reverse(); // now, rebase unchanged commits onto the new commit let commits_to_rebase = ids_to_rebase .iter() .map(|oid| project_repository.git_repository.find_commit(*oid)) .collect::, _>>() .context("failed to read commits to rebase")?; commits_to_rebase .into_iter() .fold( project_repository .git_repository .find_commit(new_commit_oid) .context("failed to find new commit"), |head, to_rebase| { let head = head?; let mut cherrypick_index = project_repository .git_repository .cherry_pick(&head, &to_rebase) .context("failed to cherry pick")?; if cherrypick_index.has_conflicts() { bail!("failed to rebase"); } let merge_tree_oid = cherrypick_index .write_tree_to(&project_repository.git_repository) .context("failed to write merge tree")?; let merge_tree = project_repository .git_repository .find_tree(merge_tree_oid) .context("failed to find merge tree")?; let commit_oid = project_repository .git_repository .commit( None, &to_rebase.author(), &to_rebase.committer(), to_rebase.message().unwrap_or_default(), &merge_tree, &[&head], ) .context("failed to create commit")?; project_repository .git_repository .find_commit(commit_oid) .context("failed to find commit") }, )? .id() } else { new_commit_oid }; // save new branch head let writer = branch::Writer::new( gb_repository, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), ) .context("failed to create writer")?; branch.head = new_head_id; writer .write(&mut branch) .context("failed to write branch")?; super::integration::update_gitbutler_integration(gb_repository, project_repository)?; Ok(()) } /// moves commit on top of the to target branch pub fn move_commit( gb_repository: &gb_repository::Repository, project_repository: &project_repository::Repository, target_branch_id: &BranchId, commit_oid: git::Oid, user: Option<&users::User>, signing_key: Option<&keys::PrivateKey>, ) -> Result<(), errors::MoveCommitError> { if project_repository.is_resolving() { return Err(errors::MoveCommitError::Conflicted( errors::ProjectConflict { project_id: project_repository.project().id, }, )); } let latest_session = gb_repository .get_latest_session() .context("failed to get or create current session")? .ok_or_else(|| { errors::MoveCommitError::DefaultTargetNotSet(errors::DefaultTargetNotSet { project_id: project_repository.project().id, }) })?; let latest_session_reader = sessions::Reader::open(gb_repository, &latest_session) .context("failed to open current session")?; let applied_branches = Iterator::new( &latest_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ) .context("failed to create branch iterator")? .collect::, reader::Error>>() .context("failed to read virtual branches")? .into_iter() .filter(|b| b.applied) .collect::>(); if !applied_branches.iter().any(|b| b.id == *target_branch_id) { return Err(errors::MoveCommitError::BranchNotFound( errors::BranchNotFound { project_id: project_repository.project().id, branch_id: *target_branch_id, }, )); } let default_target = get_default_target(&latest_session_reader, project_repository.project()) .context("failed to get default target")? .ok_or_else(|| { errors::MoveCommitError::DefaultTargetNotSet(errors::DefaultTargetNotSet { project_id: project_repository.project().id, }) })?; let integration_commit = super::integration::update_gitbutler_integration(gb_repository, project_repository)?; let (mut applied_statuses, _) = get_applied_status( gb_repository, project_repository, &integration_commit, &default_target.sha, applied_branches, )?; let (ref mut source_branch, source_status) = applied_statuses .iter_mut() .find(|(b, _)| b.head == commit_oid) .ok_or_else(|| errors::MoveCommitError::CommitNotFound(commit_oid))?; let source_branch_non_comitted_files = source_status; let source_branch_head = project_repository .git_repository .find_commit(commit_oid) .context("failed to find commit")?; let source_branch_head_parent = source_branch_head .parent(0) .context("failed to get parent commit")?; let source_branch_head_tree = source_branch_head .tree() .context("failed to get commit tree")?; let source_branch_head_parent_tree = source_branch_head_parent .tree() .context("failed to get parent tree")?; let branch_head_diff = diff::trees( &project_repository.git_repository, &source_branch_head_parent_tree, &source_branch_head_tree, 3, )?; let branch_head_diff = diff::diff_files_to_hunks(&branch_head_diff); let is_source_locked = source_branch_non_comitted_files .iter() .any(|(path, hunks)| { branch_head_diff.get(path).map_or(false, |head_diff_hunks| { hunks.iter().any(|hunk| { head_diff_hunks.iter().any(|head_hunk| { joined( head_hunk.new_start, head_hunk.new_start + head_hunk.new_lines, hunk.new_start, hunk.new_start + hunk.new_lines, ) }) }) }) }); if is_source_locked { return Err(errors::MoveCommitError::SourceLocked); } let branch_writer = branch::Writer::new( gb_repository, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), ) .context("failed to create writer")?; let branch_reader = branch::Reader::new( &latest_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ); // move files ownerships from source branch to the destination branch let ownerships_to_transfer = branch_head_diff .iter() .map(|(file_path, hunks)| { ( file_path.clone(), hunks.iter().map(Into::into).collect::>(), ) }) .map(|(file_path, hunks)| OwnershipClaim { file_path, hunks }) .flat_map(|file_ownership| source_branch.ownership.take(&file_ownership)) .collect::>(); // reset the source branch to the parent commit { source_branch.head = source_branch_head_parent.id(); branch_writer.write(source_branch)?; } // move the commit to destination branch target branch { let mut destination_branch = branch_reader .read(target_branch_id) .map_err(|error| match error { reader::Error::NotFound => { errors::MoveCommitError::BranchNotFound(errors::BranchNotFound { project_id: project_repository.project().id, branch_id: *target_branch_id, }) } error => errors::MoveCommitError::Other(error.into()), })?; for ownership in ownerships_to_transfer { destination_branch.ownership.put(&ownership); } let new_destination_tree_oid = write_tree_onto_commit( project_repository, destination_branch.head, &branch_head_diff, ) .context("failed to write tree onto commit")?; let new_destination_tree = project_repository .git_repository .find_tree(new_destination_tree_oid) .context("failed to find tree")?; let new_destination_head_oid = project_repository .commit( user, source_branch_head.message().unwrap_or_default(), &new_destination_tree, &[&project_repository .git_repository .find_commit(destination_branch.head) .context("failed to get dst branch head commit")?], signing_key, ) .context("failed to commit")?; destination_branch.head = new_destination_head_oid; branch_writer.write(&mut destination_branch)?; } super::integration::update_gitbutler_integration(gb_repository, project_repository) .context("failed to update gitbutler integration")?; Ok(()) } pub fn create_virtual_branch_from_branch( gb_repository: &gb_repository::Repository, project_repository: &project_repository::Repository, upstream: &git::Refname, signing_key: Option<&keys::PrivateKey>, user: Option<&users::User>, ) -> Result { if !matches!(upstream, git::Refname::Local(_) | git::Refname::Remote(_)) { return Err(errors::CreateVirtualBranchFromBranchError::BranchNotFound( upstream.clone(), )); } let current_session = gb_repository .get_or_create_current_session() .context("failed to get or create current session")?; let current_session_reader = sessions::Reader::open(gb_repository, ¤t_session) .context("failed to open current session")?; let default_target = get_default_target(¤t_session_reader, project_repository.project()) .context("failed to get default target")? .ok_or_else(|| { errors::CreateVirtualBranchFromBranchError::DefaultTargetNotSet( errors::DefaultTargetNotSet { project_id: project_repository.project().id, }, ) })?; if let git::Refname::Remote(remote_upstream) = upstream { if default_target.branch.eq(remote_upstream) { return Err( errors::CreateVirtualBranchFromBranchError::CantMakeBranchFromDefaultTarget, ); } } let repo = &project_repository.git_repository; let head_reference = match repo.find_reference(upstream) { Ok(head) => Ok(head), Err(git::Error::NotFound(_)) => Err( errors::CreateVirtualBranchFromBranchError::BranchNotFound(upstream.clone()), ), Err(error) => Err(errors::CreateVirtualBranchFromBranchError::Other( error.into(), )), }?; let head_commit = head_reference .peel_to_commit() .context("failed to peel to commit")?; let head_commit_tree = head_commit.tree().context("failed to find tree")?; let all_virtual_branches = Iterator::new( ¤t_session_reader, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), project_repository.project().use_toml_vbranches_state(), ) .context("failed to create branch iterator")? .collect::, reader::Error>>() .context("failed to read virtual branches")? .into_iter() .collect::>(); let order = all_virtual_branches.len(); let selected_for_changes = (!all_virtual_branches .iter() .any(|b| b.selected_for_changes.is_some())) .then_some(chrono::Utc::now().timestamp_millis()); let now = time::UNIX_EPOCH .elapsed() .context("failed to get elapsed time")? .as_millis(); // only set upstream if it's not the default target let upstream_branch = match upstream { git::Refname::Other(_) | git::Refname::Virtual(_) => { // we only support local or remote branches return Err(errors::CreateVirtualBranchFromBranchError::BranchNotFound( upstream.clone(), )); } git::Refname::Remote(remote) => Some(remote.clone()), git::Refname::Local(local) => local.remote().cloned(), }; // add file ownership based off the diff let target_commit = repo .find_commit(default_target.sha) .map_err(|error| errors::CreateVirtualBranchFromBranchError::Other(error.into()))?; let merge_base_oid = repo .merge_base(target_commit.id(), head_commit.id()) .map_err(|error| errors::CreateVirtualBranchFromBranchError::Other(error.into()))?; let merge_base_tree = repo .find_commit(merge_base_oid) .map_err(|error| errors::CreateVirtualBranchFromBranchError::Other(error.into()))? .tree() .map_err(|error| errors::CreateVirtualBranchFromBranchError::Other(error.into()))?; // do a diff between the head of this branch and the target base let diff = diff::trees( &project_repository.git_repository, &merge_base_tree, &head_commit_tree, 3, ) .context("failed to diff trees")?; let diff = diff::diff_files_to_hunks(&diff); let hunks_by_filepath = super::virtual_hunks_by_filepath(&project_repository.project().path, &diff); // assign ownership to the branch let ownership = hunks_by_filepath.values().flatten().fold( branch::BranchOwnershipClaims::default(), |mut ownership, hunk| { ownership.put( &format!("{}:{}", hunk.file_path.display(), hunk.id) .parse() .unwrap(), ); ownership }, ); let mut branch = branch::Branch { id: BranchId::generate(), name: upstream .branch() .expect("always a branch reference") .to_string(), notes: String::new(), applied: false, upstream_head: upstream_branch.is_some().then_some(head_commit.id()), upstream: upstream_branch, tree: head_commit_tree.id(), head: head_commit.id(), created_timestamp_ms: now, updated_timestamp_ms: now, ownership, order, selected_for_changes, }; let writer = branch::Writer::new( gb_repository, VirtualBranchesHandle::new(&project_repository.project().gb_dir()), ) .context("failed to create writer")?; writer .write(&mut branch) .context("failed to write branch")?; project_repository.add_branch_reference(&branch)?; match apply_branch( gb_repository, project_repository, &branch.id, signing_key, user, ) { Ok(()) => Ok(branch.id), Err(errors::ApplyBranchError::BranchConflicts(_)) => { // if branch conflicts with the workspace, it's ok. keep it unapplied Ok(branch.id) } Err(error) => Err(errors::CreateVirtualBranchFromBranchError::ApplyBranch( error, )), } } /// Just like [`diffy::apply()`], but on error it will attach hashes of the input `base_image` and `patch`. pub fn apply(base_image: &str, patch: &Patch<'_, str>) -> Result { fn md5_hash_hex(b: impl AsRef<[u8]>) -> String { format!("{:x}", md5::compute(b)) } #[derive(Debug)] #[allow(dead_code)] // Read by Debug auto-impl, which doesn't count pub enum DebugLine { // Note that each of these strings is a hash only Context(String), Delete(String), Insert(String), } impl<'a> From<&diffy::Line<'a, str>> for DebugLine { fn from(line: &Line<'a, str>) -> Self { match line { Line::Context(s) => DebugLine::Context(md5_hash_hex(s)), Line::Delete(s) => DebugLine::Delete(md5_hash_hex(s)), Line::Insert(s) => DebugLine::Insert(md5_hash_hex(s)), } } } #[derive(Debug)] #[allow(dead_code)] // Read by Debug auto-impl, which doesn't count struct DebugHunk { old_range: diffy::HunkRange, new_range: diffy::HunkRange, lines: Vec, } impl<'a> From<&diffy::Hunk<'a, str>> for DebugHunk { fn from(hunk: &diffy::Hunk<'a, str>) -> Self { Self { old_range: hunk.old_range(), new_range: hunk.new_range(), lines: hunk.lines().iter().map(Into::into).collect(), } } } #[derive(Debug)] #[allow(dead_code)] // Read by Debug auto-impl, which doesn't count struct DebugContext { base_image_hash: String, hunks: Vec, } impl std::fmt::Display for DebugContext { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { std::fmt::Debug::fmt(self, f) } } diffy_apply(base_image, patch).with_context(|| DebugContext { base_image_hash: md5_hash_hex(base_image), hunks: patch.hunks().iter().map(Into::into).collect(), }) } // Goes through a set of changes and checks if conflicts are present. If no conflicts // are present in a file it will be resolved, meaning it will be removed from the // conflicts file. fn update_conflict_markers( project_repository: &project_repository::Repository, files: &HashMap>, ) -> Result<()> { let conflicting_files = conflicts::conflicting_files(project_repository)?; for (file_path, non_commited_hunks) in files { let mut conflicted = false; if conflicting_files.contains(&file_path.display().to_string()) { // check file for conflict markers, resolve the file if there are none in any hunk for hunk in non_commited_hunks { if hunk.diff.contains("<<<<<<< ours") { conflicted = true; } if hunk.diff.contains(">>>>>>> theirs") { conflicted = true; } } if !conflicted { conflicts::resolve(project_repository, &file_path.display().to_string()).unwrap(); } } } Ok(()) } #[cfg(test)] mod tests { use super::*; #[test] fn joined_test() { assert!(!joined(1, 2, 3, 4)); assert!(joined(1, 4, 2, 3)); assert!(joined(2, 3, 1, 4)); assert!(!joined(3, 4, 1, 2)); assert!(joined(1, 2, 2, 3)); assert!(joined(1, 3, 2, 3)); assert!(joined(2, 3, 1, 2)); assert!(!joined(1, 1, 2, 2)); assert!(joined(1, 1, 1, 1)); assert!(joined(1, 1, 1, 2)); assert!(joined(1, 2, 2, 2)); } #[test] fn normalize_branch_name_test() { assert_eq!(normalize_branch_name("feature/branch"), "feature/branch"); assert_eq!(normalize_branch_name("foo#branch"), "foo#branch"); assert_eq!(normalize_branch_name("foo!branch"), "foo-branch"); } }