Adds comprehensive Git worktree support for beads issue tracking: Core changes: - New internal/git/gitdir.go package for worktree detection - GetGitDir() returns proper .git location (main repo, not worktree) - Updated all hooks to use git.GetGitDir() instead of local helper - BeadsDir() now prioritizes main repository's .beads directory Features: - Hooks auto-install in main repo when run from worktree - Shared .beads directory across all worktrees - Config option no-install-hooks to disable auto-install - New bd worktree subcommand for diagnostics Documentation: - New docs/WORKTREES.md with setup instructions - Updated CHANGELOG.md and AGENT_INSTRUCTIONS.md Testing: - Updated tests to use exported git.GetGitDir() - Added worktree detection tests Co-authored-by: Claude <noreply@anthropic.com> Closes: #478
292 lines
8.5 KiB
Go
292 lines
8.5 KiB
Go
package main
|
|
|
|
import (
|
|
"encoding/json"
|
|
"fmt"
|
|
"os"
|
|
"path/filepath"
|
|
|
|
"github.com/fatih/color"
|
|
"github.com/spf13/cobra"
|
|
"github.com/steveyegge/beads/internal/deletions"
|
|
"github.com/steveyegge/beads/internal/types"
|
|
)
|
|
|
|
var migrateTombstonesCmd = &cobra.Command{
|
|
Use: "migrate-tombstones",
|
|
Short: "Convert deletions.jsonl entries to inline tombstones",
|
|
Long: `Migrate legacy deletions.jsonl entries to inline tombstones in issues.jsonl.
|
|
|
|
This command converts existing deletion records from the legacy deletions.jsonl
|
|
manifest to inline tombstone entries in issues.jsonl. This is part of the
|
|
transition from separate deletion tracking to unified tombstone-based deletion.
|
|
|
|
The migration:
|
|
1. Reads existing deletions from deletions.jsonl
|
|
2. Checks issues.jsonl for already-existing tombstones
|
|
3. Creates tombstone entries for unmigrated deletions
|
|
4. Appends new tombstones to issues.jsonl
|
|
5. Archives deletions.jsonl with .migrated suffix
|
|
|
|
Use --dry-run to preview changes without modifying files.
|
|
|
|
Examples:
|
|
bd migrate-tombstones # Migrate deletions to tombstones
|
|
bd migrate-tombstones --dry-run # Preview what would be migrated
|
|
bd migrate-tombstones --verbose # Show detailed progress`,
|
|
Run: func(cmd *cobra.Command, _ []string) {
|
|
dryRun, _ := cmd.Flags().GetBool("dry-run")
|
|
verbose, _ := cmd.Flags().GetBool("verbose")
|
|
|
|
// Block writes in readonly mode
|
|
if !dryRun {
|
|
CheckReadonly("migrate-tombstones")
|
|
}
|
|
|
|
// Find .beads directory
|
|
beadsDir := findBeadsDir()
|
|
if beadsDir == "" {
|
|
if jsonOutput {
|
|
outputJSON(map[string]interface{}{
|
|
"error": "no_beads_directory",
|
|
"message": "No .beads directory found. Run 'bd init' first.",
|
|
})
|
|
} else {
|
|
fmt.Fprintf(os.Stderr, "Error: no .beads directory found\n")
|
|
fmt.Fprintf(os.Stderr, "Hint: run 'bd init' to initialize bd\n")
|
|
}
|
|
os.Exit(1)
|
|
}
|
|
|
|
// Check paths
|
|
deletionsPath := deletions.DefaultPath(beadsDir)
|
|
issuesPath := filepath.Join(beadsDir, "issues.jsonl")
|
|
|
|
// Load existing deletions
|
|
loadResult, err := deletions.LoadDeletions(deletionsPath)
|
|
if err != nil {
|
|
if jsonOutput {
|
|
outputJSON(map[string]interface{}{
|
|
"error": "load_deletions_failed",
|
|
"message": err.Error(),
|
|
})
|
|
} else {
|
|
fmt.Fprintf(os.Stderr, "Error loading deletions.jsonl: %v\n", err)
|
|
}
|
|
os.Exit(1)
|
|
}
|
|
|
|
if len(loadResult.Records) == 0 {
|
|
if jsonOutput {
|
|
outputJSON(map[string]interface{}{
|
|
"status": "noop",
|
|
"message": "No deletions to migrate",
|
|
"migrated": 0,
|
|
"skipped": 0,
|
|
})
|
|
} else {
|
|
fmt.Println("No deletions.jsonl entries to migrate")
|
|
}
|
|
return
|
|
}
|
|
|
|
// Print warnings from loading
|
|
for _, warning := range loadResult.Warnings {
|
|
if !jsonOutput {
|
|
color.Yellow("Warning: %s\n", warning)
|
|
}
|
|
}
|
|
|
|
// Load existing issues.jsonl to find existing tombstones
|
|
existingTombstones := make(map[string]bool)
|
|
if _, err := os.Stat(issuesPath); err == nil {
|
|
// nolint:gosec // G304: issuesPath is controlled from beadsDir
|
|
file, err := os.Open(issuesPath)
|
|
if err != nil {
|
|
if jsonOutput {
|
|
outputJSON(map[string]interface{}{
|
|
"error": "load_issues_failed",
|
|
"message": err.Error(),
|
|
})
|
|
} else {
|
|
fmt.Fprintf(os.Stderr, "Error opening issues.jsonl: %v\n", err)
|
|
}
|
|
os.Exit(1)
|
|
}
|
|
|
|
decoder := json.NewDecoder(file)
|
|
for {
|
|
var issue types.Issue
|
|
if err := decoder.Decode(&issue); err != nil {
|
|
if err.Error() == "EOF" {
|
|
break
|
|
}
|
|
// Skip corrupt lines, continue reading
|
|
continue
|
|
}
|
|
if issue.IsTombstone() {
|
|
existingTombstones[issue.ID] = true
|
|
}
|
|
}
|
|
file.Close()
|
|
}
|
|
|
|
// Determine which deletions need migration
|
|
var toMigrate []deletions.DeletionRecord
|
|
var skippedIDs []string
|
|
for id, record := range loadResult.Records {
|
|
if existingTombstones[id] {
|
|
skippedIDs = append(skippedIDs, id)
|
|
if verbose && !jsonOutput {
|
|
fmt.Printf(" Skipping %s (tombstone already exists)\n", id)
|
|
}
|
|
} else {
|
|
toMigrate = append(toMigrate, record)
|
|
}
|
|
}
|
|
|
|
if len(toMigrate) == 0 {
|
|
if jsonOutput {
|
|
outputJSON(map[string]interface{}{
|
|
"status": "noop",
|
|
"message": "All deletions already migrated to tombstones",
|
|
"migrated": 0,
|
|
"skipped": len(skippedIDs),
|
|
})
|
|
} else {
|
|
fmt.Printf("All %d deletion(s) already have tombstones in issues.jsonl\n", len(skippedIDs))
|
|
}
|
|
return
|
|
}
|
|
|
|
// Dry run - just report what would happen
|
|
if dryRun {
|
|
if jsonOutput {
|
|
outputJSON(map[string]interface{}{
|
|
"dry_run": true,
|
|
"would_migrate": len(toMigrate),
|
|
"skipped": len(skippedIDs),
|
|
"total": len(loadResult.Records),
|
|
})
|
|
} else {
|
|
fmt.Println("Dry run mode - no changes will be made")
|
|
fmt.Printf("\nWould migrate %d deletion(s) to tombstones:\n", len(toMigrate))
|
|
for _, record := range toMigrate {
|
|
fmt.Printf(" - %s (deleted %s by %s)\n",
|
|
record.ID,
|
|
record.Timestamp.Format("2006-01-02"),
|
|
record.Actor)
|
|
}
|
|
if len(skippedIDs) > 0 {
|
|
fmt.Printf("\nWould skip %d already-migrated deletion(s)\n", len(skippedIDs))
|
|
}
|
|
}
|
|
return
|
|
}
|
|
|
|
// Perform migration - append tombstones to issues.jsonl
|
|
if verbose && !jsonOutput {
|
|
fmt.Printf("Creating %d tombstone(s)...\n", len(toMigrate))
|
|
}
|
|
|
|
// Open issues.jsonl for appending
|
|
// nolint:gosec // G304: issuesPath is controlled from beadsDir
|
|
file, err := os.OpenFile(issuesPath, os.O_APPEND|os.O_CREATE|os.O_WRONLY, 0600)
|
|
if err != nil {
|
|
if jsonOutput {
|
|
outputJSON(map[string]interface{}{
|
|
"error": "open_issues_failed",
|
|
"message": err.Error(),
|
|
})
|
|
} else {
|
|
fmt.Fprintf(os.Stderr, "Error opening issues.jsonl for append: %v\n", err)
|
|
}
|
|
os.Exit(1)
|
|
}
|
|
defer file.Close()
|
|
|
|
encoder := json.NewEncoder(file)
|
|
var migratedIDs []string
|
|
for _, record := range toMigrate {
|
|
tombstone := convertDeletionRecordToTombstone(record)
|
|
if err := encoder.Encode(tombstone); err != nil {
|
|
if jsonOutput {
|
|
outputJSON(map[string]interface{}{
|
|
"error": "write_tombstone_failed",
|
|
"message": err.Error(),
|
|
"issue_id": record.ID,
|
|
})
|
|
} else {
|
|
fmt.Fprintf(os.Stderr, "Error writing tombstone for %s: %v\n", record.ID, err)
|
|
}
|
|
os.Exit(1)
|
|
}
|
|
migratedIDs = append(migratedIDs, record.ID)
|
|
if verbose && !jsonOutput {
|
|
fmt.Printf(" ✓ Created tombstone for %s\n", record.ID)
|
|
}
|
|
}
|
|
|
|
// Archive deletions.jsonl
|
|
archivePath := deletionsPath + ".migrated"
|
|
if err := os.Rename(deletionsPath, archivePath); err != nil {
|
|
// Warn but don't fail - tombstones were already created
|
|
if !jsonOutput {
|
|
color.Yellow("Warning: could not archive deletions.jsonl: %v\n", err)
|
|
}
|
|
} else if verbose && !jsonOutput {
|
|
fmt.Printf(" ✓ Archived deletions.jsonl to %s\n", filepath.Base(archivePath))
|
|
}
|
|
|
|
// Success output
|
|
if jsonOutput {
|
|
outputJSON(map[string]interface{}{
|
|
"status": "success",
|
|
"migrated": len(migratedIDs),
|
|
"skipped": len(skippedIDs),
|
|
"total": len(loadResult.Records),
|
|
"archive": archivePath,
|
|
"migrated_ids": migratedIDs,
|
|
})
|
|
} else {
|
|
color.Green("\n✓ Migration complete\n\n")
|
|
fmt.Printf(" Migrated: %d tombstone(s)\n", len(migratedIDs))
|
|
if len(skippedIDs) > 0 {
|
|
fmt.Printf(" Skipped: %d (already had tombstones)\n", len(skippedIDs))
|
|
}
|
|
fmt.Printf(" Archived: %s\n", filepath.Base(archivePath))
|
|
fmt.Println("\nNext steps:")
|
|
fmt.Println(" 1. Run 'bd sync' to propagate tombstones to remote")
|
|
fmt.Println(" 2. Other clones will receive tombstones on next sync")
|
|
}
|
|
},
|
|
}
|
|
|
|
// convertDeletionRecordToTombstone creates a tombstone issue from a deletion record.
|
|
// This is similar to the importer's convertDeletionToTombstone but operates on
|
|
// deletions.DeletionRecord directly.
|
|
func convertDeletionRecordToTombstone(del deletions.DeletionRecord) *types.Issue {
|
|
deletedAt := del.Timestamp
|
|
return &types.Issue{
|
|
ID: del.ID,
|
|
Title: "(deleted)",
|
|
Description: "",
|
|
Status: types.StatusTombstone,
|
|
Priority: 0, // Unknown priority (0 = unset)
|
|
IssueType: types.TypeTask, // Default type (must be valid)
|
|
CreatedAt: del.Timestamp,
|
|
UpdatedAt: del.Timestamp,
|
|
DeletedAt: &deletedAt,
|
|
DeletedBy: del.Actor,
|
|
DeleteReason: del.Reason,
|
|
OriginalType: "", // Not available in legacy deletions.jsonl
|
|
}
|
|
}
|
|
|
|
func init() {
|
|
migrateTombstonesCmd.Flags().Bool("dry-run", false, "Preview changes without modifying files")
|
|
migrateTombstonesCmd.Flags().Bool("verbose", false, "Show detailed progress")
|
|
migrateTombstonesCmd.Flags().BoolVar(&jsonOutput, "json", false, "Output in JSON format")
|
|
rootCmd.AddCommand(migrateTombstonesCmd)
|
|
}
|