mirror of
https://github.com/sbrow/envr.git
synced 2026-04-18 16:19:23 -04:00
Compare commits
1 Commits
zig
...
bd625a7086
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
bd625a7086 |
7
.gitignore
vendored
7
.gitignore
vendored
@@ -1,16 +1,9 @@
|
|||||||
# dev env
|
# dev env
|
||||||
.direnv
|
.direnv
|
||||||
|
|
||||||
# dependencies
|
|
||||||
deps
|
|
||||||
vendor
|
|
||||||
|
|
||||||
# docs
|
# docs
|
||||||
man
|
man
|
||||||
|
|
||||||
# build artifacts
|
# build artifacts
|
||||||
.zig-cache
|
|
||||||
builds
|
|
||||||
envr
|
envr
|
||||||
envr-go
|
|
||||||
result
|
result
|
||||||
|
|||||||
43
CHANGELOG.md
43
CHANGELOG.md
@@ -1,51 +1,8 @@
|
|||||||
# Changelog
|
# Changelog
|
||||||
|
|
||||||
## [0.2.1](https://github.com/sbrow/envr/compare/v0.2.0...v0.2.1) (2026-01-12)
|
|
||||||
|
|
||||||
|
|
||||||
### Bug Fixes
|
|
||||||
|
|
||||||
* Added `add` as an alias for backup. ([cf363ab](https://github.com/sbrow/envr/commit/cf363abc4d8cec208d23c6acedbb7e0dd6900332))
|
|
||||||
|
|
||||||
## [0.2.0](https://github.com/sbrow/envr/compare/v0.1.1...v0.2.0) (2025-11-10)
|
|
||||||
|
|
||||||
|
|
||||||
### ⚠ BREAKING CHANGES
|
|
||||||
|
|
||||||
* Dir is now derived from Path rather than stored in the DB. Your DB will need to be updated.
|
|
||||||
* **scan:** The config value `scan.Exclude` is now a list rather than a string.
|
|
||||||
* **check:** Renamed the `check` command to `deps`.
|
|
||||||
* The config value `scan.Include` is now a list rather than a string.
|
|
||||||
|
|
||||||
### Features
|
|
||||||
|
|
||||||
* Added new `check` command. ([cbd74f3](https://github.com/sbrow/envr/commit/cbd74f387e2e330b2557d07dd82ba05cc91300ac))
|
|
||||||
* **config:** The default config now filters out more junk. ([15be62b](https://github.com/sbrow/envr/commit/15be62b5a2a5a735b90b074497d645c5a2cfced8))
|
|
||||||
* **init:** Added a `--force` flag for overwriting an existing config. ([169653d](https://github.com/sbrow/envr/commit/169653d7566f63730fb9da80a18330a566223be9))
|
|
||||||
* Multiple scan includes are now supported. ([4273fa5](https://github.com/sbrow/envr/commit/4273fa58956d8736271a0af66202dca481126fe4))
|
|
||||||
* **scan:** Added support for multiple exports. ([f43705c](https://github.com/sbrow/envr/commit/f43705cd53c6d87aef1f69df4e474441f25c1dc7))
|
|
||||||
* **sync:** envr can now detect if directories have moved. ([4db0a4d](https://github.com/sbrow/envr/commit/4db0a4d33d2b6a79d13b36a8e8631f895e8fef8d))
|
|
||||||
* **sync:** Now checks files for mismatched hashes before replacing. ([8074f7a](https://github.com/sbrow/envr/commit/8074f7ae6dfa54e931a198257f3f8e6d0cfe353a))
|
|
||||||
|
|
||||||
|
|
||||||
### Bug Fixes
|
|
||||||
|
|
||||||
* **check:** `fd` now correctly gets marked as found. ([17ce49c](https://github.com/sbrow/envr/commit/17ce49cd2d33942282c6f54ce819ac25978f6b7c))
|
|
||||||
|
|
||||||
|
|
||||||
### Code Refactoring
|
|
||||||
|
|
||||||
* **check:** Renamed the `check` command to `deps`. ([c9c34ce](https://github.com/sbrow/envr/commit/c9c34ce771653da214635f1df1fef1f23265c552))
|
|
||||||
* Dir is no longer stored in the database. ([0fef74a](https://github.com/sbrow/envr/commit/0fef74a9bba0fbf3c34b66c2095955e6eee7047b))
|
|
||||||
|
|
||||||
## [0.1.1](https://github.com/sbrow/envr/compare/v0.1.0...v0.1.1) (2025-11-05)
|
## [0.1.1](https://github.com/sbrow/envr/compare/v0.1.0...v0.1.1) (2025-11-05)
|
||||||
|
|
||||||
|
|
||||||
### Features
|
|
||||||
|
|
||||||
* **sync:** Results are now displayed in a table. ([42796ec](https://github.com/sbrow/envr/commit/42796ec77b1817e1b9f09068d76a7b6e30da246b))
|
|
||||||
|
|
||||||
|
|
||||||
### Bug Fixes
|
### Bug Fixes
|
||||||
|
|
||||||
* **sync:** Fixed an issue where deleted folders would be restored. ([9ab72a2](https://github.com/sbrow/envr/commit/9ab72a25faf1af0eedb2f4574166c6ee47450ebb))
|
* **sync:** Fixed an issue where deleted folders would be restored. ([9ab72a2](https://github.com/sbrow/envr/commit/9ab72a25faf1af0eedb2f4574166c6ee47450ebb))
|
||||||
|
|||||||
92
Makefile
92
Makefile
@@ -1,92 +0,0 @@
|
|||||||
# Makefile for envr - Environment file manager
|
|
||||||
# Builds release artifacts for GitHub releases
|
|
||||||
|
|
||||||
APP_NAME := envr
|
|
||||||
VERSION := $(shell grep 'version = ' flake.nix | head -1 | sed 's/.*version = "\(.*\)";/\1/')
|
|
||||||
BUILD_DIR := builds
|
|
||||||
LDFLAGS := -X github.com/sbrow/envr/cmd.version=v$(VERSION) -s -w
|
|
||||||
|
|
||||||
# Binary names
|
|
||||||
LINUX_AMD64_BIN := $(BUILD_DIR)/$(APP_NAME)-$(VERSION)-linux-amd64
|
|
||||||
LINUX_ARM64_BIN := $(BUILD_DIR)/$(APP_NAME)-$(VERSION)-linux-arm64
|
|
||||||
DARWIN_ARM64_BIN := $(BUILD_DIR)/$(APP_NAME)-$(VERSION)-darwin-arm64
|
|
||||||
|
|
||||||
.PHONY: all clean cleanall build-linux build-darwin compress release help
|
|
||||||
|
|
||||||
# Default target
|
|
||||||
all: release clean
|
|
||||||
|
|
||||||
# Create build directory
|
|
||||||
$(BUILD_DIR):
|
|
||||||
@mkdir -p $(BUILD_DIR)
|
|
||||||
|
|
||||||
# Build Linux AMD64
|
|
||||||
$(LINUX_AMD64_BIN): $(BUILD_DIR)
|
|
||||||
@echo "Building for Linux AMD64..."
|
|
||||||
GOOS=linux GOARCH=amd64 CGO_ENABLED=0 go build -ldflags "$(LDFLAGS)" -o $(LINUX_AMD64_BIN) .
|
|
||||||
@echo "Built $(LINUX_AMD64_BIN)"
|
|
||||||
|
|
||||||
# Build Linux ARM64
|
|
||||||
$(LINUX_ARM64_BIN): $(BUILD_DIR)
|
|
||||||
@echo "Building for Linux ARM64..."
|
|
||||||
GOOS=linux GOARCH=arm64 CGO_ENABLED=0 go build -ldflags "$(LDFLAGS)" -o $(LINUX_ARM64_BIN) .
|
|
||||||
@echo "Built $(LINUX_ARM64_BIN)"
|
|
||||||
|
|
||||||
# Build Darwin ARM64 (Mac)
|
|
||||||
$(DARWIN_ARM64_BIN): $(BUILD_DIR)
|
|
||||||
@echo "Building for Darwin ARM64..."
|
|
||||||
GOOS=darwin GOARCH=arm64 CGO_ENABLED=0 go build -ldflags "$(LDFLAGS)" -o $(DARWIN_ARM64_BIN) .
|
|
||||||
@echo "Built $(DARWIN_ARM64_BIN)"
|
|
||||||
|
|
||||||
# Build all binaries
|
|
||||||
build-linux: $(LINUX_AMD64_BIN) $(LINUX_ARM64_BIN)
|
|
||||||
build-darwin: $(DARWIN_ARM64_BIN)
|
|
||||||
|
|
||||||
# Compress Linux artifacts with gzip
|
|
||||||
$(BUILD_DIR)/$(APP_NAME)-$(VERSION)-linux-amd64.tar.gz: $(LINUX_AMD64_BIN)
|
|
||||||
@echo "Compressing Linux AMD64 artifact..."
|
|
||||||
cd $(BUILD_DIR) && tar -czf $(APP_NAME)-$(VERSION)-linux-amd64.tar.gz --transform 's|.*|$(APP_NAME)|' $(shell basename $(LINUX_AMD64_BIN))
|
|
||||||
|
|
||||||
$(BUILD_DIR)/$(APP_NAME)-$(VERSION)-linux-arm64.tar.gz: $(LINUX_ARM64_BIN)
|
|
||||||
@echo "Compressing Linux ARM64 artifact..."
|
|
||||||
cd $(BUILD_DIR) && tar -czf $(APP_NAME)-$(VERSION)-linux-arm64.tar.gz --transform 's|.*|$(APP_NAME)|' $(shell basename $(LINUX_ARM64_BIN))
|
|
||||||
|
|
||||||
# Compress Darwin artifacts with zip
|
|
||||||
$(BUILD_DIR)/$(APP_NAME)-$(VERSION)-darwin-arm64.zip: $(DARWIN_ARM64_BIN)
|
|
||||||
@echo "Compressing Darwin ARM64 artifact..."
|
|
||||||
cd $(BUILD_DIR) && cp $(shell basename $(DARWIN_ARM64_BIN)) $(APP_NAME) && zip $(APP_NAME)-$(VERSION)-darwin-arm64.zip $(APP_NAME) && rm $(APP_NAME)
|
|
||||||
|
|
||||||
# Compress all artifacts
|
|
||||||
compress: $(BUILD_DIR)/$(APP_NAME)-$(VERSION)-linux-amd64.tar.gz \
|
|
||||||
$(BUILD_DIR)/$(APP_NAME)-$(VERSION)-linux-arm64.tar.gz \
|
|
||||||
$(BUILD_DIR)/$(APP_NAME)-$(VERSION)-darwin-arm64.zip
|
|
||||||
|
|
||||||
# Build and compress all release artifacts
|
|
||||||
release: build-linux build-darwin compress
|
|
||||||
@echo "Release artifacts created:"
|
|
||||||
@ls -la $(BUILD_DIR)/*.tar.gz $(BUILD_DIR)/*.zip 2>/dev/null || echo "No compressed artifacts found"
|
|
||||||
|
|
||||||
# Clean binary files only
|
|
||||||
clean:
|
|
||||||
@echo "Cleaning binary files..."
|
|
||||||
@rm -f $(LINUX_AMD64_BIN) $(LINUX_ARM64_BIN) $(DARWIN_ARM64_BIN)
|
|
||||||
|
|
||||||
# Clean everything in build directory
|
|
||||||
cleanall:
|
|
||||||
@echo "Cleaning build directory..."
|
|
||||||
@rm -rf $(BUILD_DIR)
|
|
||||||
|
|
||||||
# Show available targets
|
|
||||||
help:
|
|
||||||
@echo "Available targets:"
|
|
||||||
@echo " all - Build all release artifacts (default)"
|
|
||||||
@echo " release - Build and compress all release artifacts"
|
|
||||||
@echo " build-linux - Build Linux binaries only"
|
|
||||||
@echo " build-darwin - Build Darwin binaries only"
|
|
||||||
@echo " compress - Compress all built binaries"
|
|
||||||
@echo " clean - Remove binary files only"
|
|
||||||
@echo " cleanall - Remove entire build directory"
|
|
||||||
@echo " help - Show this help message"
|
|
||||||
@echo ""
|
|
||||||
@echo "Release artifacts will be created in $(BUILD_DIR)/"
|
|
||||||
@echo "Version: $(VERSION)"
|
|
||||||
@@ -19,11 +19,10 @@ be run on a cron.
|
|||||||
- 🔍 **Smart Scanning**: Automatically discover and import `.env` files in your
|
- 🔍 **Smart Scanning**: Automatically discover and import `.env` files in your
|
||||||
home directory.
|
home directory.
|
||||||
- ✨ **Interactive CLI**: User-friendly prompts for file selection and management.
|
- ✨ **Interactive CLI**: User-friendly prompts for file selection and management.
|
||||||
- 🗂️ **Rename Detection**: Automatically finds and updates renamed/moved
|
|
||||||
repositories.
|
|
||||||
|
|
||||||
## TODOS
|
## TODOS
|
||||||
- [x] Rename Detection: automatically update moved files.
|
|
||||||
|
- [ ] 🗂️ **Rename Detection**: Automatically handle renamed repositories.
|
||||||
- [ ] Allow use of keys from `ssh-agent`
|
- [ ] Allow use of keys from `ssh-agent`
|
||||||
- [x] Allow configuration of ssh key.
|
- [x] Allow configuration of ssh key.
|
||||||
- [x] Allow multiple ssh keys.
|
- [x] Allow multiple ssh keys.
|
||||||
|
|||||||
92
WINDOWS.md
92
WINDOWS.md
@@ -1,92 +0,0 @@
|
|||||||
# Windows Compatibility Guide
|
|
||||||
|
|
||||||
This document outlines Windows compatibility issues and solutions for the envr project.
|
|
||||||
|
|
||||||
## Critical Issues
|
|
||||||
|
|
||||||
### 1. Path Handling Bug (MUST FIX)
|
|
||||||
|
|
||||||
**File:** `app/env_file.go:209`
|
|
||||||
|
|
||||||
**Issue:** Uses `path.Join` instead of `filepath.Join`, which won't work correctly on Windows due to different path separators.
|
|
||||||
|
|
||||||
**Current code:**
|
|
||||||
```go
|
|
||||||
f.Path = path.Join(newDir, path.Base(f.Path))
|
|
||||||
```
|
|
||||||
|
|
||||||
**Fixed code:**
|
|
||||||
```go
|
|
||||||
f.Path = filepath.Join(newDir, filepath.Base(f.Path))
|
|
||||||
```
|
|
||||||
|
|
||||||
## External Dependencies
|
|
||||||
|
|
||||||
The application relies on external tools that need to be installed separately on Windows:
|
|
||||||
|
|
||||||
### Required Tools
|
|
||||||
|
|
||||||
1. **fd** - Fast file finder
|
|
||||||
- Install via: `winget install sharkdp.fd` or `choco install fd`
|
|
||||||
- Alternative: `scoop install fd`
|
|
||||||
|
|
||||||
2. **git** - Version control system
|
|
||||||
- Install via: `winget install Git.Git` or download from git-scm.com
|
|
||||||
- Usually already available on most development machines
|
|
||||||
|
|
||||||
## Minor Compatibility Notes
|
|
||||||
|
|
||||||
### File Permissions
|
|
||||||
- Unix file permissions (`0755`, `0644`) are used throughout the codebase
|
|
||||||
- These are safely ignored on Windows - no changes needed
|
|
||||||
|
|
||||||
### Editor Configuration
|
|
||||||
**File:** `cmd/edit_config.go:20-24`
|
|
||||||
|
|
||||||
**Issue:** Relies on `$EDITOR` environment variable which is less common on Windows.
|
|
||||||
|
|
||||||
**Current behavior:** Fails if `$EDITOR` is not set
|
|
||||||
|
|
||||||
**Recommended improvement:** Add fallback detection for Windows editors:
|
|
||||||
```go
|
|
||||||
editor := os.Getenv("EDITOR")
|
|
||||||
if editor == "" {
|
|
||||||
if runtime.GOOS == "windows" {
|
|
||||||
editor = "notepad.exe" // or "code.exe" for VS Code
|
|
||||||
} else {
|
|
||||||
fmt.Println("Error: $EDITOR environment variable is not set")
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
```
|
|
||||||
|
|
||||||
## Installation Instructions for Windows
|
|
||||||
|
|
||||||
1. Install required dependencies:
|
|
||||||
```powershell
|
|
||||||
winget install sharkdp.fd
|
|
||||||
winget install Git.Git
|
|
||||||
```
|
|
||||||
|
|
||||||
2. Fix the path handling bug in `app/env_file.go:209`
|
|
||||||
|
|
||||||
3. Build and run as normal:
|
|
||||||
```powershell
|
|
||||||
go build
|
|
||||||
.\envr.exe init
|
|
||||||
```
|
|
||||||
|
|
||||||
## Testing on Windows
|
|
||||||
|
|
||||||
After applying the critical path fix, the core functionality should work correctly on Windows. The application has been designed with cross-platform compatibility in mind, using:
|
|
||||||
|
|
||||||
- `filepath` package for path operations (mostly)
|
|
||||||
- `os.UserHomeDir()` for home directory detection
|
|
||||||
- Standard Go file operations
|
|
||||||
|
|
||||||
## Summary
|
|
||||||
|
|
||||||
- **1 critical bug** must be fixed for Windows compatibility
|
|
||||||
- **2 external tools** need to be installed
|
|
||||||
- **1 minor enhancement** recommended for better Windows UX
|
|
||||||
- Overall architecture is Windows-compatible
|
|
||||||
186
app/config.go
186
app/config.go
@@ -6,7 +6,6 @@ import (
|
|||||||
"fmt"
|
"fmt"
|
||||||
"os"
|
"os"
|
||||||
"os/exec"
|
"os/exec"
|
||||||
"path"
|
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
@@ -19,17 +18,15 @@ type Config struct {
|
|||||||
ScanConfig scanConfig `json:"scan"`
|
ScanConfig scanConfig `json:"scan"`
|
||||||
}
|
}
|
||||||
|
|
||||||
// Used by age to encrypt and decrypt the database.
|
|
||||||
type SshKeyPair struct {
|
type SshKeyPair struct {
|
||||||
Private string `json:"private"` // Path to the private key file
|
Private string `json:"private"` // Path to the private key file
|
||||||
Public string `json:"public"` // Path to the public key file
|
Public string `json:"public"` // Path to the public key file
|
||||||
}
|
}
|
||||||
|
|
||||||
type scanConfig struct {
|
type scanConfig struct {
|
||||||
// TODO: Support multiple matchers
|
Matcher string `json:"matcher"`
|
||||||
Matcher string `json:"matcher"`
|
Exclude string `json:"exclude"`
|
||||||
Exclude []string `json:"exclude"`
|
Include string `json:"include"`
|
||||||
Include []string `json:"include"`
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Create a fresh config with sensible defaults.
|
// Create a fresh config with sensible defaults.
|
||||||
@@ -49,13 +46,8 @@ func NewConfig(privateKeyPaths []string) Config {
|
|||||||
Keys: keys,
|
Keys: keys,
|
||||||
ScanConfig: scanConfig{
|
ScanConfig: scanConfig{
|
||||||
Matcher: "\\.env",
|
Matcher: "\\.env",
|
||||||
Exclude: []string{
|
Exclude: "*.envrc",
|
||||||
"*\\.envrc",
|
Include: "~",
|
||||||
"\\.local/",
|
|
||||||
"node_modules",
|
|
||||||
"vendor",
|
|
||||||
},
|
|
||||||
Include: []string{"~"},
|
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -115,101 +107,77 @@ func (c *Config) Save() error {
|
|||||||
return os.WriteFile(configPath, data, 0644)
|
return os.WriteFile(configPath, data, 0644)
|
||||||
}
|
}
|
||||||
|
|
||||||
// buildFdArgs builds the fd command arguments with multiple exclude patterns
|
|
||||||
func (c Config) buildFdArgs(searchPath string, includeIgnored bool) []string {
|
|
||||||
args := []string{"-a", c.ScanConfig.Matcher}
|
|
||||||
|
|
||||||
// Add exclude patterns
|
|
||||||
for _, exclude := range c.ScanConfig.Exclude {
|
|
||||||
args = append(args, "-E", exclude)
|
|
||||||
}
|
|
||||||
|
|
||||||
if includeIgnored {
|
|
||||||
args = append(args, "-HI")
|
|
||||||
} else {
|
|
||||||
args = append(args, "-H")
|
|
||||||
}
|
|
||||||
|
|
||||||
args = append(args, searchPath)
|
|
||||||
return args
|
|
||||||
}
|
|
||||||
|
|
||||||
// Use fd to find all ignored .env files that match the config's parameters
|
// Use fd to find all ignored .env files that match the config's parameters
|
||||||
func (c Config) scan() (paths []string, err error) {
|
func (c Config) scan() (paths []string, err error) {
|
||||||
searchPaths, err := c.searchPaths()
|
searchPath, err := c.searchPath()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return []string{}, err
|
return []string{}, err
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, searchPath := range searchPaths {
|
// Find all files (including ignored ones)
|
||||||
// Find all files (including ignored ones)
|
fmt.Printf("Searching for all files in \"%s\"...\n", searchPath)
|
||||||
fmt.Printf("Searching for all files in \"%s\"...\n", searchPath)
|
allCmd := exec.Command("fd", "-a", c.ScanConfig.Matcher, "-E", c.ScanConfig.Exclude, "-HI", searchPath)
|
||||||
allCmd := exec.Command("fd", c.buildFdArgs(searchPath, true)...)
|
allOutput, err := allCmd.Output()
|
||||||
allOutput, err := allCmd.Output()
|
|
||||||
if err != nil {
|
|
||||||
return paths, err
|
|
||||||
}
|
|
||||||
|
|
||||||
allFiles := strings.Split(strings.TrimSpace(string(allOutput)), "\n")
|
|
||||||
if len(allFiles) == 1 && allFiles[0] == "" {
|
|
||||||
allFiles = []string{}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Find unignored files
|
|
||||||
fmt.Printf("Search for unignored fies in \"%s\"...\n", searchPath)
|
|
||||||
unignoredCmd := exec.Command("fd", c.buildFdArgs(searchPath, false)...)
|
|
||||||
unignoredOutput, err := unignoredCmd.Output()
|
|
||||||
if err != nil {
|
|
||||||
return []string{}, err
|
|
||||||
}
|
|
||||||
|
|
||||||
unignoredFiles := strings.Split(strings.TrimSpace(string(unignoredOutput)), "\n")
|
|
||||||
if len(unignoredFiles) == 1 && unignoredFiles[0] == "" {
|
|
||||||
unignoredFiles = []string{}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Create a map for faster lookup
|
|
||||||
unignoredMap := make(map[string]bool)
|
|
||||||
for _, file := range unignoredFiles {
|
|
||||||
unignoredMap[file] = true
|
|
||||||
}
|
|
||||||
|
|
||||||
// Filter to get only ignored files
|
|
||||||
var ignoredFiles []string
|
|
||||||
for _, file := range allFiles {
|
|
||||||
if !unignoredMap[file] {
|
|
||||||
ignoredFiles = append(ignoredFiles, file)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
paths = append(paths, ignoredFiles...)
|
|
||||||
}
|
|
||||||
|
|
||||||
return paths, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (c Config) searchPaths() (paths []string, err error) {
|
|
||||||
homeDir, err := os.UserHomeDir()
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return paths, err
|
return []string{}, err
|
||||||
}
|
}
|
||||||
|
|
||||||
includes := c.ScanConfig.Include
|
allFiles := strings.Split(strings.TrimSpace(string(allOutput)), "\n")
|
||||||
|
if len(allFiles) == 1 && allFiles[0] == "" {
|
||||||
|
allFiles = []string{}
|
||||||
|
}
|
||||||
|
|
||||||
for _, include := range includes {
|
// Find unignored files
|
||||||
path := strings.Replace(include, "~", homeDir, 1)
|
fmt.Printf("Search for unignored fies in \"%s\"...\n", searchPath)
|
||||||
absPath, err := filepath.Abs(path)
|
unignoredCmd := exec.Command("fd", "-a", c.ScanConfig.Matcher, "-E", c.ScanConfig.Exclude, "-H", searchPath)
|
||||||
if err != nil {
|
unignoredOutput, err := unignoredCmd.Output()
|
||||||
return paths, err
|
if err != nil {
|
||||||
|
return []string{}, err
|
||||||
|
}
|
||||||
|
|
||||||
|
unignoredFiles := strings.Split(strings.TrimSpace(string(unignoredOutput)), "\n")
|
||||||
|
if len(unignoredFiles) == 1 && unignoredFiles[0] == "" {
|
||||||
|
unignoredFiles = []string{}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Create a map for faster lookup
|
||||||
|
unignoredMap := make(map[string]bool)
|
||||||
|
for _, file := range unignoredFiles {
|
||||||
|
unignoredMap[file] = true
|
||||||
|
}
|
||||||
|
|
||||||
|
// Filter to get only ignored files
|
||||||
|
var ignoredFiles []string
|
||||||
|
for _, file := range allFiles {
|
||||||
|
if !unignoredMap[file] {
|
||||||
|
ignoredFiles = append(ignoredFiles, file)
|
||||||
}
|
}
|
||||||
|
|
||||||
paths = append(paths, absPath)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return paths, nil
|
return ignoredFiles, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s SshKeyPair) identity() (age.Identity, error) {
|
func (c Config) searchPath() (path string, err error) {
|
||||||
|
include := c.ScanConfig.Include
|
||||||
|
|
||||||
|
if include == "~" {
|
||||||
|
homeDir, err := os.UserHomeDir()
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
return homeDir, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
absPath, err := filepath.Abs(include)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
return absPath, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// TODO: Should this be private?
|
||||||
|
func (s SshKeyPair) Identity() (age.Identity, error) {
|
||||||
sshKey, err := os.ReadFile(s.Private)
|
sshKey, err := os.ReadFile(s.Private)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, fmt.Errorf("failed to read SSH key: %w", err)
|
return nil, fmt.Errorf("failed to read SSH key: %w", err)
|
||||||
@@ -223,7 +191,8 @@ func (s SshKeyPair) identity() (age.Identity, error) {
|
|||||||
return id, nil
|
return id, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s SshKeyPair) recipient() (age.Recipient, error) {
|
// TODO: Should this be private?
|
||||||
|
func (s SshKeyPair) Recipient() (age.Recipient, error) {
|
||||||
sshKey, err := os.ReadFile(s.Public)
|
sshKey, err := os.ReadFile(s.Public)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, fmt.Errorf("failed to read SSH key: %w", err)
|
return nil, fmt.Errorf("failed to read SSH key: %w", err)
|
||||||
@@ -236,32 +205,3 @@ func (s SshKeyPair) recipient() (age.Recipient, error) {
|
|||||||
|
|
||||||
return id, nil
|
return id, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Use fd to find all git roots in the config's search paths
|
|
||||||
func (c Config) findGitRoots() (paths []string, err error) {
|
|
||||||
searchPaths, err := c.searchPaths()
|
|
||||||
if err != nil {
|
|
||||||
return []string{}, err
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, searchPath := range searchPaths {
|
|
||||||
allCmd := exec.Command("fd", "-H", "-t", "d", "^\\.git$", searchPath)
|
|
||||||
allOutput, err := allCmd.Output()
|
|
||||||
if err != nil {
|
|
||||||
return paths, err
|
|
||||||
}
|
|
||||||
|
|
||||||
allFiles := strings.Split(strings.TrimSpace(string(allOutput)), "\n")
|
|
||||||
if len(allFiles) == 1 && allFiles[0] == "" {
|
|
||||||
allFiles = []string{}
|
|
||||||
}
|
|
||||||
|
|
||||||
for i, file := range allFiles {
|
|
||||||
allFiles[i] = path.Dir(path.Clean(file))
|
|
||||||
}
|
|
||||||
|
|
||||||
paths = append(paths, allFiles...)
|
|
||||||
}
|
|
||||||
|
|
||||||
return paths, nil
|
|
||||||
}
|
|
||||||
|
|||||||
101
app/db.go
101
app/db.go
@@ -1,6 +1,5 @@
|
|||||||
package app
|
package app
|
||||||
|
|
||||||
// TODO: app/db.go should be reviewed.
|
|
||||||
import (
|
import (
|
||||||
"database/sql"
|
"database/sql"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
@@ -14,12 +13,19 @@ import (
|
|||||||
_ "modernc.org/sqlite"
|
_ "modernc.org/sqlite"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// CloseMode determines whether or not the in-memory DB should be saved to disk
|
||||||
|
// before closing the connection.
|
||||||
|
type CloseMode int
|
||||||
|
|
||||||
|
const (
|
||||||
|
ReadOnly CloseMode = iota
|
||||||
|
Write
|
||||||
|
)
|
||||||
|
|
||||||
type Db struct {
|
type Db struct {
|
||||||
db *sql.DB
|
db *sql.DB
|
||||||
cfg Config
|
cfg Config
|
||||||
features *AvailableFeatures
|
features *AvailableFeatures
|
||||||
// If true, the database will be saved to disk before closing
|
|
||||||
changed bool
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func Open() (*Db, error) {
|
func Open() (*Db, error) {
|
||||||
@@ -31,7 +37,7 @@ func Open() (*Db, error) {
|
|||||||
if _, err := os.Stat("/home/spencer/.envr/data.age"); err != nil {
|
if _, err := os.Stat("/home/spencer/.envr/data.age"); err != nil {
|
||||||
// Create a new DB
|
// Create a new DB
|
||||||
db, err := newDb()
|
db, err := newDb()
|
||||||
return &Db{db, *cfg, nil, true}, err
|
return &Db{db, *cfg, nil}, err
|
||||||
} else {
|
} else {
|
||||||
// Open the existing DB
|
// Open the existing DB
|
||||||
tmpFile, err := os.CreateTemp("", "envr-*.db")
|
tmpFile, err := os.CreateTemp("", "envr-*.db")
|
||||||
@@ -53,7 +59,7 @@ func Open() (*Db, error) {
|
|||||||
|
|
||||||
restoreDB(tmpFile.Name(), memDb)
|
restoreDB(tmpFile.Name(), memDb)
|
||||||
|
|
||||||
return &Db{memDb, *cfg, nil, false}, nil
|
return &Db{memDb, *cfg, nil}, nil
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -66,6 +72,7 @@ func newDb() (*sql.DB, error) {
|
|||||||
} else {
|
} else {
|
||||||
_, err := db.Exec(`create table envr_env_files (
|
_, err := db.Exec(`create table envr_env_files (
|
||||||
path text primary key not null
|
path text primary key not null
|
||||||
|
, dir text not null
|
||||||
, remotes text -- JSON
|
, remotes text -- JSON
|
||||||
, sha256 text not null
|
, sha256 text not null
|
||||||
, contents text not null
|
, contents text not null
|
||||||
@@ -101,7 +108,7 @@ func decryptDb(tmpFilePath string, keys []SshKeyPair) error {
|
|||||||
identities := make([]age.Identity, 0, len(keys))
|
identities := make([]age.Identity, 0, len(keys))
|
||||||
|
|
||||||
for _, key := range keys {
|
for _, key := range keys {
|
||||||
id, err := key.identity()
|
id, err := key.Identity()
|
||||||
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
@@ -143,7 +150,7 @@ func restoreDB(path string, destDB *sql.DB) error {
|
|||||||
|
|
||||||
// Returns all the EnvFiles present in the database.
|
// Returns all the EnvFiles present in the database.
|
||||||
func (db *Db) List() (results []EnvFile, err error) {
|
func (db *Db) List() (results []EnvFile, err error) {
|
||||||
rows, err := db.db.Query("select path, remotes, sha256, contents from envr_env_files")
|
rows, err := db.db.Query("select * from envr_env_files")
|
||||||
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
@@ -152,18 +159,14 @@ func (db *Db) List() (results []EnvFile, err error) {
|
|||||||
|
|
||||||
for rows.Next() {
|
for rows.Next() {
|
||||||
var envFile EnvFile
|
var envFile EnvFile
|
||||||
var remotesJson []byte
|
var remotesJSON string
|
||||||
err := rows.Scan(&envFile.Path, &remotesJson, &envFile.Sha256, &envFile.contents)
|
|
||||||
|
err := rows.Scan(&envFile.Path, &envFile.Dir, &remotesJSON, &envFile.Sha256, &envFile.contents)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
// Populate Dir from Path
|
// TODO: unmarshal remotesJSON into envFile.remotes
|
||||||
envFile.Dir = filepath.Dir(envFile.Path)
|
|
||||||
|
|
||||||
if err := json.Unmarshal(remotesJson, &envFile.Remotes); err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
results = append(results, envFile)
|
results = append(results, envFile)
|
||||||
}
|
}
|
||||||
@@ -175,10 +178,10 @@ func (db *Db) List() (results []EnvFile, err error) {
|
|||||||
return results, nil
|
return results, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (db *Db) Close() error {
|
func (db *Db) Close(mode CloseMode) error {
|
||||||
defer db.db.Close()
|
defer db.db.Close()
|
||||||
|
|
||||||
if db.changed {
|
if mode == Write {
|
||||||
// Create tmp file
|
// Create tmp file
|
||||||
tmpFile, err := os.CreateTemp("", "envr-*.db")
|
tmpFile, err := os.CreateTemp("", "envr-*.db")
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -194,8 +197,6 @@ func (db *Db) Close() error {
|
|||||||
if err := encryptDb(tmpFile.Name(), db.cfg.Keys); err != nil {
|
if err := encryptDb(tmpFile.Name(), db.cfg.Keys); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
db.changed = false
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
@@ -241,7 +242,7 @@ func encryptDb(tmpFilePath string, keys []SshKeyPair) error {
|
|||||||
|
|
||||||
recipients := make([]age.Recipient, 0, len(keys))
|
recipients := make([]age.Recipient, 0, len(keys))
|
||||||
for _, key := range keys {
|
for _, key := range keys {
|
||||||
recipient, err := key.recipient()
|
recipient, err := key.Recipient()
|
||||||
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
@@ -277,16 +278,14 @@ func (db *Db) Insert(file EnvFile) error {
|
|||||||
|
|
||||||
// Insert into database
|
// Insert into database
|
||||||
_, err = db.db.Exec(`
|
_, err = db.db.Exec(`
|
||||||
INSERT OR REPLACE INTO envr_env_files (path, remotes, sha256, contents)
|
INSERT OR REPLACE INTO envr_env_files (path, dir, remotes, sha256, contents)
|
||||||
VALUES (?, ?, ?, ?)
|
VALUES (?, ?, ?, ?, ?)
|
||||||
`, file.Path, string(remotesJSON), file.Sha256, file.contents)
|
`, file.Path, file.Dir, string(remotesJSON), file.Sha256, file.contents)
|
||||||
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to insert env file: %w", err)
|
return fmt.Errorf("failed to insert env file: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
db.changed = true
|
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -294,15 +293,12 @@ func (db *Db) Insert(file EnvFile) error {
|
|||||||
func (db *Db) Fetch(path string) (envFile EnvFile, err error) {
|
func (db *Db) Fetch(path string) (envFile EnvFile, err error) {
|
||||||
var remotesJSON string
|
var remotesJSON string
|
||||||
|
|
||||||
row := db.db.QueryRow("SELECT path, remotes, sha256, contents FROM envr_env_files WHERE path = ?", path)
|
row := db.db.QueryRow("SELECT path, dir, remotes, sha256, contents FROM envr_env_files WHERE path = ?", path)
|
||||||
err = row.Scan(&envFile.Path, &remotesJSON, &envFile.Sha256, &envFile.contents)
|
err = row.Scan(&envFile.Path, &envFile.Dir, &remotesJSON, &envFile.Sha256, &envFile.contents)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return EnvFile{}, fmt.Errorf("failed to fetch env file: %w", err)
|
return EnvFile{}, fmt.Errorf("failed to fetch env file: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Populate Dir from Path
|
|
||||||
envFile.Dir = filepath.Dir(envFile.Path)
|
|
||||||
|
|
||||||
if err = json.Unmarshal([]byte(remotesJSON), &envFile.Remotes); err != nil {
|
if err = json.Unmarshal([]byte(remotesJSON), &envFile.Remotes); err != nil {
|
||||||
return EnvFile{}, fmt.Errorf("failed to unmarshal remotes: %w", err)
|
return EnvFile{}, fmt.Errorf("failed to unmarshal remotes: %w", err)
|
||||||
}
|
}
|
||||||
@@ -326,21 +322,12 @@ func (db *Db) Delete(path string) error {
|
|||||||
return fmt.Errorf("no file found with path: %s", path)
|
return fmt.Errorf("no file found with path: %s", path)
|
||||||
}
|
}
|
||||||
|
|
||||||
db.changed = true
|
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Finds .env files in the filesystem that aren't present in the database.
|
// Finds .env files in the filesystem that aren't present in the database.
|
||||||
// path overrides the already configured
|
func (db *Db) Scan() ([]string, error) {
|
||||||
func (db *Db) Scan(paths []string) ([]string, error) {
|
all_paths, err := db.cfg.scan()
|
||||||
cfg := db.cfg
|
|
||||||
|
|
||||||
if paths != nil {
|
|
||||||
cfg.ScanConfig.Include = paths
|
|
||||||
}
|
|
||||||
|
|
||||||
all_paths, err := cfg.scan()
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return []string{}, err
|
return []string{}, err
|
||||||
}
|
}
|
||||||
@@ -387,35 +374,3 @@ func (db *Db) CanScan() error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// If true, [Db.Insert] should be called on the [EnvFile] that generated
|
|
||||||
// the given result
|
|
||||||
func (db Db) UpdateRequired(status EnvFileSyncResult) bool {
|
|
||||||
return status&(BackedUp|DirUpdated) != 0
|
|
||||||
}
|
|
||||||
|
|
||||||
func (db *Db) Sync(file *EnvFile) (result EnvFileSyncResult, err error) {
|
|
||||||
// TODO: This results in findMovedDirs being called multiple times.
|
|
||||||
return file.sync(TrustFilesystem, db)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Looks for git directories that share one or more git remotes with
|
|
||||||
// the given file.
|
|
||||||
func (db Db) findMovedDirs(file *EnvFile) (movedDirs []string, err error) {
|
|
||||||
if err = db.Features().validateFeatures(Fd, Git); err != nil {
|
|
||||||
return movedDirs, err
|
|
||||||
}
|
|
||||||
|
|
||||||
gitRoots, err := db.cfg.findGitRoots()
|
|
||||||
if err != nil {
|
|
||||||
return movedDirs, err
|
|
||||||
} else {
|
|
||||||
for _, dir := range gitRoots {
|
|
||||||
if file.sharesRemote(getGitRemotes(dir)) {
|
|
||||||
movedDirs = append(movedDirs, dir)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return movedDirs, nil
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|||||||
177
app/env_file.go
177
app/env_file.go
@@ -2,19 +2,15 @@ package app
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"crypto/sha256"
|
"crypto/sha256"
|
||||||
"errors"
|
|
||||||
"fmt"
|
"fmt"
|
||||||
"os"
|
"os"
|
||||||
"os/exec"
|
"os/exec"
|
||||||
"path"
|
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
"strings"
|
"strings"
|
||||||
)
|
)
|
||||||
|
|
||||||
type EnvFile struct {
|
type EnvFile struct {
|
||||||
// TODO: Should use FileName in the struct and derive from the path.
|
Path string
|
||||||
Path string
|
|
||||||
// Dir is derived from Path, and is not stored in the database.
|
|
||||||
Dir string
|
Dir string
|
||||||
Remotes []string // []string
|
Remotes []string // []string
|
||||||
Sha256 string
|
Sha256 string
|
||||||
@@ -25,30 +21,16 @@ type EnvFile struct {
|
|||||||
type EnvFileSyncResult int
|
type EnvFileSyncResult int
|
||||||
|
|
||||||
const (
|
const (
|
||||||
// The filesystem contents matches the struct
|
|
||||||
// no further action is required.
|
|
||||||
Noop EnvFileSyncResult = 0
|
|
||||||
// The directory changed, but the file contents matched.
|
|
||||||
// The database must be updated.
|
|
||||||
DirUpdated EnvFileSyncResult = 1
|
|
||||||
// The filesystem has been restored to match the struct
|
|
||||||
// no further action is required.
|
|
||||||
Restored EnvFileSyncResult = 1 << 1
|
|
||||||
// The filesystem has been restored to match the struct.
|
|
||||||
// The directory changed, so the database must be updated
|
|
||||||
RestoredAndDirUpdated EnvFileSyncResult = Restored | DirUpdated
|
|
||||||
// The struct has been updated from the filesystem
|
// The struct has been updated from the filesystem
|
||||||
// and should be updated in the database.
|
// and should be updated in the database.
|
||||||
BackedUp EnvFileSyncResult = 1 << 2
|
Updated EnvFileSyncResult = iota
|
||||||
Error EnvFileSyncResult = 1 << 3
|
// The filesystem has been restored to match the struct
|
||||||
)
|
// no further action is required.
|
||||||
|
Restored
|
||||||
// Determines the source of truth when calling [EnvFile.Sync] or [EnvFile.Restore]
|
Error
|
||||||
type syncDirection int
|
// The filesystem contents matches the struct
|
||||||
|
// no further action is required.
|
||||||
const (
|
Noop
|
||||||
TrustDatabase syncDirection = iota
|
|
||||||
TrustFilesystem
|
|
||||||
)
|
)
|
||||||
|
|
||||||
func NewEnvFile(path string) EnvFile {
|
func NewEnvFile(path string) EnvFile {
|
||||||
@@ -113,119 +95,62 @@ func getGitRemotes(dir string) []string {
|
|||||||
return remotes
|
return remotes
|
||||||
}
|
}
|
||||||
|
|
||||||
// Reconcile the state of the database with the state of the filesystem, using
|
// Install the file into the file system
|
||||||
// dir to determine which side to use a the source of truth.
|
func (file EnvFile) Restore() error {
|
||||||
func (f *EnvFile) sync(dir syncDirection, db *Db) (result EnvFileSyncResult, err error) {
|
// TODO: Handle restores more cleanly
|
||||||
if result != Noop {
|
// Ensure the directory exists
|
||||||
panic("Invalid state")
|
if _, err := os.Stat(file.Dir); err != nil {
|
||||||
|
return fmt.Errorf("directory missing")
|
||||||
}
|
}
|
||||||
|
|
||||||
if _, err := os.Stat(f.Dir); err != nil {
|
// Check if file already exists
|
||||||
// Directory doesn't exist
|
if _, err := os.Stat(file.Path); err == nil {
|
||||||
|
return fmt.Errorf("file already exists: %s", file.Path)
|
||||||
var movedDirs []string
|
|
||||||
|
|
||||||
if db != nil {
|
|
||||||
movedDirs, err = db.findMovedDirs(f)
|
|
||||||
}
|
|
||||||
if err != nil {
|
|
||||||
return Error, err
|
|
||||||
} else {
|
|
||||||
switch len(movedDirs) {
|
|
||||||
case 0:
|
|
||||||
return Error, fmt.Errorf("directory missing")
|
|
||||||
case 1:
|
|
||||||
f.updateDir(movedDirs[0])
|
|
||||||
result |= DirUpdated
|
|
||||||
default:
|
|
||||||
return Error, fmt.Errorf("multiple directories found")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if _, err := os.Stat(f.Path); err != nil {
|
// Write the contents to the file
|
||||||
if errors.Is(err, os.ErrNotExist) {
|
if err := os.WriteFile(file.Path, []byte(file.contents), 0644); err != nil {
|
||||||
if err := os.WriteFile(f.Path, []byte(f.contents), 0644); err != nil {
|
return fmt.Errorf("failed to write file: %w", err)
|
||||||
return Error, fmt.Errorf("failed to write file: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
return result | Restored, nil
|
|
||||||
} else {
|
|
||||||
return Error, err
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
// File exists, check its hash
|
|
||||||
contents, err := os.ReadFile(f.Path)
|
|
||||||
if err != nil {
|
|
||||||
return Error, fmt.Errorf("failed to read file for SHA comparison: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
hash := sha256.Sum256(contents)
|
|
||||||
currentSha := fmt.Sprintf("%x", hash)
|
|
||||||
|
|
||||||
// Compare the hashes
|
|
||||||
if currentSha == f.Sha256 {
|
|
||||||
// No op, or DirUpdated
|
|
||||||
return result, nil
|
|
||||||
} else {
|
|
||||||
switch dir {
|
|
||||||
case TrustDatabase:
|
|
||||||
if err := os.WriteFile(f.Path, []byte(f.contents), 0644); err != nil {
|
|
||||||
return Error, fmt.Errorf("failed to write file: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
return result | Restored, nil
|
|
||||||
case TrustFilesystem:
|
|
||||||
// Overwrite the database
|
|
||||||
if err = f.Backup(); err != nil {
|
|
||||||
return Error, err
|
|
||||||
} else {
|
|
||||||
return BackedUp, nil
|
|
||||||
}
|
|
||||||
default:
|
|
||||||
panic("unknown sync direction")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (f *EnvFile) sharesRemote(remotes []string) bool {
|
|
||||||
rMap := make(map[string]bool)
|
|
||||||
for _, remote := range f.Remotes {
|
|
||||||
rMap[remote] = true
|
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, remote := range remotes {
|
return nil
|
||||||
if rMap[remote] {
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
|
|
||||||
func (f *EnvFile) updateDir(newDir string) {
|
|
||||||
f.Dir = newDir
|
|
||||||
f.Path = path.Join(newDir, path.Base(f.Path))
|
|
||||||
f.Remotes = getGitRemotes(newDir)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Try to reconcile the EnvFile with the filesystem.
|
// Try to reconcile the EnvFile with the filesystem.
|
||||||
//
|
//
|
||||||
// If Updated is returned, [Db.Insert] should be called on file.
|
// If Updated is returned, [Db.Insert] should be called on file.
|
||||||
func (file *EnvFile) Sync() (result EnvFileSyncResult, err error) {
|
func (file *EnvFile) Sync() (result EnvFileSyncResult, err error) {
|
||||||
return file.sync(TrustFilesystem, nil)
|
// Check if the path exists in the file system
|
||||||
|
_, err = os.Stat(file.Path)
|
||||||
|
if err == nil {
|
||||||
|
contents, err := os.ReadFile(file.Path)
|
||||||
|
if err != nil {
|
||||||
|
return Error, fmt.Errorf("failed to read file for SHA comparison: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check if sha matches by reading the current file and calculating its hash
|
||||||
|
hash := sha256.Sum256(contents)
|
||||||
|
currentSha := fmt.Sprintf("%x", hash)
|
||||||
|
if file.Sha256 == currentSha {
|
||||||
|
// Nothing to do
|
||||||
|
return Noop, nil
|
||||||
|
} else {
|
||||||
|
if err = file.Backup(); err != nil {
|
||||||
|
return Error, err
|
||||||
|
} else {
|
||||||
|
return Updated, nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
if err = file.Restore(); err != nil {
|
||||||
|
return Error, err
|
||||||
|
} else {
|
||||||
|
return Restored, nil
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Install the file into the file system. If the file already exists,
|
// Update the EnvFile using the file system
|
||||||
// it will be overwritten.
|
|
||||||
func (file EnvFile) Restore() error {
|
|
||||||
_, err := file.sync(TrustDatabase, nil)
|
|
||||||
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
// Update the EnvFile using the file system.
|
|
||||||
func (file *EnvFile) Backup() error {
|
func (file *EnvFile) Backup() error {
|
||||||
// Read the contents of the file
|
// Read the contents of the file
|
||||||
contents, err := os.ReadFile(file.Path)
|
contents, err := os.ReadFile(file.Path)
|
||||||
|
|||||||
@@ -1,20 +1,9 @@
|
|||||||
package app
|
package app
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
|
||||||
"os/exec"
|
"os/exec"
|
||||||
)
|
)
|
||||||
|
|
||||||
type MissingFeatureError struct {
|
|
||||||
feature AvailableFeatures
|
|
||||||
}
|
|
||||||
|
|
||||||
func (m *MissingFeatureError) Error() string {
|
|
||||||
return fmt.Sprintf("Missing \"%s\" feature", m.feature)
|
|
||||||
}
|
|
||||||
|
|
||||||
// TODO: Features should really be renamed to Binaries
|
|
||||||
|
|
||||||
// Represents which binaries are present in $PATH.
|
// Represents which binaries are present in $PATH.
|
||||||
// Used to fail safely when required features are unavailable
|
// Used to fail safely when required features are unavailable
|
||||||
type AvailableFeatures int
|
type AvailableFeatures int
|
||||||
@@ -24,7 +13,7 @@ const (
|
|||||||
// fd
|
// fd
|
||||||
Fd AvailableFeatures = 2
|
Fd AvailableFeatures = 2
|
||||||
// All features are present
|
// All features are present
|
||||||
All AvailableFeatures = Git | Fd
|
All AvailableFeatures = Git & Fd
|
||||||
)
|
)
|
||||||
|
|
||||||
// Checks for available features.
|
// Checks for available features.
|
||||||
@@ -41,20 +30,3 @@ func checkFeatures() (feats AvailableFeatures) {
|
|||||||
|
|
||||||
return feats
|
return feats
|
||||||
}
|
}
|
||||||
|
|
||||||
// Returns a MissingFeature error if the given features aren't present.
|
|
||||||
func (a AvailableFeatures) validateFeatures(features ...AvailableFeatures) error {
|
|
||||||
var missing AvailableFeatures
|
|
||||||
|
|
||||||
for _, feat := range features {
|
|
||||||
if a&feat == 0 {
|
|
||||||
missing |= feat
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if missing == 0 {
|
|
||||||
return nil
|
|
||||||
} else {
|
|
||||||
return &MissingFeatureError{missing}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|||||||
156
build.zig
156
build.zig
@@ -1,156 +0,0 @@
|
|||||||
const std = @import("std");
|
|
||||||
|
|
||||||
// Although this function looks imperative, it does not perform the build
|
|
||||||
// directly and instead it mutates the build graph (`b`) that will be then
|
|
||||||
// executed by an external runner. The functions in `std.Build` implement a DSL
|
|
||||||
// for defining build steps and express dependencies between them, allowing the
|
|
||||||
// build runner to parallelize the build automatically (and the cache system to
|
|
||||||
// know when a step doesn't need to be re-run).
|
|
||||||
pub fn build(b: *std.Build) void {
|
|
||||||
// Standard target options allow the person running `zig build` to choose
|
|
||||||
// what target to build for. Here we do not override the defaults, which
|
|
||||||
// means any target is allowed, and the default is native. Other options
|
|
||||||
// for restricting supported target set are available.
|
|
||||||
const target = b.standardTargetOptions(.{});
|
|
||||||
// Standard optimization options allow the person running `zig build` to select
|
|
||||||
// between Debug, ReleaseSafe, ReleaseFast, and ReleaseSmall. Here we do not
|
|
||||||
// set a preferred release mode, allowing the user to decide how to optimize.
|
|
||||||
const optimize = b.standardOptimizeOption(.{});
|
|
||||||
// It's also possible to define more custom flags to toggle optional features
|
|
||||||
// of this build script using `b.option()`. All defined flags (including
|
|
||||||
// target and optimize options) will be listed when running `zig build --help`
|
|
||||||
// in this directory.
|
|
||||||
|
|
||||||
// This creates a module, which represents a collection of source files alongside
|
|
||||||
// some compilation options, such as optimization mode and linked system libraries.
|
|
||||||
// Zig modules are the preferred way of making Zig code available to consumers.
|
|
||||||
// addModule defines a module that we intend to make available for importing
|
|
||||||
// to our consumers. We must give it a name because a Zig package can expose
|
|
||||||
// multiple modules and consumers will need to be able to specify which
|
|
||||||
// module they want to access.
|
|
||||||
const mod = b.addModule("envr", .{
|
|
||||||
// The root source file is the "entry point" of this module. Users of
|
|
||||||
// this module will only be able to access public declarations contained
|
|
||||||
// in this file, which means that if you have declarations that you
|
|
||||||
// intend to expose to consumers that were defined in other files part
|
|
||||||
// of this module, you will have to make sure to re-export them from
|
|
||||||
// the root file.
|
|
||||||
.root_source_file = b.path("src/root.zig"),
|
|
||||||
// Later on we'll use this module as the root module of a test executable
|
|
||||||
// which requires us to specify a target.
|
|
||||||
.target = target,
|
|
||||||
});
|
|
||||||
|
|
||||||
// Here we define an executable. An executable needs to have a root module
|
|
||||||
// which needs to expose a `main` function. While we could add a main function
|
|
||||||
// to the module defined above, it's sometimes preferable to split business
|
|
||||||
// logic and the CLI into two separate modules.
|
|
||||||
//
|
|
||||||
// If your goal is to create a Zig library for others to use, consider if
|
|
||||||
// it might benefit from also exposing a CLI tool. A parser library for a
|
|
||||||
// data serialization format could also bundle a CLI syntax checker, for example.
|
|
||||||
//
|
|
||||||
// If instead your goal is to create an executable, consider if users might
|
|
||||||
// be interested in also being able to embed the core functionality of your
|
|
||||||
// program in their own executable in order to avoid the overhead involved in
|
|
||||||
// subprocessing your CLI tool.
|
|
||||||
//
|
|
||||||
// If neither case applies to you, feel free to delete the declaration you
|
|
||||||
// don't need and to put everything under a single module.
|
|
||||||
const exe = b.addExecutable(.{
|
|
||||||
.name = "envr",
|
|
||||||
.root_module = b.createModule(.{
|
|
||||||
// b.createModule defines a new module just like b.addModule but,
|
|
||||||
// unlike b.addModule, it does not expose the module to consumers of
|
|
||||||
// this package, which is why in this case we don't have to give it a name.
|
|
||||||
.root_source_file = b.path("src/main.zig"),
|
|
||||||
// Target and optimization levels must be explicitly wired in when
|
|
||||||
// defining an executable or library (in the root module), and you
|
|
||||||
// can also hardcode a specific target for an executable or library
|
|
||||||
// definition if desireable (e.g. firmware for embedded devices).
|
|
||||||
.target = target,
|
|
||||||
.optimize = optimize,
|
|
||||||
// List of modules available for import in source files part of the
|
|
||||||
// root module.
|
|
||||||
.imports = &.{
|
|
||||||
// Here "envr" is the name you will use in your source code to
|
|
||||||
// import this module (e.g. `@import("envr")`). The name is
|
|
||||||
// repeated because you are allowed to rename your imports, which
|
|
||||||
// can be extremely useful in case of collisions (which can happen
|
|
||||||
// importing modules from different packages).
|
|
||||||
.{ .name = "envr", .module = mod },
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
});
|
|
||||||
|
|
||||||
// This declares intent for the executable to be installed into the
|
|
||||||
// install prefix when running `zig build` (i.e. when executing the default
|
|
||||||
// step). By default the install prefix is `zig-out/` but can be overridden
|
|
||||||
// by passing `--prefix` or `-p`.
|
|
||||||
b.installArtifact(exe);
|
|
||||||
|
|
||||||
// This creates a top level step. Top level steps have a name and can be
|
|
||||||
// invoked by name when running `zig build` (e.g. `zig build run`).
|
|
||||||
// This will evaluate the `run` step rather than the default step.
|
|
||||||
// For a top level step to actually do something, it must depend on other
|
|
||||||
// steps (e.g. a Run step, as we will see in a moment).
|
|
||||||
const run_step = b.step("run", "Run the app");
|
|
||||||
|
|
||||||
// This creates a RunArtifact step in the build graph. A RunArtifact step
|
|
||||||
// invokes an executable compiled by Zig. Steps will only be executed by the
|
|
||||||
// runner if invoked directly by the user (in the case of top level steps)
|
|
||||||
// or if another step depends on it, so it's up to you to define when and
|
|
||||||
// how this Run step will be executed. In our case we want to run it when
|
|
||||||
// the user runs `zig build run`, so we create a dependency link.
|
|
||||||
const run_cmd = b.addRunArtifact(exe);
|
|
||||||
run_step.dependOn(&run_cmd.step);
|
|
||||||
|
|
||||||
// By making the run step depend on the default step, it will be run from the
|
|
||||||
// installation directory rather than directly from within the cache directory.
|
|
||||||
run_cmd.step.dependOn(b.getInstallStep());
|
|
||||||
|
|
||||||
// This allows the user to pass arguments to the application in the build
|
|
||||||
// command itself, like this: `zig build run -- arg1 arg2 etc`
|
|
||||||
if (b.args) |args| {
|
|
||||||
run_cmd.addArgs(args);
|
|
||||||
}
|
|
||||||
|
|
||||||
// Creates an executable that will run `test` blocks from the provided module.
|
|
||||||
// Here `mod` needs to define a target, which is why earlier we made sure to
|
|
||||||
// set the releative field.
|
|
||||||
const mod_tests = b.addTest(.{
|
|
||||||
.root_module = mod,
|
|
||||||
});
|
|
||||||
|
|
||||||
// A run step that will run the test executable.
|
|
||||||
const run_mod_tests = b.addRunArtifact(mod_tests);
|
|
||||||
|
|
||||||
// Creates an executable that will run `test` blocks from the executable's
|
|
||||||
// root module. Note that test executables only test one module at a time,
|
|
||||||
// hence why we have to create two separate ones.
|
|
||||||
const exe_tests = b.addTest(.{
|
|
||||||
.root_module = exe.root_module,
|
|
||||||
});
|
|
||||||
|
|
||||||
// A run step that will run the second test executable.
|
|
||||||
const run_exe_tests = b.addRunArtifact(exe_tests);
|
|
||||||
|
|
||||||
// A top level step for running all tests. dependOn can be called multiple
|
|
||||||
// times and since the two run steps do not depend on one another, this will
|
|
||||||
// make the two of them run in parallel.
|
|
||||||
const test_step = b.step("test", "Run tests");
|
|
||||||
test_step.dependOn(&run_mod_tests.step);
|
|
||||||
test_step.dependOn(&run_exe_tests.step);
|
|
||||||
|
|
||||||
// Just like flags, top level steps are also listed in the `--help` menu.
|
|
||||||
//
|
|
||||||
// The Zig build system is entirely implemented in userland, which means
|
|
||||||
// that it cannot hook into private compiler APIs. All compilation work
|
|
||||||
// orchestrated by the build system will result in other Zig compiler
|
|
||||||
// subcommands being invoked with the right flags defined. You can observe
|
|
||||||
// these invocations when one fails (or you pass a flag to increase
|
|
||||||
// verbosity) to validate assumptions and diagnose problems.
|
|
||||||
//
|
|
||||||
// Lastly, the Zig build system is relatively simple and self-contained,
|
|
||||||
// and reading its source code will allow you to master it.
|
|
||||||
}
|
|
||||||
@@ -1,81 +0,0 @@
|
|||||||
.{
|
|
||||||
// This is the default name used by packages depending on this one. For
|
|
||||||
// example, when a user runs `zig fetch --save <url>`, this field is used
|
|
||||||
// as the key in the `dependencies` table. Although the user can choose a
|
|
||||||
// different name, most users will stick with this provided value.
|
|
||||||
//
|
|
||||||
// It is redundant to include "zig" in this name because it is already
|
|
||||||
// within the Zig package namespace.
|
|
||||||
.name = .envr,
|
|
||||||
// This is a [Semantic Version](https://semver.org/).
|
|
||||||
// In a future version of Zig it will be used for package deduplication.
|
|
||||||
.version = "0.0.0",
|
|
||||||
// Together with name, this represents a globally unique package
|
|
||||||
// identifier. This field is generated by the Zig toolchain when the
|
|
||||||
// package is first created, and then *never changes*. This allows
|
|
||||||
// unambiguous detection of one package being an updated version of
|
|
||||||
// another.
|
|
||||||
//
|
|
||||||
// When forking a Zig project, this id should be regenerated (delete the
|
|
||||||
// field and run `zig build`) if the upstream project is still maintained.
|
|
||||||
// Otherwise, the fork is *hostile*, attempting to take control over the
|
|
||||||
// original project's identity. Thus it is recommended to leave the comment
|
|
||||||
// on the following line intact, so that it shows up in code reviews that
|
|
||||||
// modify the field.
|
|
||||||
.fingerprint = 0xa89bf067266a3e10, // Changing this has security and trust implications.
|
|
||||||
// Tracks the earliest Zig version that the package considers to be a
|
|
||||||
// supported use case.
|
|
||||||
.minimum_zig_version = "0.16.0",
|
|
||||||
// This field is optional.
|
|
||||||
// Each dependency must either provide a `url` and `hash`, or a `path`.
|
|
||||||
// `zig build --fetch` can be used to fetch all dependencies of a package, recursively.
|
|
||||||
// Once all dependencies are fetched, `zig build` no longer requires
|
|
||||||
// internet connectivity.
|
|
||||||
.dependencies = .{
|
|
||||||
// See `zig fetch --save <url>` for a command-line interface for adding dependencies.
|
|
||||||
//.example = .{
|
|
||||||
// // When updating this field to a new URL, be sure to delete the corresponding
|
|
||||||
// // `hash`, otherwise you are communicating that you expect to find the old hash at
|
|
||||||
// // the new URL. If the contents of a URL change this will result in a hash mismatch
|
|
||||||
// // which will prevent zig from using it.
|
|
||||||
// .url = "https://example.com/foo.tar.gz",
|
|
||||||
//
|
|
||||||
// // This is computed from the file contents of the directory of files that is
|
|
||||||
// // obtained after fetching `url` and applying the inclusion rules given by
|
|
||||||
// // `paths`.
|
|
||||||
// //
|
|
||||||
// // This field is the source of truth; packages do not come from a `url`; they
|
|
||||||
// // come from a `hash`. `url` is just one of many possible mirrors for how to
|
|
||||||
// // obtain a package matching this `hash`.
|
|
||||||
// //
|
|
||||||
// // Uses the [multihash](https://multiformats.io/multihash/) format.
|
|
||||||
// .hash = "...",
|
|
||||||
//
|
|
||||||
// // When this is provided, the package is found in a directory relative to the
|
|
||||||
// // build root. In this case the package's hash is irrelevant and therefore not
|
|
||||||
// // computed. This field and `url` are mutually exclusive.
|
|
||||||
// .path = "foo",
|
|
||||||
//
|
|
||||||
// // When this is set to `true`, a package is declared to be lazily
|
|
||||||
// // fetched. This makes the dependency only get fetched if it is
|
|
||||||
// // actually used.
|
|
||||||
// .lazy = false,
|
|
||||||
//},
|
|
||||||
},
|
|
||||||
// Specifies the set of files and directories that are included in this package.
|
|
||||||
// Only files and directories listed here are included in the `hash` that
|
|
||||||
// is computed for this package. Only files listed here will remain on disk
|
|
||||||
// when using the zig package manager. As a rule of thumb, one should list
|
|
||||||
// files required for compilation plus any license(s).
|
|
||||||
// Paths are relative to the build root. Use the empty string (`""`) to refer to
|
|
||||||
// the build root itself.
|
|
||||||
// A directory listed here means that all files within, recursively, are included.
|
|
||||||
.paths = .{
|
|
||||||
"build.zig",
|
|
||||||
"build.zig.zon",
|
|
||||||
"src",
|
|
||||||
// For example...
|
|
||||||
//"LICENSE",
|
|
||||||
//"README.md",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
@@ -13,10 +13,9 @@ import (
|
|||||||
|
|
||||||
// backupCmd represents the backup command
|
// backupCmd represents the backup command
|
||||||
var backupCmd = &cobra.Command{
|
var backupCmd = &cobra.Command{
|
||||||
Use: "backup <path>",
|
Use: "backup <path>",
|
||||||
Short: "Import a .env file into envr",
|
Short: "Import a .env file into envr",
|
||||||
Aliases: []string{"add"},
|
Args: cobra.ExactArgs(1),
|
||||||
Args: cobra.ExactArgs(1),
|
|
||||||
// Long: `Long desc`
|
// Long: `Long desc`
|
||||||
RunE: func(cmd *cobra.Command, args []string) error {
|
RunE: func(cmd *cobra.Command, args []string) error {
|
||||||
path := args[0]
|
path := args[0]
|
||||||
@@ -28,11 +27,11 @@ var backupCmd = &cobra.Command{
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
} else {
|
} else {
|
||||||
defer db.Close()
|
defer db.Close(app.Write)
|
||||||
record := app.NewEnvFile(path)
|
record := app.NewEnvFile(path)
|
||||||
|
|
||||||
if err := db.Insert(record); err != nil {
|
if err := db.Insert(record); err != nil {
|
||||||
return err
|
panic(err)
|
||||||
} else {
|
} else {
|
||||||
fmt.Printf("Saved %s into the database", path)
|
fmt.Printf("Saved %s into the database", path)
|
||||||
return nil
|
return nil
|
||||||
|
|||||||
110
cmd/check.go
110
cmd/check.go
@@ -1,106 +1,48 @@
|
|||||||
package cmd
|
package cmd
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
|
||||||
"os"
|
"os"
|
||||||
"path/filepath"
|
|
||||||
|
|
||||||
|
"github.com/olekukonko/tablewriter"
|
||||||
"github.com/sbrow/envr/app"
|
"github.com/sbrow/envr/app"
|
||||||
"github.com/spf13/cobra"
|
"github.com/spf13/cobra"
|
||||||
)
|
)
|
||||||
|
|
||||||
var checkCmd = &cobra.Command{
|
var checkCmd = &cobra.Command{
|
||||||
Use: "check [path]",
|
Use: "check",
|
||||||
Short: "check if files in the current directory are backed up",
|
Short: "Check for missing binaries",
|
||||||
// TODO: Long description for new check command
|
Long: `envr relies on external binaries for certain functionality.
|
||||||
Args: cobra.MaximumNArgs(1),
|
|
||||||
|
The check command reports on which binaries are available and which are not.`,
|
||||||
RunE: func(cmd *cobra.Command, args []string) error {
|
RunE: func(cmd *cobra.Command, args []string) error {
|
||||||
// Accept an optional path arg, default to current working directory
|
|
||||||
var checkPath string
|
|
||||||
if len(args) > 0 {
|
|
||||||
checkPath = args[0]
|
|
||||||
} else {
|
|
||||||
cwd, err := os.Getwd()
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("failed to get current working directory: %w", err)
|
|
||||||
}
|
|
||||||
checkPath = cwd
|
|
||||||
}
|
|
||||||
|
|
||||||
// Get absolute path
|
|
||||||
absPath, err := filepath.Abs(checkPath)
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("failed to get absolute path: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Open database
|
|
||||||
db, err := app.Open()
|
db, err := app.Open()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to open database: %w", err)
|
return err
|
||||||
}
|
|
||||||
defer db.Close()
|
|
||||||
|
|
||||||
// Check if the path is a file or directory
|
|
||||||
info, err := os.Stat(absPath)
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("failed to stat path: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
var filesInPath []string
|
|
||||||
|
|
||||||
if info.IsDir() {
|
|
||||||
// Find .env files in the specified directory
|
|
||||||
if err := db.CanScan(); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
// Scan only the specified path for .env files
|
|
||||||
filesInPath, err = db.Scan([]string{absPath})
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("failed to scan path for env files: %w", err)
|
|
||||||
}
|
|
||||||
} else {
|
} else {
|
||||||
// Path is a file, just check this specific file
|
defer db.Close(app.ReadOnly)
|
||||||
filesInPath = []string{absPath}
|
features := db.Features()
|
||||||
}
|
|
||||||
|
|
||||||
// Get all backed up files from the database
|
table := tablewriter.NewWriter(os.Stdout)
|
||||||
envFiles, err := db.List()
|
table.Header([]string{"Feature", "Status"})
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("failed to list files from database: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Check which files are not backed up
|
// Check Git
|
||||||
var notBackedUp []string
|
if features&app.Git == 1 {
|
||||||
for _, file := range filesInPath {
|
table.Append([]string{"Git", "✓ Available"})
|
||||||
isBackedUp := false
|
|
||||||
for _, envFile := range envFiles {
|
|
||||||
if envFile.Path == file {
|
|
||||||
isBackedUp = true
|
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if !isBackedUp {
|
|
||||||
notBackedUp = append(notBackedUp, file)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Display results
|
|
||||||
if len(notBackedUp) == 0 {
|
|
||||||
if len(filesInPath) == 0 {
|
|
||||||
fmt.Println("No .env files found in the specified directory.")
|
|
||||||
} else {
|
} else {
|
||||||
fmt.Println("✓ All .env files in the directory are backed up.")
|
table.Append([]string{"Git", "✗ Missing"})
|
||||||
}
|
}
|
||||||
} else {
|
|
||||||
fmt.Printf("Found %d .env file(s) that are not backed up:\n", len(notBackedUp))
|
|
||||||
for _, file := range notBackedUp {
|
|
||||||
fmt.Printf(" %s\n", file)
|
|
||||||
}
|
|
||||||
fmt.Println("\nRun 'envr sync' to back up these files.")
|
|
||||||
}
|
|
||||||
|
|
||||||
return nil
|
// Check fd
|
||||||
|
if features&app.Fd == 1 {
|
||||||
|
table.Append([]string{"fd", "✓ Available"})
|
||||||
|
} else {
|
||||||
|
table.Append([]string{"fd", "✗ Missing"})
|
||||||
|
}
|
||||||
|
|
||||||
|
table.Render()
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
51
cmd/deps.go
51
cmd/deps.go
@@ -1,51 +0,0 @@
|
|||||||
package cmd
|
|
||||||
|
|
||||||
import (
|
|
||||||
"os"
|
|
||||||
|
|
||||||
"github.com/olekukonko/tablewriter"
|
|
||||||
"github.com/sbrow/envr/app"
|
|
||||||
"github.com/spf13/cobra"
|
|
||||||
)
|
|
||||||
|
|
||||||
var depsCmd = &cobra.Command{
|
|
||||||
Use: "deps",
|
|
||||||
Short: "Check for missing binaries",
|
|
||||||
Long: `envr relies on external binaries for certain functionality.
|
|
||||||
|
|
||||||
The check command reports on which binaries are available and which are not.`,
|
|
||||||
RunE: func(cmd *cobra.Command, args []string) error {
|
|
||||||
db, err := app.Open()
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
} else {
|
|
||||||
defer db.Close()
|
|
||||||
features := db.Features()
|
|
||||||
|
|
||||||
table := tablewriter.NewWriter(os.Stdout)
|
|
||||||
table.Header([]string{"Feature", "Status"})
|
|
||||||
|
|
||||||
// Check Git
|
|
||||||
if features&app.Git == 1 {
|
|
||||||
table.Append([]string{"Git", "✓ Available"})
|
|
||||||
} else {
|
|
||||||
table.Append([]string{"Git", "✗ Missing"})
|
|
||||||
}
|
|
||||||
|
|
||||||
// Check fd
|
|
||||||
if features&app.Fd == app.Fd {
|
|
||||||
table.Append([]string{"fd", "✓ Available"})
|
|
||||||
} else {
|
|
||||||
table.Append([]string{"fd", "✗ Missing"})
|
|
||||||
}
|
|
||||||
|
|
||||||
table.Render()
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
func init() {
|
|
||||||
rootCmd.AddCommand(depsCmd)
|
|
||||||
}
|
|
||||||
19
cmd/init.go
19
cmd/init.go
@@ -11,9 +11,11 @@ import (
|
|||||||
"github.com/spf13/cobra"
|
"github.com/spf13/cobra"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// TODO: Add --force (-f) flag.
|
||||||
var initCmd = &cobra.Command{
|
var initCmd = &cobra.Command{
|
||||||
Use: "init",
|
Use: "init",
|
||||||
Short: "Set up envr",
|
DisableFlagsInUseLine: true,
|
||||||
|
Short: "Set up envr",
|
||||||
Long: `The init command generates your initial config and saves it to
|
Long: `The init command generates your initial config and saves it to
|
||||||
~/.envr/config in JSON format.
|
~/.envr/config in JSON format.
|
||||||
|
|
||||||
@@ -21,10 +23,11 @@ During setup, you will be prompted to select one or more ssh keys with which to
|
|||||||
encrypt your databse. **Make 100% sure** that you have **a remote copy** of this
|
encrypt your databse. **Make 100% sure** that you have **a remote copy** of this
|
||||||
key somewhere, otherwise your data could be lost forever.`,
|
key somewhere, otherwise your data could be lost forever.`,
|
||||||
RunE: func(cmd *cobra.Command, args []string) error {
|
RunE: func(cmd *cobra.Command, args []string) error {
|
||||||
force, _ := cmd.Flags().GetBool("force")
|
|
||||||
config, _ := app.LoadConfig()
|
config, _ := app.LoadConfig()
|
||||||
|
|
||||||
if config == nil || force {
|
if config != nil {
|
||||||
|
return fmt.Errorf("You have already initialized envr")
|
||||||
|
} else {
|
||||||
keys, err := selectSSHKeys()
|
keys, err := selectSSHKeys()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("Error selecting SSH keys: %v", err)
|
return fmt.Errorf("Error selecting SSH keys: %v", err)
|
||||||
@@ -40,17 +43,13 @@ key somewhere, otherwise your data could be lost forever.`,
|
|||||||
}
|
}
|
||||||
|
|
||||||
fmt.Printf("Config initialized with %d SSH key(s). You are ready to use envr.\n", len(keys))
|
fmt.Printf("Config initialized with %d SSH key(s). You are ready to use envr.\n", len(keys))
|
||||||
return nil
|
|
||||||
} else {
|
|
||||||
return fmt.Errorf(`You have already initialized envr.
|
|
||||||
Run again with the --force flag if you want to reinitialize.
|
|
||||||
`)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
initCmd.Flags().BoolP("force", "f", false, "Overwrite an existing config")
|
|
||||||
rootCmd.AddCommand(initCmd)
|
rootCmd.AddCommand(initCmd)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -24,7 +24,7 @@ var listCmd = &cobra.Command{
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
defer db.Close()
|
defer db.Close(app.ReadOnly)
|
||||||
|
|
||||||
rows, err := db.List()
|
rows, err := db.List()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|||||||
@@ -25,7 +25,7 @@ var removeCmd = &cobra.Command{
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
} else {
|
} else {
|
||||||
defer db.Close()
|
defer db.Close(app.Write)
|
||||||
if err := db.Delete(path); err != nil {
|
if err := db.Delete(path); err != nil {
|
||||||
return err
|
return err
|
||||||
} else {
|
} else {
|
||||||
|
|||||||
@@ -27,7 +27,7 @@ var restoreCmd = &cobra.Command{
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
} else {
|
} else {
|
||||||
defer db.Close()
|
defer db.Close(app.ReadOnly)
|
||||||
record, err := db.Fetch(path)
|
record, err := db.Fetch(path)
|
||||||
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|||||||
@@ -28,7 +28,7 @@ var scanCmd = &cobra.Command{
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
files, err := db.Scan(nil)
|
files, err := db.Scan()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@@ -57,7 +57,7 @@ var scanCmd = &cobra.Command{
|
|||||||
|
|
||||||
// Close database with write mode to persist changes
|
// Close database with write mode to persist changes
|
||||||
if addedCount > 0 {
|
if addedCount > 0 {
|
||||||
err = db.Close()
|
err = db.Close(app.Write)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("Error saving changes: %v\n", err)
|
return fmt.Errorf("Error saving changes: %v\n", err)
|
||||||
} else {
|
} else {
|
||||||
@@ -65,7 +65,7 @@ var scanCmd = &cobra.Command{
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
err = db.Close()
|
err = db.Close(app.ReadOnly)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("Error closing database: %v\n", err)
|
return fmt.Errorf("Error closing database: %v\n", err)
|
||||||
}
|
}
|
||||||
|
|||||||
67
cmd/sync.go
67
cmd/sync.go
@@ -1,93 +1,54 @@
|
|||||||
package cmd
|
package cmd
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"encoding/json"
|
"fmt"
|
||||||
"os"
|
|
||||||
|
|
||||||
"github.com/mattn/go-isatty"
|
|
||||||
"github.com/olekukonko/tablewriter"
|
|
||||||
"github.com/sbrow/envr/app"
|
"github.com/sbrow/envr/app"
|
||||||
"github.com/spf13/cobra"
|
"github.com/spf13/cobra"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// TODO: Detect when file paths have moved and update accordingly.
|
||||||
var syncCmd = &cobra.Command{
|
var syncCmd = &cobra.Command{
|
||||||
Use: "sync",
|
Use: "sync",
|
||||||
Short: "Update or restore your env backups",
|
Short: "Update or restore your env backups",
|
||||||
RunE: func(cmd *cobra.Command, args []string) error {
|
RunE: func(cmd *cobra.Command, args []string) error {
|
||||||
db, err := app.Open()
|
db, err := app.Open()
|
||||||
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
} else {
|
} else {
|
||||||
defer db.Close()
|
defer db.Close(app.Write)
|
||||||
files, err := db.List()
|
files, err := db.List()
|
||||||
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
} else {
|
} else {
|
||||||
type syncResult struct {
|
|
||||||
Path string `json:"path"`
|
|
||||||
Status string `json:"status"`
|
|
||||||
}
|
|
||||||
var results []syncResult
|
|
||||||
|
|
||||||
for _, file := range files {
|
for _, file := range files {
|
||||||
// Syncronize the filesystem with the database.
|
fmt.Printf("%s\n", file.Path)
|
||||||
oldPath := file.Path
|
|
||||||
changed, err := db.Sync(&file)
|
// Syncronize the filesystem with the database.
|
||||||
|
changed, err := file.Sync()
|
||||||
|
|
||||||
var status string
|
|
||||||
switch changed {
|
switch changed {
|
||||||
case app.BackedUp:
|
case app.Updated:
|
||||||
status = "Backed Up"
|
fmt.Printf("File updated - changes saved\n")
|
||||||
if err := db.Insert(file); err != nil {
|
if err := db.Insert(file); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
case app.Restored:
|
case app.Restored:
|
||||||
fallthrough
|
fmt.Printf("File missing - restored backup\n")
|
||||||
case app.RestoredAndDirUpdated:
|
|
||||||
status = "Restored"
|
|
||||||
case app.Error:
|
case app.Error:
|
||||||
if err == nil {
|
if err == nil {
|
||||||
panic("err cannot be nil when Sync returns Error")
|
panic("err cannot be nil when Sync returns Error")
|
||||||
|
} else {
|
||||||
|
fmt.Printf("%s\n", err)
|
||||||
}
|
}
|
||||||
status = err.Error()
|
|
||||||
case app.Noop:
|
case app.Noop:
|
||||||
status = "OK"
|
fmt.Println("Nothing to do")
|
||||||
case app.DirUpdated:
|
|
||||||
status = "Moved"
|
|
||||||
default:
|
default:
|
||||||
panic("Unknown result")
|
panic("Unknown result")
|
||||||
}
|
}
|
||||||
|
|
||||||
if changed&app.DirUpdated == app.DirUpdated {
|
fmt.Println("")
|
||||||
if err := db.Delete(oldPath); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if db.UpdateRequired(changed) {
|
|
||||||
if err := db.Insert(file); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
results = append(results, syncResult{
|
|
||||||
Path: file.Path,
|
|
||||||
Status: status,
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
if isatty.IsTerminal(os.Stdout.Fd()) {
|
|
||||||
table := tablewriter.NewWriter(os.Stdout)
|
|
||||||
table.Header([]string{"File", "Status"})
|
|
||||||
|
|
||||||
for _, result := range results {
|
|
||||||
table.Append([]string{result.Path, result.Status})
|
|
||||||
}
|
|
||||||
table.Render()
|
|
||||||
} else {
|
|
||||||
encoder := json.NewEncoder(os.Stdout)
|
|
||||||
return encoder.Encode(results)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
|
|||||||
@@ -44,8 +44,7 @@ at before, restore your backup with:
|
|||||||
### SEE ALSO
|
### SEE ALSO
|
||||||
|
|
||||||
* [envr backup](envr_backup.md) - Import a .env file into envr
|
* [envr backup](envr_backup.md) - Import a .env file into envr
|
||||||
* [envr check](envr_check.md) - check if files in the current directory are backed up
|
* [envr check](envr_check.md) - Check for missing binaries
|
||||||
* [envr deps](envr_deps.md) - Check for missing binaries
|
|
||||||
* [envr edit-config](envr_edit-config.md) - Edit your config with your default editor
|
* [envr edit-config](envr_edit-config.md) - Edit your config with your default editor
|
||||||
* [envr init](envr_init.md) - Set up envr
|
* [envr init](envr_init.md) - Set up envr
|
||||||
* [envr list](envr_list.md) - View your tracked files
|
* [envr list](envr_list.md) - View your tracked files
|
||||||
|
|||||||
@@ -1,9 +1,15 @@
|
|||||||
## envr check
|
## envr check
|
||||||
|
|
||||||
check if files in the current directory are backed up
|
Check for missing binaries
|
||||||
|
|
||||||
|
### Synopsis
|
||||||
|
|
||||||
|
envr relies on external binaries for certain functionality.
|
||||||
|
|
||||||
|
The check command reports on which binaries are available and which are not.
|
||||||
|
|
||||||
```
|
```
|
||||||
envr check [path] [flags]
|
envr check [flags]
|
||||||
```
|
```
|
||||||
|
|
||||||
### Options
|
### Options
|
||||||
|
|||||||
@@ -1,24 +0,0 @@
|
|||||||
## envr deps
|
|
||||||
|
|
||||||
Check for missing binaries
|
|
||||||
|
|
||||||
### Synopsis
|
|
||||||
|
|
||||||
envr relies on external binaries for certain functionality.
|
|
||||||
|
|
||||||
The check command reports on which binaries are available and which are not.
|
|
||||||
|
|
||||||
```
|
|
||||||
envr deps [flags]
|
|
||||||
```
|
|
||||||
|
|
||||||
### Options
|
|
||||||
|
|
||||||
```
|
|
||||||
-h, --help help for deps
|
|
||||||
```
|
|
||||||
|
|
||||||
### SEE ALSO
|
|
||||||
|
|
||||||
* [envr](envr.md) - Manage your .env files.
|
|
||||||
|
|
||||||
@@ -12,14 +12,13 @@ encrypt your databse. **Make 100% sure** that you have **a remote copy** of this
|
|||||||
key somewhere, otherwise your data could be lost forever.
|
key somewhere, otherwise your data could be lost forever.
|
||||||
|
|
||||||
```
|
```
|
||||||
envr init [flags]
|
envr init
|
||||||
```
|
```
|
||||||
|
|
||||||
### Options
|
### Options
|
||||||
|
|
||||||
```
|
```
|
||||||
-f, --force Overwrite an existing config
|
-h, --help help for init
|
||||||
-h, --help help for init
|
|
||||||
```
|
```
|
||||||
|
|
||||||
### SEE ALSO
|
### SEE ALSO
|
||||||
|
|||||||
30
flake.lock
generated
30
flake.lock
generated
@@ -5,11 +5,11 @@
|
|||||||
"nixpkgs-lib": "nixpkgs-lib"
|
"nixpkgs-lib": "nixpkgs-lib"
|
||||||
},
|
},
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1768135262,
|
"lastModified": 1751413152,
|
||||||
"narHash": "sha256-PVvu7OqHBGWN16zSi6tEmPwwHQ4rLPU9Plvs8/1TUBY=",
|
"narHash": "sha256-Tyw1RjYEsp5scoigs1384gIg6e0GoBVjms4aXFfRssQ=",
|
||||||
"owner": "hercules-ci",
|
"owner": "hercules-ci",
|
||||||
"repo": "flake-parts",
|
"repo": "flake-parts",
|
||||||
"rev": "80daad04eddbbf5a4d883996a73f3f542fa437ac",
|
"rev": "77826244401ea9de6e3bac47c2db46005e1f30b5",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
@@ -20,11 +20,11 @@
|
|||||||
},
|
},
|
||||||
"nixpkgs": {
|
"nixpkgs": {
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1767313136,
|
"lastModified": 1761597516,
|
||||||
"narHash": "sha256-16KkgfdYqjaeRGBaYsNrhPRRENs0qzkQVUooNHtoy2w=",
|
"narHash": "sha256-wxX7u6D2rpkJLWkZ2E932SIvDJW8+ON/0Yy8+a5vsDU=",
|
||||||
"owner": "NixOS",
|
"owner": "NixOS",
|
||||||
"repo": "nixpkgs",
|
"repo": "nixpkgs",
|
||||||
"rev": "ac62194c3917d5f474c1a844b6fd6da2db95077d",
|
"rev": "daf6dc47aa4b44791372d6139ab7b25269184d55",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
@@ -36,11 +36,11 @@
|
|||||||
},
|
},
|
||||||
"nixpkgs-lib": {
|
"nixpkgs-lib": {
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1765674936,
|
"lastModified": 1751159883,
|
||||||
"narHash": "sha256-k00uTP4JNfmejrCLJOwdObYC9jHRrr/5M/a/8L2EIdo=",
|
"narHash": "sha256-urW/Ylk9FIfvXfliA1ywh75yszAbiTEVgpPeinFyVZo=",
|
||||||
"owner": "nix-community",
|
"owner": "nix-community",
|
||||||
"repo": "nixpkgs.lib",
|
"repo": "nixpkgs.lib",
|
||||||
"rev": "2075416fcb47225d9b68ac469a5c4801a9c4dd85",
|
"rev": "14a40a1d7fb9afa4739275ac642ed7301a9ba1ab",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
@@ -51,11 +51,11 @@
|
|||||||
},
|
},
|
||||||
"nixpkgs-unstable": {
|
"nixpkgs-unstable": {
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1768178648,
|
"lastModified": 1751949589,
|
||||||
"narHash": "sha256-kz/F6mhESPvU1diB7tOM3nLcBfQe7GU7GQCymRlTi/s=",
|
"narHash": "sha256-mgFxAPLWw0Kq+C8P3dRrZrOYEQXOtKuYVlo9xvPntt8=",
|
||||||
"owner": "NixOS",
|
"owner": "NixOS",
|
||||||
"repo": "nixpkgs",
|
"repo": "nixpkgs",
|
||||||
"rev": "3fbab70c6e69c87ea2b6e48aa6629da2aa6a23b0",
|
"rev": "9b008d60392981ad674e04016d25619281550a9d",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
@@ -80,11 +80,11 @@
|
|||||||
]
|
]
|
||||||
},
|
},
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1768158989,
|
"lastModified": 1752055615,
|
||||||
"narHash": "sha256-67vyT1+xClLldnumAzCTBvU0jLZ1YBcf4vANRWP3+Ak=",
|
"narHash": "sha256-19m7P4O/Aw/6+CzncWMAJu89JaKeMh3aMle1CNQSIwM=",
|
||||||
"owner": "numtide",
|
"owner": "numtide",
|
||||||
"repo": "treefmt-nix",
|
"repo": "treefmt-nix",
|
||||||
"rev": "e96d59dff5c0d7fddb9d113ba108f03c3ef99eca",
|
"rev": "c9d477b5d5bd7f26adddd3f96cfd6a904768d4f9",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
|
|||||||
14
flake.nix
14
flake.nix
@@ -21,12 +21,7 @@
|
|||||||
imports = [
|
imports = [
|
||||||
inputs.treefmt-nix.flakeModule
|
inputs.treefmt-nix.flakeModule
|
||||||
];
|
];
|
||||||
systems = [
|
systems = [ "x86_64-linux" ];
|
||||||
"x86_64-linux"
|
|
||||||
"aarch64-linux"
|
|
||||||
|
|
||||||
"aarch64-darwin"
|
|
||||||
];
|
|
||||||
|
|
||||||
perSystem =
|
perSystem =
|
||||||
{ pkgs, system, inputs', ... }: {
|
{ pkgs, system, inputs', ... }: {
|
||||||
@@ -61,7 +56,7 @@
|
|||||||
|
|
||||||
packages.default = pkgs.buildGoModule rec {
|
packages.default = pkgs.buildGoModule rec {
|
||||||
pname = "envr";
|
pname = "envr";
|
||||||
version = "0.2.0";
|
version = "0.1.0";
|
||||||
src = ./.;
|
src = ./.;
|
||||||
# If the build complains, uncomment this line
|
# If the build complains, uncomment this line
|
||||||
# vendorHash = "sha256:0000000000000000000000000000000000000000000000000000";
|
# vendorHash = "sha256:0000000000000000000000000000000000000000000000000000";
|
||||||
@@ -97,11 +92,6 @@
|
|||||||
gotools
|
gotools
|
||||||
cobra-cli
|
cobra-cli
|
||||||
|
|
||||||
# Build tools
|
|
||||||
zip
|
|
||||||
|
|
||||||
opencode
|
|
||||||
|
|
||||||
# IDE
|
# IDE
|
||||||
unstable.helix
|
unstable.helix
|
||||||
typescript-language-server
|
typescript-language-server
|
||||||
|
|||||||
87
src/main.zig
87
src/main.zig
@@ -1,87 +0,0 @@
|
|||||||
const std = @import("std");
|
|
||||||
const Io = std.Io;
|
|
||||||
|
|
||||||
const envr = @import("envr");
|
|
||||||
|
|
||||||
const goBinary = "envr-go";
|
|
||||||
|
|
||||||
pub fn main(init: std.process.Init) !void {
|
|
||||||
// This is appropriate for anything that lives as long as the process.
|
|
||||||
const arena: std.mem.Allocator = init.arena.allocator();
|
|
||||||
|
|
||||||
const args = try init.minimal.args.toSlice(arena);
|
|
||||||
|
|
||||||
if (std.mem.eql(u8, args[1], "version")) {
|
|
||||||
version(args[1..]);
|
|
||||||
} else {
|
|
||||||
return fallbackToGo(init.io, args, arena);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
fn version(args: []const [:0]const u8) void {
|
|
||||||
// std.debug.print("hello from Zig!\n", .{});
|
|
||||||
|
|
||||||
// for (args[1..]) |arg| {
|
|
||||||
// std.debug.print("arg: {s}\n", .{arg});
|
|
||||||
// }
|
|
||||||
//
|
|
||||||
|
|
||||||
_ = args;
|
|
||||||
|
|
||||||
std.debug.print("TODO: Implement\n", .{});
|
|
||||||
}
|
|
||||||
|
|
||||||
fn fallbackToGo(
|
|
||||||
io: Io,
|
|
||||||
args: []const [:0]const u8,
|
|
||||||
arena: std.mem.Allocator,
|
|
||||||
) std.process.ReplaceError {
|
|
||||||
// Remap args
|
|
||||||
var childArgs = try std.ArrayList([]const u8).initCapacity(arena, args.len);
|
|
||||||
childArgs.appendAssumeCapacity(goBinary);
|
|
||||||
|
|
||||||
for (args[1..]) |arg| {
|
|
||||||
childArgs.appendAssumeCapacity(arg);
|
|
||||||
}
|
|
||||||
|
|
||||||
return std.process.replace(io, .{ .argv = childArgs.items });
|
|
||||||
}
|
|
||||||
|
|
||||||
test "simple test" {
|
|
||||||
const gpa = std.testing.allocator;
|
|
||||||
var list: std.ArrayList(i32) = .empty;
|
|
||||||
defer list.deinit(gpa); // Try commenting this out and see if zig detects the memory leak!
|
|
||||||
try list.append(gpa, 42);
|
|
||||||
try std.testing.expectEqual(@as(i32, 42), list.pop());
|
|
||||||
}
|
|
||||||
|
|
||||||
test "fuzz example" {
|
|
||||||
try std.testing.fuzz({}, testOne, .{});
|
|
||||||
}
|
|
||||||
|
|
||||||
fn testOne(context: void, smith: *std.testing.Smith) !void {
|
|
||||||
_ = context;
|
|
||||||
// Try passing `--fuzz` to `zig build test` and see if it manages to fail this test case!
|
|
||||||
|
|
||||||
const gpa = std.testing.allocator;
|
|
||||||
var list: std.ArrayList(u8) = .empty;
|
|
||||||
defer list.deinit(gpa);
|
|
||||||
while (!smith.eos()) switch (smith.value(enum { add_data, dup_data })) {
|
|
||||||
.add_data => {
|
|
||||||
const slice = try list.addManyAsSlice(gpa, smith.value(u4));
|
|
||||||
smith.bytes(slice);
|
|
||||||
},
|
|
||||||
.dup_data => {
|
|
||||||
if (list.items.len == 0) continue;
|
|
||||||
if (list.items.len > std.math.maxInt(u32)) return error.SkipZigTest;
|
|
||||||
const len = smith.valueRangeAtMost(u32, 1, @min(32, list.items.len));
|
|
||||||
const off = smith.valueRangeAtMost(u32, 0, @intCast(list.items.len - len));
|
|
||||||
try list.appendSlice(gpa, list.items[off..][0..len]);
|
|
||||||
try std.testing.expectEqualSlices(
|
|
||||||
u8,
|
|
||||||
list.items[off..][0..len],
|
|
||||||
list.items[list.items.len - len ..],
|
|
||||||
);
|
|
||||||
},
|
|
||||||
};
|
|
||||||
}
|
|
||||||
18
src/root.zig
18
src/root.zig
@@ -1,18 +0,0 @@
|
|||||||
//! By convention, root.zig is the root source file when making a package.
|
|
||||||
const std = @import("std");
|
|
||||||
const Io = std.Io;
|
|
||||||
|
|
||||||
/// This is a documentation comment to explain the `printAnotherMessage` function below.
|
|
||||||
///
|
|
||||||
/// Accepting an `Io.Writer` instance is a handy way to write reusable code.
|
|
||||||
pub fn printAnotherMessage(writer: *Io.Writer) Io.Writer.Error!void {
|
|
||||||
try writer.print("Run `zig build test` to run the tests.\n", .{});
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn add(a: i32, b: i32) i32 {
|
|
||||||
return a + b;
|
|
||||||
}
|
|
||||||
|
|
||||||
test "basic add functionality" {
|
|
||||||
try std.testing.expect(add(3, 7) == 10);
|
|
||||||
}
|
|
||||||
Reference in New Issue
Block a user