Files
troubleshoot/pkg/supportbundle/supportbundle.go
Andrew Lavery e8bf6435e4 add a '--metadata' flag to support-bundle (#1993)
* add a '--metadata' flag to support-bundle

* test the metadata flag e2e
2026-03-12 13:30:28 -04:00

475 lines
17 KiB
Go

package supportbundle
import (
"bytes"
"context"
"encoding/json"
"fmt"
"net/http"
"os"
"path/filepath"
"strings"
"time"
cursor "github.com/ahmetalpbalkan/go-cursor"
"github.com/fatih/color"
"github.com/pkg/errors"
"github.com/replicatedhq/troubleshoot/internal/traces"
"github.com/replicatedhq/troubleshoot/internal/util"
analyzer "github.com/replicatedhq/troubleshoot/pkg/analyze"
troubleshootv1beta2 "github.com/replicatedhq/troubleshoot/pkg/apis/troubleshoot/v1beta2"
"github.com/replicatedhq/troubleshoot/pkg/collect"
"github.com/replicatedhq/troubleshoot/pkg/constants"
"github.com/replicatedhq/troubleshoot/pkg/convert"
"github.com/replicatedhq/troubleshoot/pkg/redact"
"github.com/replicatedhq/troubleshoot/pkg/version"
"go.opentelemetry.io/otel"
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
"k8s.io/client-go/kubernetes"
"k8s.io/client-go/rest"
"k8s.io/klog/v2"
)
type SupportBundleCreateOpts struct {
CollectorProgressCallback func(chan interface{}, string)
CollectWithoutPermissions bool
RemoteHostCollectTimeoutSeconds int
HttpClient *http.Client
KubernetesRestConfig *rest.Config
Namespace string
ProgressChan chan interface{}
SinceTime *time.Time
OutputPath string
Redact bool
FromCLI bool
RunHostCollectorsInPod bool
// Phase 4: Tokenization options
Tokenize bool // Enable intelligent tokenization
RedactionMapPath string // Path for redaction mapping file
EncryptRedactionMap bool // Encrypt the redaction mapping file
TokenPrefix string // Custom token prefix format
VerifyTokenization bool // Validation mode only
BundleID string // Custom bundle identifier
TokenizationStats bool // Include detailed tokenization statistics
UserMetadata map[string]string // User-provided key=value metadata pairs
}
type SupportBundleResponse struct {
AnalyzerResults []*analyzer.AnalyzeResult
ArchivePath string
FileUploaded bool
// Phase 4: Tokenization response data
TokenizationEnabled bool // Whether tokenization was used
RedactionMapPath string // Path to generated redaction mapping file
TokenizationStats *redact.RedactionStats // Detailed tokenization statistics
BundleID string // Bundle identifier for correlation
}
// NodeList is a list of remote nodes to collect data from in a support bundle
type NodeList struct {
Nodes []string `json:"nodes"`
}
// CollectSupportBundleFromSpec collects support bundle from start to finish, including running
// collectors, analyzers and after collection steps. Input arguments are specifications.
// if FromCLI option is set to true, the output is the name of the archive on disk in the cwd.
// if FromCLI option is set to false, the support bundle is archived in the OS temp folder (os.TempDir()).
func CollectSupportBundleFromSpec(
spec *troubleshootv1beta2.SupportBundleSpec, additionalRedactors *troubleshootv1beta2.Redactor, opts SupportBundleCreateOpts,
) (*SupportBundleResponse, error) {
resultsResponse := SupportBundleResponse{}
if opts.KubernetesRestConfig == nil {
return nil, errors.New("did not receive kube rest config")
}
if opts.ProgressChan == nil {
return nil, errors.New("did not receive collector progress chan")
}
tmpDir, err := os.MkdirTemp("", "supportbundle")
if err != nil {
return nil, errors.Wrap(err, "create temp dir")
}
defer os.RemoveAll(tmpDir)
klog.V(2).Infof("Support bundle created in temporary directory: %s", tmpDir)
basename := ""
if opts.OutputPath != "" {
// use override output path
overridePath, err := convert.ValidateOutputPath(opts.OutputPath)
if err != nil {
return nil, errors.Wrap(err, "override output file path")
}
basename = strings.TrimSuffix(overridePath, ".tar.gz")
} else {
// use default output path
basename = fmt.Sprintf("support-bundle-%s", time.Now().Format("2006-01-02T15_04_05"))
if !opts.FromCLI {
basename = filepath.Join(os.TempDir(), basename)
}
}
filename, err := findFileName(basename, "tar.gz")
if err != nil {
return nil, errors.Wrap(err, "find file name")
}
resultsResponse.ArchivePath = filename
bundlePath := filepath.Join(tmpDir, strings.TrimSuffix(filename, ".tar.gz"))
if err := os.MkdirAll(bundlePath, 0777); err != nil {
return nil, errors.Wrap(err, "create bundle dir")
}
result := make(collect.CollectorResult)
ctx, root := otel.Tracer(constants.LIB_TRACER_NAME).Start(
context.Background(), constants.TROUBLESHOOT_ROOT_SPAN_NAME,
)
defer func() {
// If this function returns an error, root.End() may not be called.
// We want to ensure this happens, so we defer it. It is safe to call
// root.End() multiple times.
root.End()
}()
// Cache error returned by collectors and return it at the end of the function
// so as to have a chance to run analyzers and archive the support bundle after.
// If both host and in cluster collectors fail, the errors will be wrapped
collectorsErrs := []string{}
var files, hostFiles collect.CollectorResult
if spec.HostCollectors != nil {
// Run host collectors
hostFiles, err = runHostCollectors(ctx, spec.HostCollectors, additionalRedactors, bundlePath, opts)
if err != nil {
collectorsErrs = append(collectorsErrs, fmt.Sprintf("failed to run host collectors: %s", err))
}
}
if spec.Collectors != nil {
// Run collectors
files, err = runCollectors(ctx, spec.Collectors, additionalRedactors, bundlePath, opts)
if err != nil {
collectorsErrs = append(collectorsErrs, fmt.Sprintf("failed to run collectors: %s", err))
}
}
// merge in-cluster and host collectors results
for k, v := range files {
result[k] = v
}
for k, v := range hostFiles {
result[k] = v
}
if len(result) == 0 {
if len(collectorsErrs) > 0 {
return nil, fmt.Errorf("failed to generate support bundle: %s", strings.Join(collectorsErrs, "\n"))
}
return nil, fmt.Errorf("failed to generate support bundle")
}
version, err := version.GetVersionFile()
if err != nil {
return nil, errors.Wrap(err, "failed to get version file")
}
err = result.SaveResult(bundlePath, constants.VERSION_FILENAME, bytes.NewBuffer([]byte(version)))
if err != nil {
return nil, errors.Wrap(err, "failed to write version")
}
if len(opts.UserMetadata) > 0 {
metadataJSON, err := json.MarshalIndent(opts.UserMetadata, "", " ")
if err != nil {
return nil, errors.Wrap(err, "failed to marshal user metadata")
}
err = result.SaveResult(bundlePath, "metadata/user.json", bytes.NewBuffer(metadataJSON))
if err != nil {
return nil, errors.Wrap(err, "failed to write user metadata")
}
}
// Run Analyzers
analyzeResults, err := AnalyzeSupportBundle(ctx, spec, bundlePath)
if err != nil {
if opts.FromCLI {
c := color.New(color.FgHiRed)
c.Printf("%s\r * %v\n", cursor.ClearEntireLine(), err)
// don't die
} else {
return nil, errors.Wrap(err, "failed to run analysis")
}
}
resultsResponse.AnalyzerResults = analyzeResults
analysis, err := getAnalysisFile(analyzeResults)
if err != nil {
return nil, errors.Wrap(err, "failed to get analysis file")
}
err = result.SaveResult(bundlePath, constants.ANALYSIS_FILENAME, analysis)
if err != nil {
return nil, errors.Wrap(err, "failed to write analysis")
}
// Complete tracing by ending the root span and collecting
// the summary of the traces. Store them in the support bundle.
root.End()
summary := traces.GetExporterInstance().GetSummary()
err = result.SaveResult(bundlePath, "execution-data/summary.txt", bytes.NewReader([]byte(summary)))
if err != nil {
// Don't fail the support bundle if we can't save the execution summary
klog.Errorf("failed to save execution summary file in the support bundle: %v", err)
}
// Phase 4: Process tokenization features
if err := processTokenizationFeatures(opts, bundlePath, &resultsResponse); err != nil {
if opts.FromCLI {
c := color.New(color.FgHiYellow)
c.Printf("%s\r * Warning: %v\n", cursor.ClearEntireLine(), err)
// Don't fail the support bundle, just warn
} else {
return nil, errors.Wrap(err, "failed to process tokenization features")
}
}
// Archive Support Bundle
if err := result.ArchiveBundle(bundlePath, filename); err != nil {
return nil, errors.Wrap(err, "create bundle file")
}
fileUploaded, err := ProcessSupportBundleAfterCollection(spec, filename)
if err != nil {
if opts.FromCLI {
c := color.New(color.FgHiRed)
c.Printf("%s\r * %v\n", cursor.ClearEntireLine(), err)
// don't die
} else {
return nil, errors.Wrap(err, "failed to process bundle after collection")
}
}
resultsResponse.FileUploaded = fileUploaded
if len(collectorsErrs) > 0 {
// TODO: Consider a collectors error type
// TODO: use errors.Join in go 1.20 (https://pkg.go.dev/errors#Join)
return &resultsResponse, fmt.Errorf("%s", strings.Join(collectorsErrs, "\n"))
}
return &resultsResponse, nil
}
// CollectSupportBundleFromURI collects support bundle from start to finish, including running
// collectors, analyzers and after collection steps. Input arguments are the URIs of the support bundle and redactor specs.
// The support bundle is archived in the OS temp folder (os.TempDir()).
func CollectSupportBundleFromURI(specURI string, redactorURIs []string, opts SupportBundleCreateOpts) (*SupportBundleResponse, error) {
supportBundle, err := GetSupportBundleFromURI(specURI)
if err != nil {
return nil, errors.Wrap(err, "could not bundle from URI")
}
redactors, err := GetRedactorsFromURIs(redactorURIs)
if err != nil {
return nil, err
}
additionalRedactors := &troubleshootv1beta2.Redactor{}
additionalRedactors.Spec.Redactors = redactors
return CollectSupportBundleFromSpec(&supportBundle.Spec, additionalRedactors, opts)
}
// ProcessSupportBundleAfterCollection performs the after collection actions, like Callbacks and sending the archive to a remote server.
func ProcessSupportBundleAfterCollection(spec *troubleshootv1beta2.SupportBundleSpec, archivePath string) (bool, error) {
fileUploaded := false
if len(spec.AfterCollection) > 0 {
for _, ac := range spec.AfterCollection {
if ac.UploadResultsTo != nil {
if err := uploadSupportBundle(ac.UploadResultsTo, archivePath); err != nil {
return false, errors.Wrap(err, "failed to upload support bundle")
} else {
fileUploaded = true
}
} else if ac.Callback != nil {
if err := callbackSupportBundleAPI(ac.Callback, archivePath); err != nil {
return false, errors.Wrap(err, "failed to notify API that support bundle has been uploaded")
}
}
}
}
return fileUploaded, nil
}
// processTokenizationFeatures handles tokenization-specific processing
func processTokenizationFeatures(opts SupportBundleCreateOpts, bundlePath string, response *SupportBundleResponse) error {
// Configure tokenization if enabled
if opts.Tokenize {
// Enable tokenization directly (safer than environment variables)
redact.EnableTokenization()
defer redact.DisableTokenization() // Always cleanup, even on error
// Configure custom tokenizer if needed
if err := configureTokenizer(opts); err != nil {
return errors.Wrap(err, "failed to configure tokenizer")
}
response.TokenizationEnabled = true
// Get tokenizer for statistics and mapping
tokenizer := redact.GetGlobalTokenizer()
response.BundleID = tokenizer.GetBundleID()
// Override with custom bundle ID if provided
if opts.BundleID != "" {
response.BundleID = opts.BundleID
}
// Generate redaction mapping file if requested
if opts.RedactionMapPath != "" {
profile := "support-bundle"
if opts.BundleID != "" {
profile = fmt.Sprintf("support-bundle-%s", opts.BundleID)
}
err := tokenizer.GenerateRedactionMapFile(profile, opts.RedactionMapPath, opts.EncryptRedactionMap)
if err != nil {
return errors.Wrap(err, "failed to generate redaction mapping file")
}
response.RedactionMapPath = opts.RedactionMapPath
if opts.FromCLI {
fmt.Printf("\n✅ Redaction mapping file generated: %s\n", opts.RedactionMapPath)
if opts.EncryptRedactionMap {
fmt.Printf("🔒 Mapping file is encrypted with AES-256\n")
}
}
}
// Include tokenization statistics if requested
if opts.TokenizationStats {
redactionMap := tokenizer.GetRedactionMap("support-bundle-stats")
response.TokenizationStats = &redactionMap.Stats
if opts.FromCLI {
printTokenizationStats(redactionMap.Stats)
}
}
}
return nil
}
// configureTokenizer configures the global tokenizer with CLI options
func configureTokenizer(opts SupportBundleCreateOpts) error {
_ = redact.GetGlobalTokenizer() // Get tokenizer to ensure it's initialized
// Apply custom token prefix if specified
if opts.TokenPrefix != "" {
// Validate format
if !strings.Contains(opts.TokenPrefix, "%s") {
return errors.Errorf("custom token prefix must contain %%s placeholders: %s", opts.TokenPrefix)
}
// Note: In a more complete implementation, we'd need to modify the tokenizer config
// For now, we validate but use the default format
fmt.Printf("📝 Custom token prefix validated: %s\n", opts.TokenPrefix)
}
// Apply custom bundle ID if specified
if opts.BundleID != "" {
// Note: In a more complete implementation, we'd set the bundle ID in the tokenizer
// For now, we'll use this in the response
fmt.Printf("🆔 Custom bundle ID: %s\n", opts.BundleID)
}
return nil
}
// printTokenizationStats prints detailed tokenization statistics
func printTokenizationStats(stats redact.RedactionStats) {
fmt.Printf("\n📊 Tokenization Statistics:\n")
fmt.Printf(" Total secrets processed: %d\n", stats.TotalSecrets)
fmt.Printf(" Unique secrets: %d\n", stats.UniqueSecrets)
fmt.Printf(" Tokens generated: %d\n", stats.TokensGenerated)
fmt.Printf(" Files covered: %d\n", stats.FilesCovered)
fmt.Printf(" Duplicates detected: %d\n", stats.DuplicateCount)
fmt.Printf(" Correlations found: %d\n", stats.CorrelationCount)
totalLookups := stats.CacheHits + stats.CacheMisses
if totalLookups > 0 {
hitRate := float64(stats.CacheHits) / float64(totalLookups) * 100
fmt.Printf(" Cache hits: %d / %d (%.1f%% hit rate)\n", stats.CacheHits, totalLookups, hitRate)
} else {
fmt.Printf(" Cache hits: %d / %d (no lookups)\n", stats.CacheHits, totalLookups)
}
if len(stats.SecretsByType) > 0 {
fmt.Printf(" Secrets by type:\n")
for secretType, count := range stats.SecretsByType {
fmt.Printf(" %s: %d\n", secretType, count)
}
}
if len(stats.FileCoverage) > 0 {
fmt.Printf(" File coverage:\n")
for file, fileStats := range stats.FileCoverage {
fmt.Printf(" %s: %d secrets\n", file, fileStats.SecretsFound)
}
}
}
// AnalyzeSupportBundle performs analysis on a support bundle using the support bundle spec and an already unpacked support
// bundle on disk
func AnalyzeSupportBundle(ctx context.Context, spec *troubleshootv1beta2.SupportBundleSpec, tmpDir string) ([]*analyzer.AnalyzeResult, error) {
if len(spec.Analyzers) == 0 && len(spec.HostAnalyzers) == 0 {
return nil, nil
}
spec.Analyzers = analyzer.DedupAnalyzers(spec.Analyzers)
analyzeResults, err := analyzer.AnalyzeLocal(ctx, tmpDir, spec.Analyzers, spec.HostAnalyzers)
if err != nil {
return nil, errors.Wrap(err, "failed to analyze support bundle")
}
return analyzeResults, nil
}
// ConcatSpec the intention with these appends is to swap them out at a later date with more specific handlers for merging the spec fields
func ConcatSpec(target *troubleshootv1beta2.SupportBundle, source *troubleshootv1beta2.SupportBundle) *troubleshootv1beta2.SupportBundle {
if source == nil {
return target
}
var newBundle *troubleshootv1beta2.SupportBundle
if target == nil {
newBundle = source
} else {
newBundle = target.DeepCopy()
newBundle.Spec.Collectors = util.Append(target.Spec.Collectors, source.Spec.Collectors)
newBundle.Spec.AfterCollection = util.Append(target.Spec.AfterCollection, source.Spec.AfterCollection)
newBundle.Spec.HostCollectors = util.Append(target.Spec.HostCollectors, source.Spec.HostCollectors)
newBundle.Spec.HostAnalyzers = util.Append(target.Spec.HostAnalyzers, source.Spec.HostAnalyzers)
newBundle.Spec.Analyzers = util.Append(target.Spec.Analyzers, source.Spec.Analyzers)
// TODO: What to do with the Uri field?
}
return newBundle
}
func getNodeList(clientset kubernetes.Interface, opts SupportBundleCreateOpts) (*NodeList, error) {
// todo: any node filtering on opts?
nodes, err := clientset.CoreV1().Nodes().List(context.Background(), metav1.ListOptions{})
if err != nil {
return nil, errors.Wrap(err, "failed to list nodes")
}
nodeList := NodeList{}
for _, node := range nodes.Items {
nodeList.Nodes = append(nodeList.Nodes, node.Name)
}
return &nodeList, nil
}