Compare commits

...

17 Commits

Author SHA1 Message Date
Alon Girmonsky
1669680d10 🔖 Bump the Helm chart version to 52.12.0 2026-01-25 17:54:24 -08:00
Alon Girmonsky
19389fcba7 Bring back community license. (#1822) 2026-01-23 14:49:38 -08:00
Volodymyr Stoiko
1b027153e3 Increase dissector image limit (#1823)
Co-authored-by: Alon Girmonsky <1990761+alongir@users.noreply.github.com>
2026-01-21 19:52:45 -08:00
Volodymyr Stoiko
77d16e73e8 Migrate kubehq.com to kubeshark.com domain (#1824)
* Update labels

* Update kubeshark API url

* Update other domains

* comments

---------

Co-authored-by: Alon Girmonsky <1990761+alongir@users.noreply.github.com>
2026-01-21 19:23:50 -08:00
Volodymyr Stoiko
a73c904a9b Use console.kubeshark.com instead of kubehq (#1821) 2026-01-21 09:18:47 -08:00
Alon Girmonsky
8f3f136be6 Update README.md (#1820) 2026-01-20 17:16:02 -08:00
Alon Girmonsky
897aa44965 🔖 Bump the Helm chart version to 52.11.7 2026-01-15 13:23:40 -08:00
Alon Girmonsky
57093e8a25 Updated complementary license (#1819)
* Updated complementary license
To Feb 14th.

* Update README.md

shout out
2026-01-15 13:17:15 -08:00
Volodymyr Stoiko
1fd9dffc60 Add delayed dissection controller configuration (#1818)
* Allow managing jobs in kubeshark namespaces

* Extend with dissector image

* make dissection job resources configurable
2026-01-06 11:56:53 -08:00
Alon Girmonsky
3b315eb89f Update CONTRIBUTING.md (#1817) 2025-12-28 13:38:08 +02:00
Alon Girmonsky
6645e23704 🔖 Bump the Helm chart version to 52.11.0 2025-12-16 11:49:37 -08:00
Alon Girmonsky
b7190162ec In preparation for v200 (#1815)
* In preparation for v200

* updated README

* Enable raw capture

* changed 0.0.0.0 to 127.0.0.1
as 0.0.0.0 is insecure address

* added tip: kubeshark proxy

* added new TCP/UDP connection dissectors
Set API2 as the default

* increased storageLimit per worker.

* Updated makefile

* updated the complementary license
to the end of Jan 2026.

* readme touch ups

* Updated snapshot image

* updated license
removed dashboard subproject
2025-12-16 11:44:02 -08:00
Alon Girmonsky
9570b2e317 Update README.md (#1814) 2025-12-05 08:27:25 -08:00
Serhii Ponomarenko
b98113a2b5 🔨 Create raw-capture-enabled front env (#1813) 2025-12-01 16:02:19 -08:00
Alon Girmonsky
9724a0c279 🔖 Bump the Helm chart version to 52.10.3 2025-11-28 17:12:22 -08:00
Alon Girmonsky
47ac96a71b Adding a default license (#1812) 2025-11-28 17:06:48 -08:00
Serhii Ponomarenko
4dea643781 🚑 Use www.kubehq.com for links (#1809) 2025-11-26 08:14:40 -08:00
28 changed files with 241 additions and 170 deletions

View File

@@ -7,7 +7,7 @@ Please read and follow the guidelines below.
## Communication
* Before starting work on a major feature, please reach out to us via [GitHub](https://github.com/kubeshark/kubeshark), [Discord](https://discord.gg/WkvRGMUcx7), [Slack](https://join.slack.com/t/kubeshark/shared_invite/zt-1k3sybpq9-uAhFkuPJiJftKniqrGHGhg), [email](mailto:info@kubeshark.co), etc. We will make sure no one else is already working on it. A _major feature_ is defined as any change that is > 100 LOC altered (not including tests), or changes any user-facing behavior
* Before starting work on a major feature, please reach out to us via [GitHub](https://github.com/kubeshark/kubeshark), [Discord](https://discord.gg/WkvRGMUcx7), [Slack](https://join.slack.com/t/kubeshark/shared_invite/zt-1k3sybpq9-uAhFkuPJiJftKniqrGHGhg), [email](mailto:info@kubeshark.com), etc. We will make sure no one else is already working on it. A _major feature_ is defined as any change that is > 100 LOC altered (not including tests), or changes any user-facing behavior
* Small patches and bug fixes don't need prior communication.
## Contribution Requirements

View File

@@ -84,8 +84,10 @@ kubectl-view-kubeshark-resources: ## This command outputs all Kubernetes resourc
./kubectl.sh view-kubeshark-resources
generate-helm-values: ## Generate the Helm values from config.yaml
mv ~/.kubeshark/config.yaml ~/.kubeshark/config.yaml.old; bin/kubeshark__ config>helm-chart/values.yaml;mv ~/.kubeshark/config.yaml.old ~/.kubeshark/config.yaml
sed -i 's/^license:.*/license: ""/' helm-chart/values.yaml && sed -i '1i # find a detailed description here: https://github.com/kubeshark/kubeshark/blob/master/helm-chart/README.md' helm-chart/values.yaml
# [ -f ~/.kubeshark/config.yaml ] && mv ~/.kubeshark/config.yaml ~/.kubeshark/config.yaml.old
bin/kubeshark__ config>helm-chart/values.yaml
# [ -f ~/.kubeshark/config.yaml.old ] && mv ~/.kubeshark/config.yaml.old ~/.kubeshark/config.yaml
# sed -i 's/^license:.*/license: ""/' helm-chart/values.yaml && sed -i '1i # find a detailed description here: https://github.com/kubeshark/kubeshark/blob/master/helm-chart/README.md' helm-chart/values.yaml
generate-manifests: ## Generate the manifests from the Helm chart using default configuration
helm template kubeshark -n default ./helm-chart > ./manifests/complete.yaml
@@ -189,8 +191,8 @@ release:
@make generate-helm-values && make generate-manifests
@git add -A . && git commit -m ":bookmark: Bump the Helm chart version to $(VERSION)" && git push
@git tag -d v$(VERSION); git tag v$(VERSION) && git push origin --tags
@cd helm-chart && rm -rf ../../kubeshark.github.io/charts/chart && mkdir ../../kubeshark.github.io/charts/chart && cp -r . ../../kubeshark.github.io/charts/chart/
@cd ../../kubeshark.github.io/ && git add -A . && git commit -m ":sparkles: Update the Helm chart" && git push
@rm -rf ../kubeshark.github.io/charts/chart && mkdir ../kubeshark.github.io/charts/chart && cp -r helm-chart/ ../kubeshark.github.io/charts/chart/
@cd ../kubeshark.github.io/ && git add -A . && git commit -m ":sparkles: Update the Helm chart" && git push
@cd ../kubeshark
release-dry-run:
@@ -198,11 +200,14 @@ release-dry-run:
@cd ../tracer && git checkout master && git pull
@cd ../hub && git checkout master && git pull
@cd ../front && git checkout master && git pull
@cd ../kubeshark && git checkout master && git pull && sed -i "s/^version:.*/version: \"$(shell echo $(VERSION) | sed -E 's/^([0-9]+\.[0-9]+\.[0-9]+)\..*/\1/')\"/" helm-chart/Chart.yaml && make
@cd ../kubeshark && sed -i "s/^version:.*/version: \"$(shell echo $(VERSION) | sed -E 's/^([0-9]+\.[0-9]+\.[0-9]+)\..*/\1/')\"/" helm-chart/Chart.yaml && make
@if [ "$(shell uname)" = "Darwin" ]; then \
codesign --sign - --force --preserve-metadata=entitlements,requirements,flags,runtime ./bin/kubeshark__; \
fi
@make generate-helm-values && make generate-manifests
@rm -rf ../kubeshark.github.io/charts/chart && mkdir ../kubeshark.github.io/charts/chart && cp -r helm-chart/ ../kubeshark.github.io/charts/chart/
@cd ../kubeshark.github.io/
@cd ../kubeshark
branch:
@cd ../worker && git checkout master && git pull && git checkout -b $(name); git push --set-upstream origin $(name)

View File

@@ -21,26 +21,40 @@
</p>
<p align="center">
<b>
We're currently experiencing issues and our team is working around the clock to resolve them as quickly as possible. We'll keep you updated once everything is back to normal. In the meantime, feel free to reach out to us on <a href="https://join.slack.com/t/kubeshark/shared_invite/zt-3jdcdgxdv-1qNkhBh9c6CFoE7bSPkpBQ">Slack</a> or email us at <a href="mailto:support@kubehq.io">support@kubehq.io</a>.
Want to see Kubeshark in action right now? Visit this
<a href="https://demo.kubeshark.com/">live demo deployment</a> of Kubeshark.
</b>
</p>
**Kubeshark** is a network observability platform for Kubernetes, providing real-time, cluster-wide visibility into Kubernetes network. It enables users to inspect all internal and external cluster communications, API calls, and data in transit. Additionally, Kubeshark detects anomalies and emergent behaviors, trigger autonomous remediations, and generate deep network insights.
**Kubeshark** is an API traffic analyzer for Kubernetes, providing deep packet inspection with complete API and Kubernetes contexts, retaining cluster-wide L4 traffic (PCAP), and using minimal production compute resources.
![Simple UI](https://github.com/kubeshark/assets/raw/master/png/kubeshark-ui.png)
Think [TCPDump](https://en.wikipedia.org/wiki/Tcpdump) and [Wireshark](https://www.wireshark.org/) reimagined for Kubernetes.
Access cluster-wide PCAP traffic by pressing a single button, without the need to install `tcpdump` or manually copy files. Understand the traffic context in relation to the API and Kubernetes contexts.
#### Service-Map w/Kubernetes Context
![Service Map with Kubernetes Context](https://github.com/kubeshark/assets/raw/master/png/kubeshark-servicemap.png)
#### Cluster-Wide PCAP Recording
#### Export Cluster-Wide L4 Traffic (PCAP)
![Cluster-Wide PCAP Recording](https://github.com/kubeshark/assets/raw/master/png/pcap-recording.png)
Imagine having a cluster-wide [TCPDump](https://www.tcpdump.org/)-like capability—exporting a single [PCAP](https://www.ietf.org/archive/id/draft-gharris-opsawg-pcap-01.html) file that consolidates traffic from multiple nodes, all accessible with a single click.
1. Go to the **Snapshots** tab
2. Create a new snapshot
3. **Optionally** select the nodes (default: all nodes)
4. **Optionally** select the time frame (default: last one hour)
5. Press **Create**
<img width="3342" height="1206" alt="image" src="https://github.com/user-attachments/assets/e8e47996-52b7-4028-9698-f059a13ffdb7" />
Once the snapshot is ready, click the PCAP file to export its contents and open it in Wireshark.
## Getting Started
Download **Kubeshark**'s binary distribution [latest release](https://github.com/kubeshark/kubeshark/releases/latest) or use one of the following methods to deploy **Kubeshark**. The [web-based dashboard](https://docs.kubehq.com/en/ui) should open in your browser, showing a real-time view of your cluster's traffic.
Download **Kubeshark**'s binary distribution [latest release](https://github.com/kubeshark/kubeshark/releases/latest) or use one of the following methods to deploy **Kubeshark**. The [web-based dashboard](https://docs.kubeshark.com/en/ui) should open in your browser, showing a real-time view of your cluster's traffic.
### Homebrew
@@ -61,7 +75,7 @@ kubeshark clean
Add the Helm repository and install the chart:
```shell
helm repo add kubeshark https://helm.kubehq.com
helm repo add kubeshark https://helm.kubeshark.com
helm install kubeshark kubeshark/kubeshark
```
Follow the on-screen instructions how to connect to the dashboard.
@@ -77,7 +91,7 @@ Clone this repository and run the `make` command to build it. After the build is
## Documentation
To learn more, read the [documentation](https://docs.kubehq.com).
To learn more, read the [documentation](https://docs.kubeshark.com).
## Contributing

View File

@@ -24,7 +24,7 @@ import (
)
const (
label = "app.kubehq.com/app=worker"
label = "app.kubeshark.com/app=worker"
srcDir = "pcapdump"
maxSnaplen uint32 = 262144
maxTimePerFile = time.Minute * 5

View File

@@ -137,6 +137,8 @@ func CreateDefaultConfig() ConfigStruct {
"ldap",
"radius",
"diameter",
"udp-flow",
"tcp-flow",
},
PortMapping: configStructs.PortMapping{
HTTP: []uint16{80, 443, 8080},

View File

@@ -198,7 +198,7 @@ type RoutingConfig struct {
}
type DashboardConfig struct {
StreamingType string `yaml:"streamingType" json:"streamingType" default:""`
StreamingType string `yaml:"streamingType" json:"streamingType" default:"connect-rpc"`
CompleteStreamingEnabled bool `yaml:"completeStreamingEnabled" json:"completeStreamingEnabled" default:"true"`
}
@@ -207,7 +207,7 @@ type FrontRoutingConfig struct {
}
type ReleaseConfig struct {
Repo string `yaml:"repo" json:"repo" default:"https://helm.kubehq.com"`
Repo string `yaml:"repo" json:"repo" default:"https://helm.kubeshark.com"`
Name string `yaml:"name" json:"name" default:"kubeshark"`
Namespace string `yaml:"namespace" json:"namespace" default:"default"`
}
@@ -251,8 +251,8 @@ type PprofConfig struct {
type MiscConfig struct {
JsonTTL string `yaml:"jsonTTL" json:"jsonTTL" default:"5m"`
PcapTTL string `yaml:"pcapTTL" json:"pcapTTL" default:"10s"`
PcapErrorTTL string `yaml:"pcapErrorTTL" json:"pcapErrorTTL" default:"60s"`
PcapTTL string `yaml:"pcapTTL" json:"pcapTTL" default:"0"`
PcapErrorTTL string `yaml:"pcapErrorTTL" json:"pcapErrorTTL" default:"0"`
TrafficSampleRate int `yaml:"trafficSampleRate" json:"trafficSampleRate" default:"100"`
TcpStreamChannelTimeoutMs int `yaml:"tcpStreamChannelTimeoutMs" json:"tcpStreamChannelTimeoutMs" default:"10000"`
TcpStreamChannelTimeoutShow bool `yaml:"tcpStreamChannelTimeoutShow" json:"tcpStreamChannelTimeoutShow" default:"false"`
@@ -263,7 +263,7 @@ type MiscConfig struct {
}
type PcapDumpConfig struct {
PcapDumpEnabled bool `yaml:"enabled" json:"enabled" default:"true"`
PcapDumpEnabled bool `yaml:"enabled" json:"enabled" default:"false"`
PcapTimeInterval string `yaml:"timeInterval" json:"timeInterval" default:"1m"`
PcapMaxTime string `yaml:"maxTime" json:"maxTime" default:"1h"`
PcapMaxSize string `yaml:"maxSize" json:"maxSize" default:"500MB"`
@@ -301,13 +301,19 @@ type SeLinuxOptionsConfig struct {
}
type RawCaptureConfig struct {
Enabled bool `yaml:"enabled" json:"enabled" default:"false"`
Enabled bool `yaml:"enabled" json:"enabled" default:"true"`
StorageSize string `yaml:"storageSize" json:"storageSize" default:"1Gi"`
}
type SnapshotsConfig struct {
StorageClass string `yaml:"storageClass" json:"storageClass" default:""`
StorageSize string `yaml:"storageSize" json:"storageSize" default:"1Gi"`
StorageSize string `yaml:"storageSize" json:"storageSize" default:"20Gi"`
}
type DelayedDissectionConfig struct {
Image string `yaml:"image" json:"image" default:"kubeshark/worker:master"`
CPU string `yaml:"cpu" json:"cpu" default:"1"`
Memory string `yaml:"memory" json:"memory" default:"4Gi"`
}
type CaptureConfig struct {
@@ -325,6 +331,7 @@ type TapConfig struct {
ExcludedNamespaces []string `yaml:"excludedNamespaces" json:"excludedNamespaces" default:"[]"`
BpfOverride string `yaml:"bpfOverride" json:"bpfOverride" default:""`
Capture CaptureConfig `yaml:"capture" json:"capture"`
DelayedDissection DelayedDissectionConfig `yaml:"delayedDissection" json:"delayedDissection"`
Snapshots SnapshotsConfig `yaml:"snapshots" json:"snapshots"`
Release ReleaseConfig `yaml:"release" json:"release"`
PersistentStorage bool `yaml:"persistentStorage" json:"persistentStorage" default:"false"`
@@ -332,7 +339,7 @@ type TapConfig struct {
PersistentStoragePvcVolumeMode string `yaml:"persistentStoragePvcVolumeMode" json:"persistentStoragePvcVolumeMode" default:"FileSystem"`
EfsFileSytemIdAndPath string `yaml:"efsFileSytemIdAndPath" json:"efsFileSytemIdAndPath" default:""`
Secrets []string `yaml:"secrets" json:"secrets" default:"[]"`
StorageLimit string `yaml:"storageLimit" json:"storageLimit" default:"5Gi"`
StorageLimit string `yaml:"storageLimit" json:"storageLimit" default:"10Gi"`
StorageClass string `yaml:"storageClass" json:"storageClass" default:"standard"`
DryRun bool `yaml:"dryRun" json:"dryRun" default:"false"`
DnsConfig DnsConfig `yaml:"dns" json:"dns"`

View File

@@ -1,8 +1,8 @@
apiVersion: v2
name: kubeshark
version: "52.10.0"
version: "52.12.0"
description: The API Traffic Analyzer for Kubernetes
home: https://kubehq.com
home: https://kubeshark.com
keywords:
- kubeshark
- packet capture
@@ -16,9 +16,9 @@ keywords:
- api
kubeVersion: '>= 1.16.0-0'
maintainers:
- email: info@kubehq.com
- email: support@kubeshark.com
name: Kubeshark
url: https://kubehq.com
url: https://kubeshark.com
sources:
- https://github.com/kubeshark/kubeshark/tree/master/helm-chart
type: application

View File

@@ -5,7 +5,7 @@
Add the Helm repo for Kubeshark:
```shell
helm repo add kubeshark https://helm.kubehq.com
helm repo add kubeshark https://helm.kubeshark.com
```
then install Kubeshark:
@@ -69,7 +69,7 @@ When it's necessary, you can use:
--set license=YOUR_LICENSE_GOES_HERE
```
Get your license from Kubeshark's [Admin Console](https://console.kubehq.com/).
Get your license from Kubeshark's [Admin Console](https://console.kubeshark.com/).
## Installing with Ingress (EKS) enabled
@@ -140,12 +140,12 @@ Example for overriding image names:
| `tap.bpfOverride` | When using AF_PACKET as a traffic capture backend, override any existing pod targeting rules and set explicit BPF expression (e.g. `net 0.0.0.0/0`). | `[]` |
| `tap.capture.stopped` | Set to `false` to have traffic processing start automatically. When set to `true`, traffic processing is stopped by default, resulting in almost no resource consumption (e.g. Kubeshark is dormant). This property can be dynamically control via the dashboard. | `false` |
| `tap.capture.stopAfter` | Set to a duration (e.g. `30s`) to have traffic processing stop after no websocket activity between worker and hub. | `30s` |
| `tap.capture.raw.enabled` | Enable raw capture of packets and syscalls to disk for offline analysis | `false` |
| `tap.capture.raw.enabled` | Enable raw capture of packets and syscalls to disk for offline analysis | `true` |
| `tap.capture.raw.storageSize` | Maximum storage size for raw capture files (supports K8s quantity format: `1Gi`, `500Mi`, etc.) | `1Gi` |
| `tap.capture.dbMaxSize` | Maximum size for capture database (e.g., `4Gi`, `2000Mi`). When empty, automatically uses 80% of allocated storage (`tap.storageLimit`). | `""` |
| `tap.snapshots.storageClass` | Storage class for snapshots volume. When empty, uses `emptyDir`. When set, creates a PVC with this storage class | `""` |
| `tap.snapshots.storageSize` | Storage size for snapshots volume (supports K8s quantity format: `1Gi`, `500Mi`, etc.) | `1Gi` |
| `tap.release.repo` | URL of the Helm chart repository | `https://helm.kubehq.com` |
| `tap.snapshots.storageSize` | Storage size for snapshots volume (supports K8s quantity format: `1Gi`, `500Mi`, etc.) | `10Gi` |
| `tap.release.repo` | URL of the Helm chart repository | `https://helm.kubeshark.com` |
| `tap.release.name` | Helm release name | `kubeshark` |
| `tap.release.namespace` | Helm release namespace | `default` |
| `tap.persistentStorage` | Use `persistentVolumeClaim` instead of `emptyDir` | `false` |
@@ -221,7 +221,7 @@ Example for overriding image names:
| `tap.hostNetwork` | Enable host network mode for worker DaemonSet pods. When enabled, worker pods use the host's network namespace for direct network access. | `true` |
| `tap.gitops.enabled` | Enable GitOps functionality. This will allow you to use GitOps to manage your Kubeshark configuration. | `false` |
| `logs.file` | Logs dump path | `""` |
| `pcapdump.enabled` | Enable recording of all traffic captured according to other parameters. Whatever Kubeshark captures, considering pod targeting rules, will be stored in pcap files ready to be viewed by tools | `true` |
| `pcapdump.enabled` | Enable recording of all traffic captured according to other parameters. Whatever Kubeshark captures, considering pod targeting rules, will be stored in pcap files ready to be viewed by tools | `false` |
| `pcapdump.maxTime` | The time window into the past that will be stored. Older traffic will be discarded. | `2h` |
| `pcapdump.maxSize` | The maximum storage size the PCAP files will consume. Old files that cause to surpass storage consumption will get discarded. | `500MB` |
| `kube.configPath` | Path to the `kubeconfig` file (`$HOME/.kube/config`) | `""` |
@@ -308,7 +308,7 @@ tap:
# Installing with Dex OIDC authentication
[**Click here to see full docs**](https://docs.kubehq.com/en/saml#installing-with-oidc-enabled-dex-idp).
[**Click here to see full docs**](https://docs.kubeshark.com/en/saml#installing-with-oidc-enabled-dex-idp).
Choose this option, if **you already have a running instance** of Dex in your cluster &
you want to set up Dex OIDC authentication for Kubeshark users.

View File

@@ -86,3 +86,9 @@ rules:
verbs:
- create
- get
- apiGroups:
- batch
resources:
- jobs
verbs:
- "*"

View File

@@ -3,7 +3,7 @@ apiVersion: apps/v1
kind: Deployment
metadata:
labels:
app.kubehq.com/app: hub
app.kubeshark.com/app: hub
{{- include "kubeshark.labels" . | nindent 4 }}
{{- if .Values.tap.annotations }}
annotations:
@@ -15,12 +15,12 @@ spec:
replicas: 1 # Set the desired number of replicas
selector:
matchLabels:
app.kubehq.com/app: hub
app.kubeshark.com/app: hub
{{- include "kubeshark.selectorLabels" . | nindent 6 }}
template:
metadata:
labels:
app.kubehq.com/app: hub
app.kubeshark.com/app: hub
{{- include "kubeshark.labels" . | nindent 8 }}
spec:
dnsPolicy: ClusterFirstWithHostNet
@@ -40,6 +40,18 @@ spec:
- "{{ if hasKey .Values.tap.capture "stopAfter" }}{{ .Values.tap.capture.stopAfter }}{{ else }}5m{{ end }}"
- -snapshot-size-limit
- '{{ .Values.tap.snapshots.storageSize }}'
{{- if .Values.tap.delayedDissection.image }}
- -dissector-image
- '{{ .Values.tap.delayedDissection.image }}'
{{- end }}
{{- if .Values.tap.delayedDissection.cpu }}
- -dissector-cpu
- '{{ .Values.tap.delayedDissection.cpu }}'
{{- end }}
{{- if .Values.tap.delayedDissection.memory }}
- -dissector-memory
- '{{ .Values.tap.delayedDissection.memory }}'
{{- end }}
{{- if .Values.tap.gitops.enabled }}
- -gitops
{{- end }}
@@ -64,7 +76,7 @@ spec:
- name: SENTRY_ENVIRONMENT
value: '{{ .Values.tap.sentry.environment }}'
- name: KUBESHARK_CLOUD_API_URL
value: 'https://api.kubehq.com'
value: 'https://api.kubeshark.com'
- name: PROFILING_ENABLED
value: '{{ .Values.tap.pprof.enabled }}'
{{- if .Values.tap.docker.overrideImage.hub }}

View File

@@ -3,7 +3,7 @@ apiVersion: v1
kind: Service
metadata:
labels:
app.kubehq.com/app: hub
app.kubeshark.com/app: hub
{{- include "kubeshark.labels" . | nindent 4 }}
{{- if .Values.tap.annotations }}
annotations:
@@ -17,5 +17,5 @@ spec:
port: 80
targetPort: 8080
selector:
app.kubehq.com/app: hub
app.kubeshark.com/app: hub
type: ClusterIP

View File

@@ -2,7 +2,7 @@ apiVersion: apps/v1
kind: Deployment
metadata:
labels:
app.kubehq.com/app: front
app.kubeshark.com/app: front
{{- include "kubeshark.labels" . | nindent 4 }}
{{- if .Values.tap.annotations }}
annotations:
@@ -14,12 +14,12 @@ spec:
replicas: 1 # Set the desired number of replicas
selector:
matchLabels:
app.kubehq.com/app: front
app.kubeshark.com/app: front
{{- include "kubeshark.selectorLabels" . | nindent 6 }}
template:
metadata:
labels:
app.kubehq.com/app: front
app.kubeshark.com/app: front
{{- include "kubeshark.labels" . | nindent 8 }}
spec:
containers:
@@ -86,6 +86,8 @@ spec:
value: '{{ default false .Values.betaEnabled | ternary "true" "false" }}'
- name: REACT_APP_DISSECTORS_UPDATING_ENABLED
value: '{{ .Values.tap.liveConfigMapChangesDisabled | ternary "false" "true" }}'
- name: REACT_APP_RAW_CAPTURE_ENABLED
value: '{{ .Values.tap.capture.raw.enabled | ternary "true" "false" }}'
- name: REACT_APP_SENTRY_ENABLED
value: '{{ (include "sentry.enabled" .) }}'
- name: REACT_APP_SENTRY_ENVIRONMENT

View File

@@ -16,5 +16,5 @@ spec:
port: 80
targetPort: 8080
selector:
app.kubehq.com/app: front
app.kubeshark.com/app: front
type: ClusterIP

View File

@@ -3,7 +3,7 @@ apiVersion: apps/v1
kind: DaemonSet
metadata:
labels:
app.kubehq.com/app: worker
app.kubeshark.com/app: worker
sidecar.istio.io/inject: "false"
{{- include "kubeshark.labels" . | nindent 4 }}
{{- if .Values.tap.annotations }}
@@ -15,12 +15,12 @@ metadata:
spec:
selector:
matchLabels:
app.kubehq.com/app: worker
app.kubeshark.com/app: worker
{{- include "kubeshark.selectorLabels" . | nindent 6 }}
template:
metadata:
labels:
app.kubehq.com/app: worker
app.kubeshark.com/app: worker
{{- include "kubeshark.labels" . | nindent 8 }}
name: kubeshark-worker-daemon-set
namespace: kubeshark
@@ -130,7 +130,7 @@ spec:
- name: TCP_STREAM_CHANNEL_TIMEOUT_SHOW
value: '{{ .Values.tap.misc.tcpStreamChannelTimeoutShow }}'
- name: KUBESHARK_CLOUD_API_URL
value: 'https://api.kubehq.com'
value: 'https://api.kubeshark.com'
- name: PROFILING_ENABLED
value: '{{ .Values.tap.pprof.enabled }}'
- name: SENTRY_ENABLED

View File

@@ -4,7 +4,7 @@ metadata:
name: {{ include "kubeshark.configmapName" . }}
namespace: {{ .Release.Namespace }}
labels:
app.kubehq.com/app: hub
app.kubeshark.com/app: hub
{{- include "kubeshark.labels" . | nindent 4 }}
data:
POD_REGEX: '{{ .Values.tap.regex }}'

View File

@@ -4,7 +4,7 @@ metadata:
name: {{ include "kubeshark.secretName" . }}
namespace: {{ .Release.Namespace }}
labels:
app.kubehq.com/app: hub
app.kubeshark.com/app: hub
{{- include "kubeshark.labels" . | nindent 4 }}
stringData:
LICENSE: '{{ .Values.license }}'
@@ -20,7 +20,7 @@ metadata:
name: kubeshark-saml-x509-crt-secret
namespace: {{ .Release.Namespace }}
labels:
app.kubehq.com/app: hub
app.kubeshark.com/app: hub
{{- include "kubeshark.labels" . | nindent 4 }}
stringData:
AUTH_SAML_X509_CRT: |
@@ -34,7 +34,7 @@ metadata:
name: kubeshark-saml-x509-key-secret
namespace: {{ .Release.Namespace }}
labels:
app.kubehq.com/app: hub
app.kubeshark.com/app: hub
{{- include "kubeshark.labels" . | nindent 4 }}
stringData:
AUTH_SAML_X509_KEY: |

View File

@@ -14,7 +14,7 @@ metadata:
namespace: {{ .Release.Namespace }}
spec:
selector:
app.kubehq.com/app: worker
app.kubeshark.com/app: worker
{{- include "kubeshark.labels" . | nindent 4 }}
ports:
- name: metrics

View File

@@ -14,7 +14,7 @@ metadata:
namespace: {{ .Release.Namespace }}
spec:
selector:
app.kubehq.com/app: hub
app.kubeshark.com/app: hub
{{- include "kubeshark.labels" . | nindent 4 }}
ports:
- name: metrics

View File

@@ -12,7 +12,7 @@ metadata:
spec:
podSelector:
matchLabels:
app.kubehq.com/app: hub
app.kubeshark.com/app: hub
policyTypes:
- Ingress
- Egress
@@ -40,7 +40,7 @@ metadata:
spec:
podSelector:
matchLabels:
app.kubehq.com/app: front
app.kubeshark.com/app: front
policyTypes:
- Ingress
- Egress
@@ -65,7 +65,7 @@ metadata:
spec:
podSelector:
matchLabels:
app.kubehq.com/app: dex
app.kubeshark.com/app: dex
policyTypes:
- Ingress
- Egress
@@ -90,7 +90,7 @@ metadata:
spec:
podSelector:
matchLabels:
app.kubehq.com/app: worker
app.kubeshark.com/app: worker
policyTypes:
- Ingress
- Egress

View File

@@ -5,7 +5,7 @@ apiVersion: apps/v1
kind: Deployment
metadata:
labels:
app.kubehq.com/app: dex
app.kubeshark.com/app: dex
{{- include "kubeshark.labels" . | nindent 4 }}
{{- if .Values.tap.annotations }}
annotations:
@@ -17,12 +17,12 @@ spec:
replicas: 1 # Set the desired number of replicas
selector:
matchLabels:
app.kubehq.com/app: dex
app.kubeshark.com/app: dex
{{- include "kubeshark.selectorLabels" . | nindent 6 }}
template:
metadata:
labels:
app.kubehq.com/app: dex
app.kubeshark.com/app: dex
{{- include "kubeshark.labels" . | nindent 8 }}
spec:
containers:

View File

@@ -5,7 +5,7 @@ apiVersion: v1
kind: Service
metadata:
labels:
app.kubehq.com/app: dex
app.kubeshark.com/app: dex
{{- include "kubeshark.labels" . | nindent 4 }}
{{- if .Values.tap.annotations }}
annotations:
@@ -19,7 +19,7 @@ spec:
port: 80
targetPort: 5556
selector:
app.kubehq.com/app: dex
app.kubeshark.com/app: dex
type: ClusterIP
{{- end }}

View File

@@ -6,7 +6,7 @@ metadata:
name: kubeshark-dex-conf-secret
namespace: {{ .Release.Namespace }}
labels:
app.kubehq.com/app: hub
app.kubeshark.com/app: hub
{{- include "kubeshark.labels" . | nindent 4 }}
data:
dex-config.yaml: {{ .Values.tap.auth.dexConfig | toYaml | b64enc | quote }}

View File

@@ -28,9 +28,12 @@ Notices:
- Support chat using Intercom is enabled. It can be disabled using `--set supportChatEnabled=false`
{{- end }}
{{- if eq .Values.license ""}}
- No license key was detected. You can either log-in/sign-up through the dashboard, or download the license key from https://console.kubehq.com/ and add it as 'license: <license>' in helm values or as `--set license=<license>` or as `LICENSE` via mounted secret (`tap.secrets`).
- No license key was detected.
- Authenticate through the dashboard to activate a complementary COMMUNITY license.
- If you have an Enterprise license, download the license key from https://console.kubeshark.com/
- An Enterprise license-key can be added as 'license: <license>' in helm values or as `--set license=<license>` or as `LICENSE` via mounted secret (`tap.secrets`).
- Contact us to get an Enterprise license: https://kubeshark.com/contact-us.
{{- end }}
{{ if .Values.tap.ingress.enabled }}
You can now access the application through the following URL:
@@ -42,8 +45,9 @@ To access the application, follow these steps:
1. Perform port forwarding with the following commands:
kubectl port-forward -n {{ .Release.Namespace }} service/kubeshark-front 8899:80
you could also run: `kubeshark proxy` (which simply manages the port-forward connection)
2. Once port forwarding is done, you can access the application by visiting the following URL in your web browser:
http://0.0.0.0:8899{{ default "" (((.Values.tap).routing).front).basePath }}/
http://127.0.0.1:8899{{ default "" (((.Values.tap).routing).front).basePath }}/
{{- end }}

View File

@@ -1,4 +1,3 @@
# find a detailed description here: https://github.com/kubeshark/kubeshark/blob/master/helm-chart/README.md
tap:
docker:
registry: docker.io/kubeshark
@@ -30,14 +29,18 @@ tap:
stopped: false
stopAfter: 5m
raw:
enabled: false
enabled: true
storageSize: 1Gi
dbMaxSize: 500Mi
delayedDissection:
image: kubeshark/worker:master
cpu: "1"
memory: 4Gi
snapshots:
storageClass: ""
storageSize: 1Gi
storageSize: 20Gi
release:
repo: https://helm.kubehq.com
repo: https://helm.kubeshark.com
name: kubeshark
namespace: default
persistentStorage: false
@@ -45,7 +48,7 @@ tap:
persistentStoragePvcVolumeMode: FileSystem
efsFileSytemIdAndPath: ""
secrets: []
storageLimit: 5Gi
storageLimit: 10Gi
storageClass: standard
dryRun: false
dns:
@@ -156,7 +159,7 @@ tap:
ipv6: true
debug: false
dashboard:
streamingType: ""
streamingType: connect-rpc
completeStreamingEnabled: true
telemetry:
enabled: true
@@ -183,6 +186,8 @@ tap:
- ldap
- radius
- diameter
- udp-flow
- tcp-flow
portMapping:
http:
- 80
@@ -209,8 +214,8 @@ tap:
view: flamegraph
misc:
jsonTTL: 5m
pcapTTL: 10s
pcapErrorTTL: 60s
pcapTTL: "0"
pcapErrorTTL: "0"
trafficSampleRate: 100
tcpStreamChannelTimeoutMs: 10000
tcpStreamChannelTimeoutShow: false
@@ -247,7 +252,7 @@ logs:
file: ""
grep: ""
pcapdump:
enabled: true
enabled: false
timeInterval: 1m
maxTime: 1h
maxSize: 500MB

View File

@@ -8,5 +8,5 @@ const (
HubServiceName = HubPodName
K8sAllNamespaces = ""
MinKubernetesServerVersion = "1.16.0"
AppLabelKey = "app.kubehq.com/app"
AppLabelKey = "app.kubeshark.com/app"
)

View File

@@ -4,17 +4,17 @@ apiVersion: networking.k8s.io/v1
kind: NetworkPolicy
metadata:
labels:
helm.sh/chart: kubeshark-52.10.0
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
name: kubeshark-hub-network-policy
namespace: default
spec:
podSelector:
matchLabels:
app.kubehq.com/app: hub
app.kubeshark.com/app: hub
policyTypes:
- Ingress
- Egress
@@ -33,10 +33,10 @@ apiVersion: networking.k8s.io/v1
kind: NetworkPolicy
metadata:
labels:
helm.sh/chart: kubeshark-52.10.0
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
annotations:
name: kubeshark-front-network-policy
@@ -44,7 +44,7 @@ metadata:
spec:
podSelector:
matchLabels:
app.kubehq.com/app: front
app.kubeshark.com/app: front
policyTypes:
- Ingress
- Egress
@@ -60,10 +60,10 @@ apiVersion: networking.k8s.io/v1
kind: NetworkPolicy
metadata:
labels:
helm.sh/chart: kubeshark-52.10.0
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
annotations:
name: kubeshark-dex-network-policy
@@ -71,7 +71,7 @@ metadata:
spec:
podSelector:
matchLabels:
app.kubehq.com/app: dex
app.kubeshark.com/app: dex
policyTypes:
- Ingress
- Egress
@@ -87,10 +87,10 @@ apiVersion: networking.k8s.io/v1
kind: NetworkPolicy
metadata:
labels:
helm.sh/chart: kubeshark-52.10.0
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
annotations:
name: kubeshark-worker-network-policy
@@ -98,7 +98,7 @@ metadata:
spec:
podSelector:
matchLabels:
app.kubehq.com/app: worker
app.kubeshark.com/app: worker
policyTypes:
- Ingress
- Egress
@@ -116,10 +116,10 @@ apiVersion: v1
kind: ServiceAccount
metadata:
labels:
helm.sh/chart: kubeshark-52.10.0
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
name: kubeshark-service-account
namespace: default
@@ -131,11 +131,11 @@ metadata:
name: kubeshark-secret
namespace: default
labels:
app.kubehq.com/app: hub
helm.sh/chart: kubeshark-52.10.0
app.kubeshark.com/app: hub
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
stringData:
LICENSE: ''
@@ -150,11 +150,11 @@ metadata:
name: kubeshark-saml-x509-crt-secret
namespace: default
labels:
app.kubehq.com/app: hub
helm.sh/chart: kubeshark-52.10.0
app.kubeshark.com/app: hub
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
stringData:
AUTH_SAML_X509_CRT: |
@@ -166,11 +166,11 @@ metadata:
name: kubeshark-saml-x509-key-secret
namespace: default
labels:
app.kubehq.com/app: hub
helm.sh/chart: kubeshark-52.10.0
app.kubeshark.com/app: hub
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
stringData:
AUTH_SAML_X509_KEY: |
@@ -182,10 +182,10 @@ metadata:
name: kubeshark-nginx-config-map
namespace: default
labels:
helm.sh/chart: kubeshark-52.10.0
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
data:
default.conf: |
@@ -245,11 +245,11 @@ metadata:
name: kubeshark-config-map
namespace: default
labels:
app.kubehq.com/app: hub
helm.sh/chart: kubeshark-52.10.0
app.kubeshark.com/app: hub
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
data:
POD_REGEX: '.*'
@@ -281,22 +281,22 @@ data:
DEFAULT_FILTER: ""
TRAFFIC_SAMPLE_RATE: '100'
JSON_TTL: '5m'
PCAP_TTL: '10s'
PCAP_ERROR_TTL: '60s'
PCAP_TTL: '0'
PCAP_ERROR_TTL: '0'
TIMEZONE: ' '
CLOUD_LICENSE_ENABLED: 'true'
AI_ASSISTANT_ENABLED: 'true'
DUPLICATE_TIMEFRAME: '200ms'
ENABLED_DISSECTORS: 'amqp,dns,http,icmp,kafka,redis,ws,ldap,radius,diameter'
ENABLED_DISSECTORS: 'amqp,dns,http,icmp,kafka,redis,ws,ldap,radius,diameter,udp-flow,tcp-flow'
CUSTOM_MACROS: '{"https":"tls and (http or http2)"}'
DISSECTORS_UPDATING_ENABLED: 'true'
DETECT_DUPLICATES: 'false'
PCAP_DUMP_ENABLE: 'true'
PCAP_DUMP_ENABLE: 'false'
PCAP_TIME_INTERVAL: '1m'
PCAP_MAX_TIME: '1h'
PCAP_MAX_SIZE: '500MB'
PORT_MAPPING: '{"amqp":[5671,5672],"diameter":[3868],"http":[80,443,8080],"kafka":[9092],"ldap":[389],"redis":[6379]}'
RAW_CAPTURE: 'false'
RAW_CAPTURE: 'true'
RAW_CAPTURE_STORAGE_SIZE: '1Gi'
---
# Source: kubeshark/templates/02-cluster-role.yaml
@@ -304,10 +304,10 @@ apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
labels:
helm.sh/chart: kubeshark-52.10.0
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
name: kubeshark-cluster-role-default
namespace: default
@@ -351,10 +351,10 @@ apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
labels:
helm.sh/chart: kubeshark-52.10.0
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
name: kubeshark-cluster-role-binding-default
namespace: default
@@ -372,10 +372,10 @@ apiVersion: rbac.authorization.k8s.io/v1
kind: Role
metadata:
labels:
helm.sh/chart: kubeshark-52.10.0
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
annotations:
name: kubeshark-self-config-role
@@ -410,16 +410,22 @@ rules:
verbs:
- create
- get
- apiGroups:
- batch
resources:
- jobs
verbs:
- "*"
---
# Source: kubeshark/templates/03-cluster-role-binding.yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: RoleBinding
metadata:
labels:
helm.sh/chart: kubeshark-52.10.0
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
annotations:
name: kubeshark-self-config-role-binding
@@ -438,11 +444,11 @@ apiVersion: v1
kind: Service
metadata:
labels:
app.kubehq.com/app: hub
helm.sh/chart: kubeshark-52.10.0
app.kubeshark.com/app: hub
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
name: kubeshark-hub
namespace: default
@@ -452,7 +458,7 @@ spec:
port: 80
targetPort: 8080
selector:
app.kubehq.com/app: hub
app.kubeshark.com/app: hub
type: ClusterIP
---
# Source: kubeshark/templates/07-front-service.yaml
@@ -460,10 +466,10 @@ apiVersion: v1
kind: Service
metadata:
labels:
helm.sh/chart: kubeshark-52.10.0
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
name: kubeshark-front
namespace: default
@@ -473,7 +479,7 @@ spec:
port: 80
targetPort: 8080
selector:
app.kubehq.com/app: front
app.kubeshark.com/app: front
type: ClusterIP
---
# Source: kubeshark/templates/15-worker-service-metrics.yaml
@@ -481,10 +487,10 @@ kind: Service
apiVersion: v1
metadata:
labels:
helm.sh/chart: kubeshark-52.10.0
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
annotations:
prometheus.io/scrape: 'true'
@@ -493,11 +499,11 @@ metadata:
namespace: default
spec:
selector:
app.kubehq.com/app: worker
helm.sh/chart: kubeshark-52.10.0
app.kubeshark.com/app: worker
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
ports:
- name: metrics
@@ -510,10 +516,10 @@ kind: Service
apiVersion: v1
metadata:
labels:
helm.sh/chart: kubeshark-52.10.0
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
annotations:
prometheus.io/scrape: 'true'
@@ -522,11 +528,11 @@ metadata:
namespace: default
spec:
selector:
app.kubehq.com/app: hub
helm.sh/chart: kubeshark-52.10.0
app.kubeshark.com/app: hub
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
ports:
- name: metrics
@@ -539,29 +545,29 @@ apiVersion: apps/v1
kind: DaemonSet
metadata:
labels:
app.kubehq.com/app: worker
app.kubeshark.com/app: worker
sidecar.istio.io/inject: "false"
helm.sh/chart: kubeshark-52.10.0
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
name: kubeshark-worker-daemon-set
namespace: default
spec:
selector:
matchLabels:
app.kubehq.com/app: worker
app.kubeshark.com/app: worker
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
template:
metadata:
labels:
app.kubehq.com/app: worker
helm.sh/chart: kubeshark-52.10.0
app.kubeshark.com/app: worker
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
name: kubeshark-worker-daemon-set
namespace: kubeshark
@@ -571,7 +577,7 @@ spec:
- /bin/sh
- -c
- mkdir -p /sys/fs/bpf && mount | grep -q '/sys/fs/bpf' || mount -t bpf bpf /sys/fs/bpf
image: 'docker.io/kubeshark/worker:v52.10'
image: 'docker.io/kubeshark/worker:v52.12'
imagePullPolicy: Always
name: mount-bpf
securityContext:
@@ -601,10 +607,10 @@ spec:
- -staletimeout
- '30'
- -storage-size
- '5Gi'
- '10Gi'
- -capture-db-max-size
- '500Mi'
image: 'docker.io/kubeshark/worker:v52.10'
image: 'docker.io/kubeshark/worker:v52.12'
imagePullPolicy: Always
name: sniffer
ports:
@@ -625,7 +631,7 @@ spec:
- name: TCP_STREAM_CHANNEL_TIMEOUT_SHOW
value: 'false'
- name: KUBESHARK_CLOUD_API_URL
value: 'https://api.kubehq.com'
value: 'https://api.kubeshark.com'
- name: PROFILING_ENABLED
value: 'false'
- name: SENTRY_ENABLED
@@ -678,7 +684,7 @@ spec:
- -disable-tls-log
- -loglevel
- 'warning'
image: 'docker.io/kubeshark/worker:v52.10'
image: 'docker.io/kubeshark/worker:v52.12'
imagePullPolicy: Always
name: tracer
env:
@@ -762,18 +768,18 @@ spec:
name: root
- name: data
emptyDir:
sizeLimit: 5Gi
sizeLimit: 10Gi
---
# Source: kubeshark/templates/04-hub-deployment.yaml
apiVersion: apps/v1
kind: Deployment
metadata:
labels:
app.kubehq.com/app: hub
helm.sh/chart: kubeshark-52.10.0
app.kubeshark.com/app: hub
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
name: kubeshark-hub
namespace: default
@@ -781,17 +787,17 @@ spec:
replicas: 1 # Set the desired number of replicas
selector:
matchLabels:
app.kubehq.com/app: hub
app.kubeshark.com/app: hub
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
template:
metadata:
labels:
app.kubehq.com/app: hub
helm.sh/chart: kubeshark-52.10.0
app.kubeshark.com/app: hub
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
spec:
dnsPolicy: ClusterFirstWithHostNet
@@ -807,7 +813,13 @@ spec:
- -capture-stop-after
- "5m"
- -snapshot-size-limit
- '1Gi'
- '20Gi'
- -dissector-image
- 'kubeshark/worker:master'
- -dissector-cpu
- '1'
- -dissector-memory
- '4Gi'
env:
- name: POD_NAME
valueFrom:
@@ -822,10 +834,10 @@ spec:
- name: SENTRY_ENVIRONMENT
value: 'production'
- name: KUBESHARK_CLOUD_API_URL
value: 'https://api.kubehq.com'
value: 'https://api.kubeshark.com'
- name: PROFILING_ENABLED
value: 'false'
image: 'docker.io/kubeshark/hub:v52.10'
image: 'docker.io/kubeshark/hub:v52.12'
imagePullPolicy: Always
readinessProbe:
periodSeconds: 5
@@ -885,18 +897,18 @@ spec:
path: kubeshark.key
- name: snapshots-volume
emptyDir:
sizeLimit: 1Gi
sizeLimit: 20Gi
---
# Source: kubeshark/templates/06-front-deployment.yaml
apiVersion: apps/v1
kind: Deployment
metadata:
labels:
app.kubehq.com/app: front
helm.sh/chart: kubeshark-52.10.0
app.kubeshark.com/app: front
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
name: kubeshark-front
namespace: default
@@ -904,17 +916,17 @@ spec:
replicas: 1 # Set the desired number of replicas
selector:
matchLabels:
app.kubehq.com/app: front
app.kubeshark.com/app: front
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
template:
metadata:
labels:
app.kubehq.com/app: front
helm.sh/chart: kubeshark-52.10.0
app.kubeshark.com/app: front
helm.sh/chart: kubeshark-52.12.0
app.kubernetes.io/name: kubeshark
app.kubernetes.io/instance: kubeshark
app.kubernetes.io/version: "52.10.0"
app.kubernetes.io/version: "52.12.0"
app.kubernetes.io/managed-by: Helm
spec:
containers:
@@ -926,7 +938,7 @@ spec:
- name: REACT_APP_COMPLETE_STREAMING_ENABLED
value: 'true'
- name: REACT_APP_STREAMING_TYPE
value: ''
value: 'connect-rpc'
- name: REACT_APP_AUTH_SAML_IDP_METADATA_URL
value: ' '
- name: REACT_APP_TIMEZONE
@@ -953,11 +965,13 @@ spec:
value: 'false'
- name: REACT_APP_DISSECTORS_UPDATING_ENABLED
value: 'true'
- name: REACT_APP_RAW_CAPTURE_ENABLED
value: 'true'
- name: REACT_APP_SENTRY_ENABLED
value: 'false'
- name: REACT_APP_SENTRY_ENVIRONMENT
value: 'production'
image: 'docker.io/kubeshark/front:v52.10'
image: 'docker.io/kubeshark/front:v52.12'
imagePullPolicy: Always
name: kubeshark-front
livenessProbe:

View File

@@ -5,7 +5,7 @@ metadata:
spec:
acme:
server: https://acme-v02.api.letsencrypt.org/directory
email: info@kubehq.com
email: info@kubeshark.com
privateKeySecretRef:
name: letsencrypt-prod-key
solvers:

View File

@@ -10,8 +10,8 @@ var (
Software = "Kubeshark"
Program = "kubeshark"
Description = "The API Traffic Analyzer for Kubernetes"
Website = "https://kubehq.com"
Email = "support@kubehq.io"
Website = "https://kubeshark.com"
Email = "support@kubeshark.com"
Ver = "0.0.0"
Branch = "master"
GitCommitHash = "" // this var is overridden using ldflags in makefile when building