Skip to content

Latest commit

 

History

History
94 lines (78 loc) · 3.86 KB

README.md

File metadata and controls

94 lines (78 loc) · 3.86 KB

MeshInsight

Overview

MeshInsight is a tool to systematically characterize the overhead of service meshes and to help developers quantify the latency and CPU overhead in deployment scenarios of interest. Read the paper for how MeshInsight works!

Note: MeshInsight currently only works on Istio. We plan to extend it to other service meshes (e.g., Cilium or Linkerd) in the future.

Installation

Requirements

  • Python (Version >= 3.7)
  • Kubernetes (Idealy version >=1.24)
    • Run . ./utils/k8s_setup.sh to install Kubernetes. (Note - this script is only tested on Ubuntu.)

Run the following script to install necessary tools for MeshInsight:

. ./install.sh

For details see INSTALL.md.

Run MeshInsight

workflow
MeshInsight Workflow

MeshInsight has an offline profiling phase and an online prediction phase. The offline phase generates performance profiles of individual service mesh components, and the online phase predicts overhead based on these profiles, service mesh configuration, and application workload. Optionally, MeshInsight can predict the performance of an optimization based on a speedup profile.

Step 1: Run offline profiler

# Run profiler for latency and cpu overhead profile. Root access is needed for BCC and perf
# Once it completes, the profile will be saved to meshinsight/profiles/
sudo -E python3 ./meshinsight/profiler/offline_profiler.py --latency --cpu

usage: offline_profiler.py [-h] [-v] [-c] [-l] [-d DURATION] 

optional arguments:
  -h, --help            show this help message and exit
  -v, --verbose
  -c, --cpu             run offline latency profiling
  -l, --latency         run offline cpu profiling
  -d DURATION, --duration DURATION
                        the duration for each measurement
                        default duration is 15s                    

Note:

  1. The profiler will delete all kubernetes deployments before it starts to minimize interference.
  2. We recommend running the offline profiler on a machine that has at least 16 cores.
  3. The offline profiling phase will take a few hours (~3 hours with duration of 15s)

Step 2: Run online predictor

# Run prediction based on an example call graph
python3 ./meshinsight/predictor/online_predictor.py -c ./meshinsight/predictor/config/example1.yml -p meshinsight/profiles/profile.pkl 

usage: online_predictor.py [-h] [-v] [-p PROFILE] -c CONFIG

optional arguments:
  -h, --help            show this help message and exit
  -v, --verbose
  -p PROFILE, --profile PROFILE
                        path to the profile
  -c CONFIG, --config CONFIG
                        path to config file

Please see PREDICTION.md for reference of configuration files.

Repo Structure

Repo Root
|---- benchmark   
|---- meshinsight   # MeshInsight source code
  |---- profiler    # Source code for offline profiler
    |---- benchmark   # Benchmarks used in the paper and their wrk scripts
  |---- predictor   # Source code for online predictor
|---- utils         # Some other tools for analyzing service mesh deployments
|---- samples       # Demo of service mesh performance analysis

Reference

Please consider citing our paper if you find MeshInsight related to your research.

@inproceedings{meshinsight,
  title={Dissecting Overheads of Service Mesh Sidecars},
  author={Zhu, Xiangfeng and She, Guozhen and Xue, Bowen and Zhang, Yu and Zhang, Yongsu and Zou, Xuan Kelvin and Duan, XiongChun and He, Peng and Krishnamurthy, Arvind and Lentz, Matthew and others},
  booktitle={Proceedings of the 2023 ACM Symposium on Cloud Computing},
  pages={142--157},
  year={2023}
}

Contact

If you have any questions or comments, please get in touch with Xiangfeng Zhu ([email protected]).