Inferring Channel Buffer Bounds Via Linear Programming

* Final gross prices may vary according to local VAT.

Get Access

Abstract

We present a static analysis for inferring the maximum amount of buffer space used by a program consisting of concurrently running processes communicating via buffered channels. We reduce the problem to linear programming by casting the analysis as a fractional capability calculus system. Our analysis can reason about buffers used by multiple processes concurrently, and runs in time polynomial in the size of the program.