)]}'
{
  "id": "11d69fc5b2e15cc82b570dfb2cfcc82882df513b",
  "repo": "external/github.com/python/cpython",
  "revision": "e9791ba35175171170ff09094ea46b91fc18c654",
  "path": "Parser/tokenizer.h"
}
