)]}'
{
  "id": "463c0e00ca1411af5624ab82881011a42bbc992c",
  "repo": "external/github.com/python/cpython",
  "revision": "005e69403d638f9ff8f71e59960c600016e101a4",
  "path": "Parser/tokenizer.c"
}
