)]}'
{
  "id": "ce72e1529024c138f499b9319283cd8ebaadd25c",
  "repo": "external/github.com/python/cpython",
  "revision": "003f341e99234cf6088341e746ffef15e12ccda2",
  "path": "Parser/tokenizer.c"
}
