)]}'
{
  "id": "92669bfd8a1607fbe450ee9f904f949dac9075f4",
  "repo": "external/github.com/python/cpython",
  "revision": "80428ed4e19b31071433806b4d89465c88e084c6",
  "path": "Parser/tokenizer.h"
}
