)]}'
{
  "id": "a40f7d9687b44078065b3c1d751010bd3592b086",
  "repo": "external/github.com/python/cpython",
  "revision": "d003a5bd2505a7fa04f50504b68ba8fca67349cd",
  "path": "Parser/tokenizer.h"
}
