)]}'
{
  "id": "5e2171885ac75b4ee37c164828b07d4d756da7de",
  "repo": "external/github.com/python/cpython",
  "revision": "798bcaa1eb01de7db9ff1881a3088603ad09b096",
  "path": "Parser/tokenizer.h"
}
