)]}'
{
  "id": "2542d30e1da0ed1624ce6edcd7311eff8c8fe424",
  "repo": "external/github.com/python/cpython",
  "revision": "003f341e99234cf6088341e746ffef15e12ccda2",
  "path": "Parser/tokenizer.h"
}
