)]}'
{
  "id": "16a94d5f51d664b58d8fd9749295e06a3f5070db",
  "repo": "external/github.com/python/cpython",
  "revision": "36854bbb240e417c0df6f0014924fcc899388186",
  "path": "Parser/tokenizer.h"
}
