1.1.0 • Published 1 year ago
gpt-token-counter v1.1.0
gpt-token-counter
A command-line tool to count tokens in files and folders using GPT-4 tokenization.
Features
- Count tokens in individual files or entire directories
- Supports multiple file types: .md, .txt, .json, .csv
- Excludes common non-text files and directories
- Uses GPT-4 tokenization method
- No API key or internet connection required
Installation
Install the package globally:
npm install -g gpt-token-counterUsage
You can use the command tokens to count tokens in a file or directory.
Count Tokens in a File
tokens <path-to-file>Example:
tokens /path/to/your/file.txtCount Tokens in a Directory
tokens <path-to-directory>Example:
tokens /path/to/your/directoryAdding Alias to zshrc
To make it easier to use, you can add an alias to your ~/.zshrc file:
echo "alias 'tokens file'='tokens'" >> ~/.zshrc
source ~/.zshrcNow you can use the command tokens file <path> to count tokens in any file.
Configuration
The tool uses a config.json file to specify which file types to include and exclude. Here is an example configuration:
{
"include": ["**/*.md", "**/*.txt", "**/*.json", "**/*.csv"],
"exclude": ["node_modules/**", "**/*.log", "**/*.pdf", "**/*.docx", "**/*.xlsx", "**/*.xls", "**/*.pptx", "**/*.ppt", "**/*.odt", "**/*.ods", "**/*.odp"]
}Example Output
tokens /path/to/your/directory
Processing folder: /path/to/your/directory
Include pattern: **/*.{md,txt,json,csv}
Exclude patterns: node_modules, log, pdf, docx, xlsx, xls, pptx, ppt, odt, ods, odp
Found files: 3
Processing file: example-file-1.json
Token count for example-file-1.json: 37857
Processing file: example-file-2.json
Token count for example-file-2.json: 379634
Processing file: example-file-3.json
Token count for example-file-3.json: 10039
Total token count: 427530Author
Repository
License
This project is licensed under the ISC License.
1.1.0
1 year ago