JSPM

  • Created
  • Published
  • Downloads 7219
  • Score
    100M100P100Q114742F
  • License MIT

Information on LLM models, context window token limit, output token limit, pricing and more

Package Exports

  • llm-info

Readme

llm-info

NPM version CI

Information on LLM models, context window token limit, output token limit, pricing and more.

Information provided includes:

  • context window token limit
  • output token limit
  • pricing
  • and more

Models included:

  • GPT-4
  • GPT-4o
  • GPT-4o mini
  • Claude 3.5 Sonnet

Install:

$ yarn add llm-info

Usage

import { AllModels, ModelEnum, ModelLikeEnum, ModelInfoMap } from 'llm-info';

console.log(AllModels);
// [ 'gpt-4', 'gpt-4o', 'gpt-4o-mini', 'claude-3-5-sonnet-20240620' ]

const model = ModelEnum['gpt-4'];
const modelInfo = ModelInfoMap[model];
console.log(modelInfo);
/*
{
  name: 'GPT-4',
  provider: 'openai',
  contextWindowTokenLimit: 128000,
  outputTokenLimit: 4096,
  pricePerMillionInputTokens: 30,
  pricePerMillionOutputTokens: 60
}
*/

console.log(ModelInfoMap[ModelEnum['claude-3-5-sonnet-20240620']]);
/*
{
  name: 'Claude 3.5 Sonnet',
  provider: 'anthropic',
  contextWindowTokenLimit: 200000,
  outputTokenLimit: 4096,
  pricePerMillionInputTokens: 3,
  pricePerMillionOutputTokens: 15
}
*/

const modelLike = ModelLikeEnum['chatgpt'];
const modelLikeInfo = ModelInfoMap[modelLike];
console.log(modelLikeInfo);
/*
{
  name: 'ChatGPT',
  provider: 'openai',
  contextWindowTokenLimit: 4096,
  outputTokenLimit: 4096,
  pricePerMillionInputTokens: null,
  pricePerMillionOutputTokens: null
}
*/

Testing

$ yarn test