Package Exports
- @faiss-node/native
- @faiss-node/native/src/js/index.js
This package does not declare an exports field, so the exports above have been automatically detected and optimized by JSPM instead. If any package subpath is missing, it is recommended to post an issue to the original package (@faiss-node/native) to support the "exports" field. If that is not possible, create a JSPM override to customize the exports field for this package.
Readme
@faiss-node/native
High-performance Node.js native bindings for Facebook FAISS - the industry-standard vector similarity search library. Built for semantic search, RAG applications, and vector databases.
Features
- 🚀 Async Operations - Non-blocking Promise-based API that never blocks the event loop
- 🔒 Thread-Safe - Mutex-protected concurrent operations for production workloads
- 📦 Multiple Index Types - FLAT_L2, IVF_FLAT, and HNSW with optimized defaults
- 💾 Persistence - Save/load indexes to disk or serialize to buffers
- ⚡ High Performance - Direct C++ bindings with zero-copy data transfer
- 📚 TypeScript Support - Full type definitions included
Installation
Quick Install
npm install @faiss-node/nativePrerequisites
macOS:
brew install cmake libomp openblas faissLinux (Ubuntu/Debian):
sudo apt-get update
sudo apt-get install -y cmake libopenblas-dev libomp-dev
# Build FAISS from source (see below)Building FAISS from Source (Linux):
git clone https://github.com/facebookresearch/faiss.git
cd faiss
cmake -B build -DFAISS_ENABLE_GPU=OFF -DFAISS_ENABLE_PYTHON=OFF
cmake --build build -j$(nproc)
sudo cmake --install buildBuild Native Module
After installing prerequisites:
npm run buildQuick Start
const { FaissIndex } = require('@faiss-node/native');
// Create an index
const index = new FaissIndex({ type: 'FLAT_L2', dims: 128 });
// Add vectors (single or batch)
const vectors = new Float32Array([
1.0, 0.0, 0.0, 0.0, // Vector 1
0.0, 1.0, 0.0, 0.0, // Vector 2
0.0, 0.0, 1.0, 0.0 // Vector 3
]);
await index.add(vectors);
// Search for nearest neighbors
const query = new Float32Array([1.0, 0.0, 0.0, 0.0]);
const results = await index.search(query, 2);
console.log('Labels:', results.labels); // Int32Array: [0, 1]
console.log('Distances:', results.distances); // Float32Array: [0, 2]
// Cleanup
index.dispose();API Reference
Constructor
Create a new FAISS index with the specified configuration.
const index = new FaissIndex(config);Parameters:
config.type(string, required): Index type -'FLAT_L2','IVF_FLAT', or'HNSW'config.dims(number, required): Vector dimensions (must be positive integer)config.nlist(number, optional): Number of clusters for IVF_FLAT (default: 100)config.nprobe(number, optional): Clusters to search for IVF_FLAT (default: 10)config.M(number, optional): Connections per node for HNSW (default: 16)config.efConstruction(number, optional): HNSW construction parameter (default: 200)config.efSearch(number, optional): HNSW search parameter (default: 50)
Examples:
// FLAT_L2 - Exact search (best for small datasets < 10k vectors)
const flatIndex = new FaissIndex({ type: 'FLAT_L2', dims: 128 });
// IVF_FLAT - Fast approximate search (best for 10k - 1M vectors)
const ivfIndex = new FaissIndex({
type: 'IVF_FLAT',
dims: 768,
nlist: 100, // Number of clusters
nprobe: 10 // Clusters to search (higher = more accurate, slower)
});
await ivfIndex.train(trainingVectors); // Must train before adding vectors!
// HNSW - State-of-the-art approximate search (best for large datasets)
const hnswIndex = new FaissIndex({
type: 'HNSW',
dims: 1536,
M: 16, // Connections per node (higher = more accurate, slower)
efConstruction: 200, // Construction parameter
efSearch: 50 // Search parameter (higher = more accurate, slower)
});Methods
add(vectors: Float32Array): Promise<void>
Add vectors to the index. Can add a single vector or a batch of vectors.
// Single vector
await index.add(new Float32Array([1, 2, 3, 4]));
// Batch of vectors (4 vectors of 4 dimensions each)
await index.add(new Float32Array([
1, 0, 0, 0, // Vector 1
0, 1, 0, 0, // Vector 2
0, 0, 1, 0, // Vector 3
0, 0, 0, 1 // Vector 4
]));Note: For IVF_FLAT indexes, you must call train() before adding vectors.
search(query: Float32Array, k: number): Promise<SearchResults>
Search for k nearest neighbors.
const query = new Float32Array([1, 0, 0, 0]);
const results = await index.search(query, 5);
// results.distances: Float32Array of L2 distances
// results.labels: Int32Array of vector indicesReturns:
distances(Float32Array): L2 distances to nearest neighborslabels(Int32Array): Indices of nearest neighbors
searchBatch(queries: Float32Array, k: number): Promise<SearchResults>
Perform batch search for multiple queries efficiently.
// 3 queries of 4 dimensions each
const queries = new Float32Array([
1, 0, 0, 0, // Query 1
0, 1, 0, 0, // Query 2
0, 0, 1, 0 // Query 3
]);
const results = await index.searchBatch(queries, 5);
// results.distances: Float32Array of shape [3 * 5]
// results.labels: Int32Array of shape [3 * 5]train(vectors: Float32Array): Promise<void>
Train an IVF_FLAT index. Required before adding vectors.
// Training vectors (typically 10k-100k vectors)
const trainingVectors = new Float32Array(/* ... */);
await ivfIndex.train(trainingVectors);
await ivfIndex.add(dataVectors); // Now you can add vectorssetNprobe(nprobe: number): Promise<void>
Set the number of clusters to search for IVF_FLAT indexes.
await ivfIndex.setNprobe(20); // Search more clusters (more accurate, slower)getStats(): IndexStats
Get index statistics.
const stats = index.getStats();
// {
// ntotal: number, // Total vectors in index
// dims: number, // Vector dimensions
// isTrained: boolean, // Whether index is trained (IVF only)
// type: string // Index type
// }save(filename: string): Promise<void>
Save index to disk.
await index.save('./my-index.faiss');static load(filename: string): Promise<FaissIndex>
Load index from disk.
const index = await FaissIndex.load('./my-index.faiss');toBuffer(): Promise<Buffer>
Serialize index to a Node.js Buffer (useful for databases, network transfer, etc.).
const buffer = await index.toBuffer();
// Store in database, send over network, etc.static fromBuffer(buffer: Buffer): Promise<FaissIndex>
Deserialize index from Buffer.
const index = await FaissIndex.fromBuffer(buffer);mergeFrom(otherIndex: FaissIndex): Promise<void>
Merge vectors from another index into this index.
const index1 = new FaissIndex({ type: 'FLAT_L2', dims: 128 });
const index2 = new FaissIndex({ type: 'FLAT_L2', dims: 128 });
await index1.add(vectors1);
await index2.add(vectors2);
await index1.mergeFrom(index2); // index1 now contains vectors from both
// Note: index2 is now empty (FAISS behavior)dispose(): void
Explicitly dispose of the index and free resources. Optional - automatic on garbage collection.
index.dispose();
// Index is now unusable - all operations will throw errorsChoosing the Right Index Type
FLAT_L2 (IndexFlatL2)
- Best for: Small datasets (< 10k vectors), exact search required
- Speed: O(n) per query - linear scan
- Accuracy: 100% recall (exact results)
- Memory: 4 × dims × n bytes
- Use case: Prototyping, small production datasets, when accuracy is critical
IVF_FLAT (IndexIVFFlat)
- Best for: Medium datasets (10k - 1M vectors), can tolerate ~90-95% recall
- Speed: O(nprobe × n/nlist) per query - much faster than FLAT
- Accuracy: ~90-95% recall (configurable via nprobe)
- Memory: Similar to FLAT + cluster overhead
- Requires: Training on sample data before use
- Use case: Production systems with medium-sized datasets
HNSW (IndexHNSW)
- Best for: Large datasets (> 100k vectors), best speed/accuracy tradeoff
- Speed: O(log n) per query - logarithmic search
- Accuracy: ~95-99% recall (configurable via efSearch)
- Memory: ~1.5-2× more than FLAT
- No training required
- Use case: Large-scale production systems, best overall performance
Examples
Basic Semantic Search
const { FaissIndex } = require('@faiss-node/native');
// Create index for 768-dimensional embeddings (e.g., OpenAI)
const index = new FaissIndex({ type: 'HNSW', dims: 768 });
// Add document embeddings
const documents = [
{ id: 0, text: "JavaScript is a programming language" },
{ id: 1, text: "Python is great for data science" },
{ id: 2, text: "Node.js runs JavaScript on the server" }
];
const embeddings = new Float32Array(/* ... your embeddings ... */);
await index.add(embeddings);
// Search for similar documents
const queryEmbedding = new Float32Array(/* ... query embedding ... */);
const results = await index.search(queryEmbedding, 3);
console.log('Most similar documents:', results.labels);RAG Pipeline
const { FaissIndex } = require('@faiss-node/native');
class RAGSystem {
constructor() {
this.index = new FaissIndex({ type: 'HNSW', dims: 1536 });
this.documents = [];
}
async addDocuments(docs, embeddings) {
this.documents.push(...docs);
await this.index.add(embeddings);
}
async search(queryEmbedding, k = 5) {
const results = await this.index.search(queryEmbedding, k);
return results.labels.map(idx => this.documents[idx]);
}
async save(path) {
await this.index.save(path);
// Also save documents mapping
}
}Persistence
const { FaissIndex } = require('@faiss-node/native');
// Save to disk
const index = new FaissIndex({ type: 'HNSW', dims: 128 });
await index.add(vectors);
await index.save('./index.faiss');
// Load from disk
const loadedIndex = await FaissIndex.load('./index.faiss');
// Or serialize to buffer (for databases)
const buffer = await index.toBuffer();
// Store in MongoDB, Redis, etc.
const restoredIndex = await FaissIndex.fromBuffer(buffer);Performance Tips
- Use HNSW for large datasets - Best overall performance
- Batch operations - Use
searchBatch()for multiple queries - Train IVF properly - Use 10k-100k training vectors
- Tune parameters - Increase
nprobe(IVF) orefSearch(HNSW) for accuracy - Reuse indexes - Save/load instead of recreating
Thread Safety
All operations are thread-safe and can be called concurrently:
// Safe to call from multiple async operations
await Promise.all([
index.add(vectors1),
index.add(vectors2),
index.search(query1),
index.search(query2)
]);The implementation uses mutex locks to ensure FAISS operations are serialized safely.
Error Handling
All methods throw JavaScript errors (not C++ exceptions):
try {
await index.add(vectors);
} catch (error) {
if (error.message.includes('disposed')) {
console.error('Index was disposed');
} else if (error.message.includes('dimensions')) {
console.error('Vector dimensions mismatch');
}
}TypeScript Support
Full TypeScript definitions are included:
import { FaissIndex, FaissIndexConfig, SearchResults } from '@faiss-node/native';
const config: FaissIndexConfig = {
type: 'HNSW',
dims: 768
};
const index = new FaissIndex(config);
const results: SearchResults = await index.search(query, 10);Updating
To update to the latest version:
npm update @faiss-node/nativeOr install a specific version:
npm install @faiss-node/native@0.1.2Development
Building from Source
# Clone repository
git clone https://github.com/anupammaurya6767/faiss-node-native.git
cd faiss-node-native
# Install dependencies
npm install
# Build native module
npm run build
# Run tests
npm testRunning Tests
npm test # All tests
npm run test:unit # Unit tests only
npm run test:integration # Integration tests only
npm run test:ci # CI tests (faster, no manual tests)Generating Documentation
npm run docs # Generate Doxygen documentation
npm run docs:serve # Serve docs locally at http://localhost:8000Documentation
- API Documentation: GitHub Pages
- Examples: See
examples/directory - Contributing: See CONTRIBUTING.md
Troubleshooting
Build Errors
macOS: "library not found"
# Ensure FAISS is installed
brew install faiss
# Check installation
ls /usr/local/lib/libfaiss*Linux: "faiss/Index.h: No such file or directory"
# Build and install FAISS from source (see Prerequisites)
# Ensure CMAKE_INSTALL_PREFIX=/usr/localRuntime Errors
"Index has been disposed"
- You called
dispose()or the index was garbage collected - Create a new index or don't dispose until done
"Vector dimensions don't match"
- Check that your vectors are the correct size
- For batch operations:
vectors.length % dims === 0
Contributing
Contributions are welcome! Please see CONTRIBUTING.md for guidelines.
License
MIT License - see LICENSE file for details.
Author
Anupam Maurya
- GitHub: @anupammaurya6767
- Email: anupammaurya6767@gmail.com
Acknowledgments
- Built on Facebook FAISS - the amazing vector similarity search library
- Inspired by the need for high-performance vector search in Node.js
- Thanks to the open-source community for feedback and contributions
Made with ❤️ for the Node.js community