ripgrep.ts
utils/ripgrep.ts
680
Lines
21188
Bytes
6
Exports
15
Imports
10
Keywords
What this is
This page documents one file from the repository and includes its full source so you can read it without leaving the docs site.
Beginner explanation
This file is one piece of the larger system. Its name, directory, imports, and exports show where it fits. Start by reading the exports and related files first.
How it is used
Start from the exports list and related files. Those are the easiest clues for where this file fits into the system.
Expert explanation
Architecturally, this file intersects with repo-context. It contains 680 lines, 15 detected imports, and 6 detected exports.
Important relationships
Detected exports
ripgrepCommandRipgrepTimeoutErrorripGrepStreamripGrepcountFilesRoundedRggetRipgrepStatus
Keywords
ripgrepstdoutcodeabortsignalstderrlinesconfigargv0childpath
Detected imports
child_processchild_processlodash-es/memoize.jsospathsrc/services/analytics/index.jsurl./bundledMode.js./debug.js./envUtils.js./execFileNoThrow.js./findExecutable.js./log.js./platform.js./stringUtils.js
Source notes
This page embeds the full file contents. Small or leaf files are still indexed honestly instead of being over-explained.
Full source
import type { ChildProcess, ExecFileException } from 'child_process'
import { execFile, spawn } from 'child_process'
import memoize from 'lodash-es/memoize.js'
import { homedir } from 'os'
import * as path from 'path'
import { logEvent } from 'src/services/analytics/index.js'
import { fileURLToPath } from 'url'
import { isInBundledMode } from './bundledMode.js'
import { logForDebugging } from './debug.js'
import { isEnvDefinedFalsy } from './envUtils.js'
import { execFileNoThrow } from './execFileNoThrow.js'
import { findExecutable } from './findExecutable.js'
import { logError } from './log.js'
import { getPlatform } from './platform.js'
import { countCharInString } from './stringUtils.js'
const __filename = fileURLToPath(import.meta.url)
// we use node:path.join instead of node:url.resolve because the former doesn't encode spaces
const __dirname = path.join(
__filename,
process.env.NODE_ENV === 'test' ? '../../../' : '../',
)
type RipgrepConfig = {
mode: 'system' | 'builtin' | 'embedded'
command: string
args: string[]
argv0?: string
}
const getRipgrepConfig = memoize((): RipgrepConfig => {
const userWantsSystemRipgrep = isEnvDefinedFalsy(
process.env.USE_BUILTIN_RIPGREP,
)
// Try system ripgrep if user wants it
if (userWantsSystemRipgrep) {
const { cmd: systemPath } = findExecutable('rg', [])
if (systemPath !== 'rg') {
// SECURITY: Use command name 'rg' instead of systemPath to prevent PATH hijacking
// If we used systemPath, a malicious ./rg.exe in current directory could be executed
// Using just 'rg' lets the OS resolve it safely with NoDefaultCurrentDirectoryInExePath protection
return { mode: 'system', command: 'rg', args: [] }
}
}
// In bundled (native) mode, ripgrep is statically compiled into bun-internal
// and dispatches based on argv[0]. We spawn ourselves with argv0='rg'.
if (isInBundledMode()) {
return {
mode: 'embedded',
command: process.execPath,
args: ['--no-config'],
argv0: 'rg',
}
}
const rgRoot = path.resolve(__dirname, 'vendor', 'ripgrep')
const command =
process.platform === 'win32'
? path.resolve(rgRoot, `${process.arch}-win32`, 'rg.exe')
: path.resolve(rgRoot, `${process.arch}-${process.platform}`, 'rg')
return { mode: 'builtin', command, args: [] }
})
export function ripgrepCommand(): {
rgPath: string
rgArgs: string[]
argv0?: string
} {
const config = getRipgrepConfig()
return {
rgPath: config.command,
rgArgs: config.args,
argv0: config.argv0,
}
}
const MAX_BUFFER_SIZE = 20_000_000 // 20MB; large monorepos can have 200k+ files
/**
* Check if an error is EAGAIN (resource temporarily unavailable).
* This happens in resource-constrained environments (Docker, CI) when
* ripgrep tries to spawn too many threads.
*/
function isEagainError(stderr: string): boolean {
return (
stderr.includes('os error 11') ||
stderr.includes('Resource temporarily unavailable')
)
}
/**
* Custom error class for ripgrep timeouts.
* This allows callers to distinguish between "no matches" and "timed out".
*/
export class RipgrepTimeoutError extends Error {
constructor(
message: string,
public readonly partialResults: string[],
) {
super(message)
this.name = 'RipgrepTimeoutError'
}
}
function ripGrepRaw(
args: string[],
target: string,
abortSignal: AbortSignal,
callback: (
error: ExecFileException | null,
stdout: string,
stderr: string,
) => void,
singleThread = false,
): ChildProcess {
// NB: When running interactively, ripgrep does not require a path as its last
// argument, but when run non-interactively, it will hang unless a path or file
// pattern is provided
const { rgPath, rgArgs, argv0 } = ripgrepCommand()
// Use single-threaded mode only if explicitly requested for this call's retry
const threadArgs = singleThread ? ['-j', '1'] : []
const fullArgs = [...rgArgs, ...threadArgs, ...args, target]
// Allow timeout to be configured via env var (in seconds), otherwise use platform defaults
// WSL has severe performance penalty for file reads (3-5x slower on WSL2)
const defaultTimeout = getPlatform() === 'wsl' ? 60_000 : 20_000
const parsedSeconds =
parseInt(process.env.CLAUDE_CODE_GLOB_TIMEOUT_SECONDS || '', 10) || 0
const timeout = parsedSeconds > 0 ? parsedSeconds * 1000 : defaultTimeout
// For embedded ripgrep, use spawn with argv0 (execFile doesn't support argv0 properly)
if (argv0) {
const child = spawn(rgPath, fullArgs, {
argv0,
signal: abortSignal,
// Prevent visible console window on Windows (no-op on other platforms)
windowsHide: true,
})
let stdout = ''
let stderr = ''
let stdoutTruncated = false
let stderrTruncated = false
child.stdout?.on('data', (data: Buffer) => {
if (!stdoutTruncated) {
stdout += data.toString()
if (stdout.length > MAX_BUFFER_SIZE) {
stdout = stdout.slice(0, MAX_BUFFER_SIZE)
stdoutTruncated = true
}
}
})
child.stderr?.on('data', (data: Buffer) => {
if (!stderrTruncated) {
stderr += data.toString()
if (stderr.length > MAX_BUFFER_SIZE) {
stderr = stderr.slice(0, MAX_BUFFER_SIZE)
stderrTruncated = true
}
}
})
// Set up timeout with SIGKILL escalation.
// SIGTERM alone may not kill ripgrep if it's blocked in uninterruptible I/O
// (e.g., deep filesystem traversal). If SIGTERM doesn't work within 5 seconds,
// escalate to SIGKILL which cannot be caught or ignored.
// On Windows, child.kill('SIGTERM') throws; use default signal.
let killTimeoutId: ReturnType<typeof setTimeout> | undefined
const timeoutId = setTimeout(() => {
if (process.platform === 'win32') {
child.kill()
} else {
child.kill('SIGTERM')
killTimeoutId = setTimeout(c => c.kill('SIGKILL'), 5_000, child)
}
}, timeout)
// On Windows, both 'close' and 'error' can fire for the same process
// (e.g. when AbortSignal kills the child). Guard against double-callback.
let settled = false
child.on('close', (code, signal) => {
if (settled) return
settled = true
clearTimeout(timeoutId)
clearTimeout(killTimeoutId)
if (code === 0 || code === 1) {
// 0 = matches found, 1 = no matches (both are success)
callback(null, stdout, stderr)
} else {
const error: ExecFileException = new Error(
`ripgrep exited with code ${code}`,
)
error.code = code ?? undefined
error.signal = signal ?? undefined
callback(error, stdout, stderr)
}
})
child.on('error', (err: NodeJS.ErrnoException) => {
if (settled) return
settled = true
clearTimeout(timeoutId)
clearTimeout(killTimeoutId)
const error: ExecFileException = err
callback(error, stdout, stderr)
})
return child
}
// For non-embedded ripgrep, use execFile
// Use SIGKILL as killSignal because SIGTERM may not terminate ripgrep
// when it's blocked in uninterruptible filesystem I/O.
// On Windows, SIGKILL throws; use default (undefined) which sends SIGTERM.
return execFile(
rgPath,
fullArgs,
{
maxBuffer: MAX_BUFFER_SIZE,
signal: abortSignal,
timeout,
killSignal: process.platform === 'win32' ? undefined : 'SIGKILL',
},
callback,
)
}
/**
* Stream-count lines from `rg --files` without buffering stdout.
*
* On large repos (e.g. 247k files, 16MB of paths), calling `ripGrep()` just
* to read `.length` materializes the full stdout string plus a 247k-element
* array. This counts newline bytes per chunk instead; peak memory is one
* stream chunk (~64KB).
*
* Intentionally minimal: the only caller is telemetry (countFilesRoundedRg),
* which swallows all errors. No EAGAIN retry, no stderr capture, no internal
* timeout (callers pass AbortSignal.timeout; spawn's signal option kills rg).
*/
async function ripGrepFileCount(
args: string[],
target: string,
abortSignal: AbortSignal,
): Promise<number> {
await codesignRipgrepIfNecessary()
const { rgPath, rgArgs, argv0 } = ripgrepCommand()
return new Promise<number>((resolve, reject) => {
const child = spawn(rgPath, [...rgArgs, ...args, target], {
argv0,
signal: abortSignal,
windowsHide: true,
stdio: ['ignore', 'pipe', 'ignore'],
})
let lines = 0
child.stdout?.on('data', (chunk: Buffer) => {
lines += countCharInString(chunk, '\n')
})
// On Windows, both 'close' and 'error' can fire for the same process.
let settled = false
child.on('close', code => {
if (settled) return
settled = true
if (code === 0 || code === 1) resolve(lines)
else reject(new Error(`rg --files exited ${code}`))
})
child.on('error', err => {
if (settled) return
settled = true
reject(err)
})
})
}
/**
* Stream lines from ripgrep as they arrive, calling `onLines` per stdout chunk.
*
* Unlike `ripGrep()` which buffers the entire stdout, this flushes complete
* lines as soon as each chunk arrives — first results paint while rg is still
* walking the tree (the fzf `change:reload` pattern). Partial trailing lines
* are carried across chunk boundaries.
*
* Callers that want to stop early (e.g. after N matches) should abort the
* signal — spawn's signal option kills rg. No EAGAIN retry, no internal
* timeout, stderr is ignored; interactive callers own recovery.
*/
export async function ripGrepStream(
args: string[],
target: string,
abortSignal: AbortSignal,
onLines: (lines: string[]) => void,
): Promise<void> {
await codesignRipgrepIfNecessary()
const { rgPath, rgArgs, argv0 } = ripgrepCommand()
return new Promise<void>((resolve, reject) => {
const child = spawn(rgPath, [...rgArgs, ...args, target], {
argv0,
signal: abortSignal,
windowsHide: true,
stdio: ['ignore', 'pipe', 'ignore'],
})
const stripCR = (l: string) => (l.endsWith('\r') ? l.slice(0, -1) : l)
let remainder = ''
child.stdout?.on('data', (chunk: Buffer) => {
const data = remainder + chunk.toString()
const lines = data.split('\n')
remainder = lines.pop() ?? ''
if (lines.length) onLines(lines.map(stripCR))
})
// On Windows, both 'close' and 'error' can fire for the same process.
let settled = false
child.on('close', code => {
if (settled) return
// Abort races close — don't flush a torn tail from a killed process.
// Promise still settles: spawn's signal option fires 'error' with
// AbortError → reject below.
if (abortSignal.aborted) return
settled = true
if (code === 0 || code === 1) {
if (remainder) onLines([stripCR(remainder)])
resolve()
} else {
reject(new Error(`ripgrep exited with code ${code}`))
}
})
child.on('error', err => {
if (settled) return
settled = true
reject(err)
})
})
}
export async function ripGrep(
args: string[],
target: string,
abortSignal: AbortSignal,
): Promise<string[]> {
await codesignRipgrepIfNecessary()
// Test ripgrep on first use and cache the result (fire and forget)
void testRipgrepOnFirstUse().catch(error => {
logError(error)
})
return new Promise((resolve, reject) => {
const handleResult = (
error: ExecFileException | null,
stdout: string,
stderr: string,
isRetry: boolean,
): void => {
// Success case
if (!error) {
resolve(
stdout
.trim()
.split('\n')
.map(line => line.replace(/\r$/, ''))
.filter(Boolean),
)
return
}
// Exit code 1 is normal "no matches"
if (error.code === 1) {
resolve([])
return
}
// Critical errors that indicate ripgrep is broken, not "no matches"
// These should be surfaced to the user rather than silently returning empty results
const CRITICAL_ERROR_CODES = ['ENOENT', 'EACCES', 'EPERM']
if (CRITICAL_ERROR_CODES.includes(error.code as string)) {
reject(error)
return
}
// If we hit EAGAIN and haven't retried yet, retry with single-threaded mode
// Note: We only use -j 1 for this specific retry, not for future calls.
// Persisting single-threaded mode globally caused timeouts on large repos
// where EAGAIN was just a transient startup error.
if (!isRetry && isEagainError(stderr)) {
logForDebugging(
`rg EAGAIN error detected, retrying with single-threaded mode (-j 1)`,
)
logEvent('tengu_ripgrep_eagain_retry', {})
ripGrepRaw(
args,
target,
abortSignal,
(retryError, retryStdout, retryStderr) => {
handleResult(retryError, retryStdout, retryStderr, true)
},
true, // Force single-threaded mode for this retry only
)
return
}
// For all other errors, try to return partial results if available
const hasOutput = stdout && stdout.trim().length > 0
const isTimeout =
error.signal === 'SIGTERM' ||
error.signal === 'SIGKILL' ||
error.code === 'ABORT_ERR'
const isBufferOverflow =
error.code === 'ERR_CHILD_PROCESS_STDIO_MAXBUFFER'
let lines: string[] = []
if (hasOutput) {
lines = stdout
.trim()
.split('\n')
.map(line => line.replace(/\r$/, ''))
.filter(Boolean)
// Drop last line for timeouts and buffer overflow - it may be incomplete
if (lines.length > 0 && (isTimeout || isBufferOverflow)) {
lines = lines.slice(0, -1)
}
}
logForDebugging(
`rg error (signal=${error.signal}, code=${error.code}, stderr: ${stderr}), ${lines.length} results`,
)
// code 2 = ripgrep usage error (already handled); ABORT_ERR = caller
// explicitly aborted (not an error, just a cancellation — interactive
// callers may abort on every keystroke-after-debounce).
if (error.code !== 2 && error.code !== 'ABORT_ERR') {
logError(error)
}
// If we timed out with no results, throw an error so Claude knows the search
// didn't complete rather than thinking there were no matches
if (isTimeout && lines.length === 0) {
reject(
new RipgrepTimeoutError(
`Ripgrep search timed out after ${getPlatform() === 'wsl' ? 60 : 20} seconds. The search may have matched files but did not complete in time. Try searching a more specific path or pattern.`,
lines,
),
)
return
}
resolve(lines)
}
ripGrepRaw(args, target, abortSignal, (error, stdout, stderr) => {
handleResult(error, stdout, stderr, false)
})
})
}
/**
* Count files in a directory recursively using ripgrep and round to the nearest power of 10 for privacy
*
* This is much more efficient than using native Node.js methods for counting files
* in large directories since it uses ripgrep's highly optimized file traversal.
*
* @param path Directory path to count files in
* @param abortSignal AbortSignal to cancel the operation
* @param ignorePatterns Optional additional patterns to ignore (beyond .gitignore)
* @returns Approximate file count rounded to the nearest power of 10
*/
export const countFilesRoundedRg = memoize(
async (
dirPath: string,
abortSignal: AbortSignal,
ignorePatterns: string[] = [],
): Promise<number | undefined> => {
// Skip file counting if we're in the home directory to avoid triggering
// macOS TCC permission dialogs for Desktop, Downloads, Documents, etc.
if (path.resolve(dirPath) === path.resolve(homedir())) {
return undefined
}
try {
// Build ripgrep arguments:
// --files: List files that would be searched (rather than searching them)
// --count: Only print a count of matching lines for each file
// --no-ignore-parent: Don't respect ignore files in parent directories
// --hidden: Search hidden files and directories
const args = ['--files', '--hidden']
// Add ignore patterns if provided
ignorePatterns.forEach(pattern => {
args.push('--glob', `!${pattern}`)
})
const count = await ripGrepFileCount(args, dirPath, abortSignal)
// Round to nearest power of 10 for privacy
if (count === 0) return 0
const magnitude = Math.floor(Math.log10(count))
const power = Math.pow(10, magnitude)
// Round to nearest power of 10
// e.g., 8 -> 10, 42 -> 100, 350 -> 100, 750 -> 1000
return Math.round(count / power) * power
} catch (error) {
// AbortSignal.timeout firing is expected on large/slow repos, not an error.
if ((error as Error)?.name !== 'AbortError') logError(error)
}
},
// lodash memoize's default resolver only uses the first argument.
// ignorePatterns affect the result, so include them in the cache key.
// abortSignal is intentionally excluded — it doesn't affect the count.
(dirPath, _abortSignal, ignorePatterns = []) =>
`${dirPath}|${ignorePatterns.join(',')}`,
)
// Singleton to store ripgrep availability status
let ripgrepStatus: {
working: boolean
lastTested: number
config: RipgrepConfig
} | null = null
/**
* Get ripgrep status and configuration info
* Returns current configuration immediately, with working status if available
*/
export function getRipgrepStatus(): {
mode: 'system' | 'builtin' | 'embedded'
path: string
working: boolean | null // null if not yet tested
} {
const config = getRipgrepConfig()
return {
mode: config.mode,
path: config.command,
working: ripgrepStatus?.working ?? null,
}
}
/**
* Test ripgrep availability on first use and cache the result
*/
const testRipgrepOnFirstUse = memoize(async (): Promise<void> => {
// Already tested
if (ripgrepStatus !== null) {
return
}
const config = getRipgrepConfig()
try {
let test: { code: number; stdout: string }
// For embedded ripgrep, use Bun.spawn with argv0
if (config.argv0) {
// Only Bun embeds ripgrep.
// eslint-disable-next-line custom-rules/require-bun-typeof-guard
const proc = Bun.spawn([config.command, '--version'], {
argv0: config.argv0,
stderr: 'ignore',
stdout: 'pipe',
})
// Bun's ReadableStream has .text() at runtime, but TS types don't reflect it
const [stdout, code] = await Promise.all([
(proc.stdout as unknown as Blob).text(),
proc.exited,
])
test = {
code,
stdout,
}
} else {
test = await execFileNoThrow(
config.command,
[...config.args, '--version'],
{
timeout: 5000,
},
)
}
const working =
test.code === 0 && !!test.stdout && test.stdout.startsWith('ripgrep ')
ripgrepStatus = {
working,
lastTested: Date.now(),
config,
}
logForDebugging(
`Ripgrep first use test: ${working ? 'PASSED' : 'FAILED'} (mode=${config.mode}, path=${config.command})`,
)
// Log telemetry for actual ripgrep availability
logEvent('tengu_ripgrep_availability', {
working: working ? 1 : 0,
using_system: config.mode === 'system' ? 1 : 0,
})
} catch (error) {
ripgrepStatus = {
working: false,
lastTested: Date.now(),
config,
}
logError(error)
}
})
let alreadyDoneSignCheck = false
async function codesignRipgrepIfNecessary() {
if (process.platform !== 'darwin' || alreadyDoneSignCheck) {
return
}
alreadyDoneSignCheck = true
// Only sign the standalone vendored rg binary (npm builds)
const config = getRipgrepConfig()
if (config.mode !== 'builtin') {
return
}
const builtinPath = config.command
// First, check to see if ripgrep is already signed
const lines = (
await execFileNoThrow('codesign', ['-vv', '-d', builtinPath], {
preserveOutputOnError: false,
})
).stdout.split('\n')
const needsSigned = lines.find(line => line.includes('linker-signed'))
if (!needsSigned) {
return
}
try {
const signResult = await execFileNoThrow('codesign', [
'--sign',
'-',
'--force',
'--preserve-metadata=entitlements,requirements,flags,runtime',
builtinPath,
])
if (signResult.code !== 0) {
logError(
new Error(
`Failed to sign ripgrep: ${signResult.stdout} ${signResult.stderr}`,
),
)
}
const quarantineResult = await execFileNoThrow('xattr', [
'-d',
'com.apple.quarantine',
builtinPath,
])
if (quarantineResult.code !== 0) {
logError(
new Error(
`Failed to remove quarantine: ${quarantineResult.stdout} ${quarantineResult.stderr}`,
),
)
}
} catch (e) {
logError(e)
}
}