apps/docs/content/troubleshooting/edge-function-546-error-response.mdx
A 546 error indicates that an edge function used more resources (CPU or Memory) than it was allocated.
Edge functions run in transient servers called isolates. Each isolate:
func_one will never serve func_two)When a request arrives, the runtime assigns it to a free isolate or spins up a new one if all existing isolates are busy. Each isolate also has resource limitations.
| Resource | Limit |
|---|---|
| CPU cycles | 2s |
| Memory | 250MB |
Once an isolate uses 50% of any resource, it will finish the current request and then shut down.
However, if that remaining request exhausts all CPU or memory before completion, the isolate will terminate immediately and return a 546 response.
When an edge function fails due to internal CPU or memory limits, it will return the error:
{
"code": "WORKER_LIMIT",
"message": "Function failed due to not having enough compute resources (please check logs)"
}
In the function dashboard's Logs tab, you can find the specific error message:
Memory limit exceededCPU Time exceededAlternatively, you can filter for the specific errors from the function using the log explorer
select
fl.event_message,
content.timestamp,
fel.function_name,
fel.status_code
from
function_logs as fl
left join UNNEST(fl.metadata) as content on true
left join (
select
em.execution_id,
req.pathname as function_name,
res.status_code
from
function_edge_logs
left join UNNEST(metadata) as em on true
left join UNNEST(em.request) as req on true
left join UNNEST(em.response) as res on true
) as fel
on content.execution_id = fel.execution_id
where content.level = 'error' and fel.status_code = 546
order by timestamp, function_name
limit 20;
Before optimizing, run the below query in the Log Explorer to understand how often 546s are occurring relative to total requests:
select
COUNT(id) as total_responses,
COUNTIF(response.status_code = 546) as total_546,
SAFE_DIVIDE(COUNTIF(response.status_code = 546), COUNT(*)) * 100 as pct_546
from
function_edge_logs
cross join UNNEST(function_edge_logs.metadata) as metadata
cross join UNNEST(metadata.response) as response
cross join UNNEST(metadata.request) as request
where method != 'OPTIONS' and pathname = '/functions/v1/YOUR_FUNCTION_NAME';
-- <-- add your function name to inspect specific endpoints
Depending on the results, you may be able to determine if the event is an edge case or affecting a function's overall behavior.
| 546-rate | What it likely means |
|---|---|
| < 5% | May be an anomaly or edge case with how your function is structure or responds to payloads. May be acceptable for your use case |
| 5-50% | Affecting a meaningful portion of traffic |
| > 50% | Nearly all requests are over-resourced; the function needs significant work |
The same constraints placed on edge function's hosted by Supabase are also imposed by the test environment spun-up by the CLI. You can follow the function's local development guide to set up a test environment and then serve your function locally:
supabase functions serve your-function --debug
Then try experimenting with different stress tests to see if you can induce 546s. Some tests worth trying may involve:
If you find a reliable way to induce the error, you may want to log between operations to gain more visibility or configure chrome dev-tools to pinpoint the underlying logic that is failing.
There are a few other queries that may be useful for identifying patterns around 546 errors.
<Accordion type="default" openBehaviour="multiple" chevronAlign="right" justified size="medium" className="text-foreground-light mt-8 mb-6"
<div className="border-b mt-3 pb-3"> <AccordionItem header="Checking if a specific version is an offender" id="item-1" > Every time you update a function, its version number increments. It may be that you made an update and it's only a specific version that is problematic.
You can check error frequency by version with the below query:
select
COUNT(id) as total_responses,
version,
COUNTIF(response.status_code = 546) as total_546,
SAFE_DIVIDE(COUNTIF(response.status_code = 546), COUNT(*)) * 100 as pct_546
from
function_edge_logs
cross join UNNEST(function_edge_logs.metadata) as metadata
cross join UNNEST(metadata.response) as response
cross join UNNEST(metadata.request) as request
where method != 'OPTIONS' and pathname = '/functions/v1/FUNCTION_NAME' -- <--OPTIONAL FILTER: add specific function name to target query
group by version
having pct_546 > 5 -- <--Failure percentage threshold. The query only shows versions with a 5% or above 546 error rate
order by pct_546;
</AccordionItem>
You can check to see how frequent 546 errors are per hour with the below query:
```sql
SELECT FORMAT_TIMESTAMP("%Y-%m-%d %H:00", TIMESTAMP(timestamp), "UTC") AS hour, COUNT(id) AS total_responses,
COUNTIF(response.status_code = 546) AS total_546,
SAFE_DIVIDE(COUNTIF(response.status_code = 546), COUNT(id)) * 100 AS pct_546
FROM function_edge_logs CROSS JOIN UNNEST(function_edge_logs.metadata) AS metadata CROSS JOIN UNNEST(metadata.response) AS response CROSS JOIN UNNEST(metadata.request) AS request WHERE pathname = '/functions/v1/FUNCTION_NAME'--<--OPTIONAL FILTER: add specific function name to target query group by hour ORDER by hour DESC LIMIT 24;
The output may look like:

If the failures are concentrated to a specific time, you can check if you made any updates around that period or if users were engaging in atypical behavior, such as sending larger payloads.
</AccordionItem>
</div>
<div className="border-b mt-3 pb-3">
<AccordionItem
header="Check requests per isolate"
id="item-3"
>
If your isolates are serving more than 2 requests before retiring, it suggests variability in how much processing each request needs. In that case, you may want to cross compare your successful requests with your failed ones. Maybe there's a query parameter or specific content-length header that makes failures more likely.
```sql
SELECT
COUNT(fel.id) AS requests_served,
metadata.execution_id AS isolate_id
FROM function_logs
LEFT JOIN UNNEST(function_logs.metadata) AS metadata ON TRUE
LEFT JOIN (
SELECT
em.execution_id,
id,
pathname,
method
FROM function_edge_logs
LEFT JOIN UNNEST(function_edge_logs.metadata) AS em ON TRUE
LEFT JOIN UNNEST(em.request) AS req ON TRUE
) fel ON metadata.execution_id = fel.execution_id
WHERE
metadata.reason IN ('Memory', 'CPUTime')
AND
method <> 'OPTIONS' --ignore OPTION requests
AND
pathname = '/functions/v1/FUNCTION_NAME' --<-- add your function name to inspect specific endpoints
GROUP BY metadata.execution_id
</AccordionItem>
The only way to manage the error is to reduce resource consumption per request. There are a few strategies one can go about.
If you believe a portion of your function is overly aggressive, try testing locally whether refactoring reduces resource overuse.
Common culprits:
CPU intensive recursions: intensive loops or recursion can quickly exhaust CPU
// This will exhaust CPU allocation when called repeatedly
function fib(n: number): number {
if (n <= 1) return n;
return fib(n - 1) + fib(n - 2); // high levels of recursion
}
for (let i = 0; i < 100; i++) {
fib(40);
}
Unbounded memory allocation: filling large arrays in a tight loop prevents the garbage collector from freeing memory
// Each iteration allocates ~100s of KB. During the loops, all memory is consumed before GC can intervene
let ref = []
for (let i = 0; i < 1000; i++) {
ref.push(new Array(10e4).fill('data'))
}
You can compare your function against working examples in the Edge Function docs for insight on how to rework your code.
If you're using a dependency that does more than you need, look for a lighter or more performant alternative.
If you are performing logic to process data from Supabase Postgres, you may be able to handle the processing within the database directly by using database functions or refactored queries.
Instead of managing all operations within the function itself, there may be an external API that can execute CPU or memory intensive jobs on its behalf. One example would be using an external API for orchestrating a headless browser and then just using the edge function to manage the output of the activity instead of everything all in place.
Break a large function into smaller ones, each responsible for a single sub-task. Stitch the results together at the app level or via an orchestrating function.
<Admonition type="caution">If you have functions that call other functions, always implement an escape condition. Supabase will terminate functions that recursively self-call past a certain depth, but your code should enforce its own limit.
</Admonition>Edge functions have a hard resource limit. If your work requires more resources than we permit, you can look into other solutions, such as AWS Lambda, that are less restrictive, or self-host edge functions and reconfigure the settings.
Performing edits against images or other large files can be both CPU and Memory intensive. Some approaches for reducing load is using more performant processing libraries, processing outside by using an API or the requester's server, or restricting the file size to reduce strain.
AI models process data into embeddings (large arrays), that they can more understand. Edge Functions are capable of managing some small models directly; however, some require more processing power than what the edge function can support directly. In these cases, the solution is to manage the embeddings via an external source, such as OpenAI, Anthropic, etc. and to just use the edge function for light processing and coordination.
Web scraping often requires a headless browser operator, such as puppeteer or playwright for rendering web pages. In this case, it is better to use an external API to manage the headless browser for you and then parse the results it returns with the edge function. There's an example in the function docs: Taking Screenshots with Puppeteer