68 results found
-
Automatic Deployment of Changed Firebase Functions
Firebase should offer a built-in feature to automatically detect changes to Cloud Functions and deploy only those functions, grouped into manageable batches to avoid rate limits. This feature would streamline deployments by:
- Scanning the project directory to identify functions with modified code since the last deployment.
- Automatically batching the deployment of changed functions based on rate limit thresholds (e.g., 8-10 functions per batch).
- Handling the deployment process transparently within the Firebase CLI or CI/CD integrations, reducing manual intervention.
This functionality would save time, minimize deployment errors, and improve developer workflows, particularly in large projects with numerous functions.
4 votes -
Make Firebase Functions Test a First Grade citizen
A solid Firebase Functions experience requires a fully functional test suite.
Firebase-functions-test is already there, and provides the experience, but is often forgotten for months.Firebase functions 6 was released a month ago, and it broke functions-test.
There's currently a patch that fixes it, and it's been there for a week.
This idea isn't about this issue. It's about giving resources to the firebase-functions-text project, so it can continue growing and thus giving cloud firebase functions users a solid testing experience before deploying.
4 votes -
not upgrading blaze plan
firebase is not updateing to blaze plan i already have an verified billing account you should have to give more information about how to setup properly
4 votes -
Add an argument that automatically chunks cloud function deployments according to the maximal possible quota
When deploying multiple cloud functions at once via
firebase deploy --only functionswe frequently run into quota issues like:
functions: got "Quota Exceeded" error while trying to update projects/.... Waiting to retry...
It never retries though.
It would make it way easier if you could add an argument to the deployment command like:
firebase deploy --only functions --chunked 504 votes -
Supports Cloud SQL connection with Firebase Functions V2 natively as it was with V1
Migrating a Firebase Functions project from V1 to V2 using Cloud SQL is not feasible. It appears that Cloud SQL connections are not yet supported in V2. Cloud SQL connection works automatically and required no actions in V1. It would be beneficial if connecting to a Cloud SQL instance using V2 could be made as simple as it was before.
4 votes -
Infinite loop/recursion detection
It's possible to make a mistake that causes a function to trigger itself (for example, an onDocumentWritten Firestore trigger that writes to the same document). If Cloud Functions for Firebase could detect this and shut down the infinite loop, it could prevent runaway bills.
It could also be a feature in the emulator suite so that we could check for them before deploying to production.
3 votes -
Allow secrets on scheduled functions
It's not clear in the documentation nor by implementation that secrets work with scheduled cloud functions. This seems to me like a major missing feature in security and consistency with the v2 cloud functions. Is the only solution using the cloud config for api keys/secrets?
We need to have something like:
exports.someScheduledFunction = onSchedule(
{
schedule: "every 60 minutes",
timeoutSeconds: 100,
secrets: [someToken.name],
},
async () => {
logInfo("Running every 60 min");...
}
);3 votes -
Rolex Enterprises
Chair service
3 votes -
Tell people that mixing V1 and V2 config files in the same project will create a port and container error.
Tell people that mixing V1 and V2 config files in the same project (note: even outside the gen2 function) will create a port and container error. Right now you don't mention in the docs or anywhere that having ANY function.config files in a project with a gen 2 function will cause a deploy error that says a container is broken and the port is unreachable. This error is misleading. It will also show the gen 2 function in the console as "unknown trigger" which also is misleading.
In all your documentation you say you can have gen1 and gen2 functions…
3 votes -
Flutterflow despliegues correctos
muchas veces flutterflow no permite desplegar cloudFuctions y es muy importante el poder crear cloudFuctions desde flutterflow
3 votes -
Ability to split Python Functions into multiple files (=enable local imports)
Right now it is only possible to write functions in
main.py, which very quickly gets overwhelmingly large. It would be good to be able to organize code into multiple files.3 votes -
Green/Blue deployment pattern
Allow for deployments to be either a green or blue deployment to be able to have multiple instances of the same functions deployed within a single Firebase project.
This would help cover the gap of time between releases of client apps and Cloud Functions, which if not handled correctly cause production issues.
3 votes -
Allow ordering functions when they are displayed in firebase console.
As a non-expert user, I am trying to avoid navigating the Google Cloud Console ("GCC") when possible (which should be understandable since that is largely the point of Firebase). So when an internal error in one of my functions occurs, I go to the firebase console, find the relevant function and then select "see logs" which takes me the to the precise place in GCC. The issue is that the function in the Firebase console seem to be displayed in random order with no way of ordering them... so it is always a pain to scan through them to find…
2 votes -
Should add basic REDIS/ Memcache Like Capability
Why I Want This?
Firebase splits functions to individual Docker Container. This works great for scaling, but with that it is not easy to share global values, for example, user's access key to external services and session specific values.
While we can use REDIS and such hosted on GCP, it would be more "batteries included" platform if this feature is offered as a built-in functionality for the functions without resorting to Firestore or RTDB.
What I look for is to globally share key values and also auto-expire them.
2 votes -
Sub Collection View
I feel the window to view sub collection on the firebase firestore console could be expandable.
2 votes -
It is necessary to change and update in advance the spread of better ideas that benefit people and reduce harmful materials
It is necessary to change and update in advance the spread of better ideas that benefit people and reduce harmful materials
2 votes -
Increase request size over 32Mb
Right now, a Firebase function can only accept a request payload size up to 32Mb.
Using it as webhook for some external APIs, we would need to have no limits or at least 100Mb limit2 votes -
Cretoin
Checking app
2 votes -
next.js dockerfile
I am deploying using the Firebase CLI with Next.js. I find it incredibly effective and I am very grateful for it, however, I've encountered a bit of a problem.
When running code with SSR that requires a certain package, an error occurs and I'm unable to proceed.
Is it possible to use a Dockerfile or a Docker container image that I've prepared myself?
Thank you very much for your assistance.
2 votes -
I want to able to set Cloud Run Execution Environment in Cloud Functions
https://docs.cloud.google.com/run/docs/configuring/execution-environments
Since Cloud Functions v2 run on Cloud Run and Cloud Run supports two different execution environment, I want to be able to set them, e.g.
const { onRequest } = require("firebase-functions/v2/https");
// Set it for a specific function
exports.myFunction = onRequest({
executionEnvironment: 'gen1', // This tells Cloud Run to use its Gen1 environment
memory: '256MiB',
}, (req, res) => {
res.send("Running on Cloud Run Gen1 environment!");
});1 vote
- Don't see your idea?